Benchmark MEDIUM
Amir Al-Maamari
Large Language Models (LLMs) show promise for Automated Program Repair (APR), yet their effectiveness on security vulnerabilities remains poorly...
2 weeks ago cs.CR cs.AI
PDF
Attack MEDIUM
Shaswata Mitra, Raj Patel, Sudip Mittal +2 more
Multi-agent systems (MAS) powered by LLMs promise adaptive, reasoning-driven enterprise workflows, yet granting agents autonomous control over tools,...
2 weeks ago cs.CR cs.MA cs.SE
PDF
Defense MEDIUM
Harry Owiredu-Ashley
Most adversarial evaluations of large language model (LLM) safety assess single prompts and report binary pass/fail outcomes, which fails to capture...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Tool MEDIUM
Yinpeng Wu, Yitong Chen, Lixiang Wang +3 more
Device-side Large Language Models (LLMs) have witnessed explosive growth, offering higher privacy and availability compared to cloud-side LLMs....
2 weeks ago cs.CR cs.LG cs.OS
PDF
Attack MEDIUM
Alexander Erlei, Lukas Meub
As AI agents increasingly act on behalf of human stakeholders in economic settings, understanding their behavior in complex market environments...
Defense MEDIUM
Bo Jiang
Knowledge distillation from proprietary LLM APIs poses a growing threat to model providers, yet defenses against this attack remain fragmented and...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Defense MEDIUM
Sumit Ranjan, Sugandha Sharma, Ubaid Abbas +1 more
Voice interfaces are quickly becoming a common way for people to interact with AI systems. This also brings new security risks, such as prompt...
2 weeks ago cs.SD cs.AI
PDF
Benchmark MEDIUM
Chenxi Li, Xianggan Liu, Dake Shen +9 more
Despite the rapid progress of Large Vision-Language Models (LVLMs), the integration of visual modalities introduces new safety vulnerabilities that...
2 weeks ago cs.CV cs.LG
PDF
Tool MEDIUM
Yuhang Huang, Boyang Ma, Biwei Yan +5 more
The Model Context Protocol (MCP) is an open and standardized interface that enables large language models (LLMs) to interact with external tools and...
2 weeks ago cs.CR cs.AI
PDF
Tool MEDIUM
Neha Nagaraja, Hayretdin Bahsi
Large Language Models (LLMs) are increasingly integrated into safety-critical workflows, yet existing security analyses remain fragmented and often...
2 weeks ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Yige Li, Wei Zhao, Zhe Li +6 more
Backdoor mechanisms have traditionally been studied as security threats that compromise the integrity of machine learning models. However, the same...
2 weeks ago cs.CR cs.AI
PDF
Attack MEDIUM
Eduard Hirsch, Kristina Raab, Tobias J. Bauer +1 more
IT systems are facing an increasing number of security threats, including advanced persistent attacks and future quantum-computing vulnerabilities....
2 weeks ago cs.CR cs.IR
PDF
Benchmark MEDIUM
Yuxu Ge
Autonomous agents powered by large language models introduce a class of execution-layer vulnerabilities -- prompt injection, retrieval poisoning, and...
2 weeks ago cs.CR cs.AI
PDF
Tool MEDIUM
Punyajoy Saha, Sudipta Halder, Debjyoti Mondal +1 more
Safety alignment is critical for deploying large language models (LLMs) in real-world applications, yet most existing approaches rely on large...
2 weeks ago cs.CL cs.AI cs.LG
PDF
Survey MEDIUM
Elzo Brito dos Santos Filho
AI-assisted software generation has increased development speed, but it has also amplified a persistent engineering problem: systems that are...
2 weeks ago cs.CR cs.AI
PDF
Attack MEDIUM
Donghwa Kang, Hojun Choe, Doohyun Kim +2 more
Deploying deep neural networks (DNNs) on edge devices exposes valuable intellectual property to model-stealing attacks. While TEE-shielded DNN...
Defense MEDIUM
Xisen Jin, Michael Duan, Qin Lin +4 more
As AI agents become widely deployed as online services, users often rely on an agent developer's claim about how safety is enforced, which introduces...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Defense MEDIUM
Jinman Wu, Yi Xie, Shen Lin +2 more
Safety alignment is often conceptualized as a monolithic process wherein harmfulness detection automatically triggers refusal. However, the...
2 weeks ago cs.CR cs.AI cs.LG
PDF
Defense MEDIUM
Ved Sriraman, Adam Block
Best-of-N (BoN) sampling is a widely used inference-time alignment method for language models, whereby N candidate responses are sampled from a...
2 weeks ago cs.LG cs.AI
PDF
Benchmark MEDIUM
Xiaoguang Li, Hanyi Wang, Yaowei Huang +6 more
Shuffler-based differential privacy (shuffle-DP) is a privacy paradigm providing high utility by involving a shuffler to permute noisy report from...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial