Tool HIGH
Yihao Zhang, Zeming Wei, Xiaokun Luan +7 more
Autonomous LLM-based agents increasingly operate as long-running processes forming densely interconnected multi-agent ecosystems, whose security...
1 months ago cs.CR cs.AI cs.LG
PDF
Tool HIGH
Yihao Zhang, Zeming Wei, Xiaokun Luan +7 more
Autonomous LLM-based agents increasingly operate as long-running processes forming densely interconnected multi-agent ecosystems, whose security...
1 months ago cs.CR cs.AI cs.LG
PDF
Attack HIGH
Mateusz Dziemian, Maxwell Lin, Xiaohan Fu +28 more
LLM based agents are increasingly deployed in high stakes settings where they process external data sources such as emails, documents, and code...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Zhenlin Xu, Xiaogang Zhu, Yu Yao +2 more
Modern agentic systems allow Large Language Model (LLM) agents to tackle complex tasks through extensive tool usage, forming structured control flows...
Benchmark HIGH
Lidor Erez, Omer Hofman, Tamir Nizri +1 more
Automated LLM vulnerability scanners are increasingly used to assess security risks by measuring different attack type success rates (ASR). Yet the...
2 months ago cs.CR cs.PF
PDF
Attack HIGH
Maël Jenny, Jérémie Dentan, Sonia Vanier +1 more
Most jailbreak techniques for Large Language Models (LLMs) primarily rely on prompt modifications, including paraphrasing, obfuscation, or...
Attack HIGH
Chongxin Li, Hanzhang Wang, Lian Duan
Safety prompts constitute an interpretable layer of defense against jailbreak attacks in vision-language models (VLMs); however, their efficacy is...
Attack HIGH
Yiling Tao, Xinran Zheng, Shuo Yang +2 more
While large language model-based agents demonstrate great potential in collaborative tasks, their interactivity also introduces security...
Attack HIGH
Zijian Ling, Pingyi Hu, Xiuyong Gao +6 more
Speech-driven large language models (LLMs) are increasingly accessed through speech interfaces, introducing new security risks via open acoustic...
2 months ago cs.CR cs.AI cs.SD
PDF
Attack HIGH
Chenlong Yin, Runpeng Geng, Yanting Wang +1 more
Prompt injection poses serious security risks to real-world LLM applications, particularly autonomous agents. Although many defenses have been...
2 months ago cs.LG cs.CR
PDF
Attack HIGH
Zheng Gao, Yifan Yang, Xiaoyu Li +4 more
Watermarking the initial noise of diffusion models has emerged as a promising approach for image provenance, but content-independent noise patterns...
2 months ago cs.CV cs.CR cs.LG
PDF
Attack HIGH
Sihao Ding
We introduce Colluding LoRA (CoLoRA), an attack in which each adapter appears benign and plausibly functional in isolation, yet their linear...
2 months ago cs.CR cs.LG
PDF
Attack HIGH
Darren Cheng, Wen-Kwang Tsao
Prompt injection remains one of the most practical attack vectors against LLM-integrated applications. We replicate the Microsoft LLMail-Inject...
2 months ago cs.CR cs.AI
PDF
Benchmark HIGH
Siddharth Srikanth, Freddie Liang, Sophie Hsu +9 more
Vision-Language-Action (VLA) models have significant potential to enable general-purpose robotic systems for a range of vision-language tasks....
2 months ago cs.RO cs.AI cs.CL
PDF
Attack HIGH
Xinhai Wang, Shaopeng Fu, Shu Yang +3 more
Suffix jailbreak attacks serve as a systematic method for red-teaming Large Language Models (LLMs) but suffer from prohibitive computational costs,...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Davi Bonetto
State Space Models (SSMs) such as Mamba achieve linear-time sequence processing through input-dependent recurrence, but this mechanism introduces a...
2 months ago cs.LG cs.CR
PDF
Attack HIGH
Alexandre Le Mercier, Thomas Demeester, Chris Develder
State space models (SSMs) like Mamba have gained significant traction as efficient alternatives to Transformers, achieving linear complexity while...
Tool HIGH
Sarbartha Banerjee, Prateek Sahu, Anjo Vahldiek-Oberwagner +2 more
Rapid progress in generative AI has given rise to Compound AI systems - pipelines comprised of multiple large language models (LLM), software tools...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
J Alex Corll
Prompt injection defenses are often framed as semantic understanding problems and delegated to increasingly large neural detectors. For the first...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Indranil Halder, Annesya Banerjee, Cengiz Pehlevan
Adversarial attacks can reliably steer safety-aligned large language models toward unsafe behavior. Empirically, we find that adversarial...
2 months ago cs.LG cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial