Benchmark LOW
Raquib Bin Yousuf, Aadyant Khatri, Shengzhe Xu +2 more
Recently proposed evaluation benchmarks aim to characterize the effective context length and the forgetting tendencies of large language models...
7 months ago cs.CL cs.AI cs.LG
PDF
Attack MEDIUM
Fatmazohra Rezkellah, Ramzi Dakhmouche
With the increasing adoption of Large Language Models (LLMs), more customization is needed to ensure privacy-preserving and safe generation. We...
7 months ago cs.LG cs.CL cs.CR
PDF
Benchmark MEDIUM
Kartik Pandit, Sourav Ganguly, Arnesh Banerjee +2 more
Ensuring safety is a foundational requirement for large language models (LLMs). Achieving an appropriate balance between enhancing the utility of...
7 months ago cs.LG cs.AI eess.SY
PDF
Attack HIGH
Javad Rafiei Asl, Sidhant Narula, Mohammad Ghasemigol +2 more
Large Language Models (LLMs) have revolutionized natural language processing but remain vulnerable to jailbreak attacks, especially multi-turn...
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Sanket Badhe
We present LegalSim, a modular multi-agent simulation of adversarial legal proceedings that explores how AI systems can exploit procedural weaknesses...
7 months ago cs.MA cs.AI cs.CR
PDF
Benchmark MEDIUM
Imene Kerboua, Sahar Omidi Shayegan, Megh Thakkar +7 more
Web agents powered by large language models (LLMs) must process lengthy web page observations to complete user goals; these pages often exceed tens...
Attack HIGH
Xinzhe Huang, Wenjing Hu, Tianhang Zheng +5 more
Existing gradient-based jailbreak attacks on Large Language Models (LLMs) typically optimize adversarial suffixes to align the LLM output with...
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Yu He, Yifei Chen, Yiming Li +5 more
In recent years, RAG has emerged as a key paradigm for enhancing large language models (LLMs). By integrating externally retrieved information, RAG...
Benchmark MEDIUM
Léo Boisvert, Abhay Puri, Chandra Kiran Reddy Evuru +6 more
While finetuning AI agents on interaction data -- such as web browsing or tool use -- improves their capabilities, it also introduces critical...
7 months ago cs.CR cs.AI cs.LG
PDF
Benchmark MEDIUM
Nikoo Naghavian, Mostafa Tavassolipour
Vision-language models like CLIP demonstrate impressive zero-shot generalization but remain highly vulnerable to adversarial attacks. In this work,...
Attack HIGH
Zhixin Xie, Xurui Song, Jun Luo
Despite substantial efforts in safety alignment, recent research indicates that Large Language Models (LLMs) remain highly susceptible to jailbreak...
Attack MEDIUM
Abrar Shahid, Ibteeker Mahir Ishum, AKM Tahmidul Haque +2 more
This paper presents a controlled study of adversarial reinforcement learning in network security through a custom OpenAI Gym environment that models...
7 months ago cs.LG cs.AI cs.CR
PDF
Defense MEDIUM
Lesly Miculicich, Mihir Parmar, Hamid Palangi +4 more
The deployment of autonomous AI agents in sensitive domains, such as healthcare, introduces critical risks to safety, security, and privacy. These...
7 months ago cs.SE cs.AI cs.CR
PDF
Attack HIGH
Chinthana Wimalasuriya, Spyros Tragoudas
Adversarial attacks present a significant threat to modern machine learning systems. Yet, existing detection methods often lack the ability to detect...
7 months ago cs.CR cs.CV cs.LG
PDF
Tool MEDIUM
Bowei Ning, Xuejun Zong, Kan He
Industrial control systems (ICS) are vital to modern infrastructure but increasingly vulnerable to cybersecurity threats, particularly through...
Attack HIGH
Zhaorun Chen, Xun Liu, Mintong Kang +4 more
As vision-language models (VLMs) gain prominence, their multimodal interfaces also introduce new safety vulnerabilities, making the safety evaluation...
7 months ago cs.AI cs.LG
PDF
Benchmark HIGH
Chengquan Guo, Chulin Xie, Yu Yang +6 more
Code agents have gained widespread adoption due to their strong code generation capabilities and integration with code interpreters, enabling dynamic...
Benchmark MEDIUM
Chenpei Huang, Lingfeng Yao, Hui Zhong +5 more
Ear canal scanning/sensing (ECS) has emerged as a novel biometric authentication method for mobile devices paired with wireless earbuds. Existing...
7 months ago cs.CR cs.HC
PDF
Tool HIGH
Jonathan Sneh, Ruomei Yan, Jialin Yu +6 more
As LLMs increasingly power agents that interact with external tools, tool use has become an essential mechanism for extending their capabilities....
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Ruohao Guo, Afshin Oroojlooy, Roshan Sridhar +3 more
Despite recent rapid progress in AI safety, current large language models remain vulnerable to adversarial attacks in multi-turn interaction...
7 months ago cs.LG cs.AI cs.CL
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial