AI Security Research
2,529+ academic papers on AI security, attacks, and defenses
Attack HIGH
Mengqi He, Xinyu Tian, Xin Shen +6 more
Recent studies show that gradient-based universal image jailbreaks on vision-language models (VLMs) exhibit little or no cross-model transferability,...
Yesterday cs.CV cs.AI
PDF
Attack HIGH
Zheng Lin, Zhenxing Niu, Haoxuan Ji +2 more
This paper proposes a jailbreaking prompt detection method for large language models (LLMs) to defend against jailbreak attacks. Although recent LLMs...
Yesterday cs.CR cs.AI
PDF
Attack HIGH
Desen Sun, Jason Hon, Howe Wang +3 more
With the rapid advancement of generative AI, users increasingly rely on image-generation models for image design and creation. To achieve faithful...
Attack HIGH
Zheng Lin, Zhenxing Niu, Haoxuan Ji +1 more
This paper proposes a guaranteed defense method for large language models (LLMs) to safeguard against jailbreaking attacks. Drawing inspiration from...
Yesterday cs.CR cs.AI
PDF
Attack HIGH
Peiru Yang, Haoran Zheng, Tong Ju +6 more
Retrieval-augmented generation (RAG) is a widely adopted paradigm for enhancing LLMs in medical applications by incorporating expert multimodal...
Yesterday cs.CR cs.AI
PDF
Attack HIGH
Farzad Nourmohammadzadeh Motlagh, Mehrdad Hajizadeh, Mehryar Majd +3 more
Natural language interfaces to structured databases are becoming increasingly common, largely due to advances in large language models (LLMs) that...
Yesterday cs.CR cs.AI
PDF
Attack HIGH
Yue Li, Xiao Li, Hao Wu +5 more
Large Language Models (LLMs) are increasingly used for automated software development, making their ability to preserve secure coding practices...
Yesterday cs.CR cs.SE
PDF
Attack HIGH
Huilin Zhou, Jian Zhao, Yilu Zhong +7 more
Red teaming is critical for uncovering vulnerabilities in Large Language Models (LLMs). While automated methods have improved scalability, existing...
Yesterday cs.LG cs.AI
PDF
Attack HIGH
Yiyong Liu, Chia-Yi Hsu, Chun-Ying Huang +3 more
LLM-powered coding agents increasingly make software supply chain decisions. They generate imports, recommend packages, and write installation...
Attack HIGH
Zeyuan Chen, Yihan Ma, Xinyue Shen +2 more
Large language models (LLMs) show strong performance across many applications, but their ability to memorize and potentially reveal training data...
Attack HIGH
Huiyu Xu, Zhibo Wang, Wenhui Zhang +4 more
Modern LLM agents solve complex tasks by operating in iterative execution loops, where they repeatedly reason, act, and self-evaluate progress to...
5 days ago cs.CR cs.AI
PDF
Attack HIGH
Md Farhamdur Reza, Richeng Jin, Tianfu Wu +1 more
Intent-obfuscation-based jailbreak attacks on multimodal large language models (MLLMs) transform a harmful query into a concealed multimodal input to...
Attack HIGH
Wesley Hanwen Deng, Mingxi Yan, Sunnie S. Y. Kim +5 more
Recent developments in AI safety research have called for red-teaming methods that effectively surface potential risks posed by generative AI models,...
5 days ago cs.HC cs.AI cs.CY
PDF
Attack HIGH
Zheng Fang, Xiaosen Wang, Shenyi Zhang +2 more
Jailbreak attacks on audio language models (ALMs) optimize audio perturbations to elicit unsafe generations, and they typically update the entire...
6 days ago cs.CR cs.AI cs.CL
PDF
Attack HIGH
Zekun Fei, Zihao Wang, Weijie Liu +4 more
Mixture-of-Experts (MoE) architectures have emerged as a leading paradigm for scaling large language models through sparse, routing-based...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial