AI Security Research
2,529+ academic papers on AI security, attacks, and defenses
Tool MEDIUM
Joel Rorseth, Parke Godfrey, Lukasz Golab +2 more
This paper demonstrates RUBEN, an interactive tool for discovering minimal rules to explain the outputs of retrieval-augmented large language models...
Tool MEDIUM
Michael A. Riegler, Inga Strümke
We present swarm-attack, an open-source adversarial testing framework in which multiple lightweight LLM agents coordinate through shared memory,...
2 days ago cs.CR cs.AI cs.LG
PDF
Tool MEDIUM
Chengjie Wang, Jingzheng Wu, Xiang Ling +2 more
Large language models (LLMs) are now largely involved in software development workflows, and the code they generate routinely includes third-party...
5 days ago cs.SE cs.AI
PDF
Tool MEDIUM
Kerri Prinos, Lilianne Brush, Cameron Denton +5 more
Agentic systems involved in high-stake decision-making under adversarial pressure need formal guarantees not offered by existing approaches....
1 weeks ago cs.AI cs.CR eess.SY
PDF
Tool MEDIUM
Mingming Zha, Xiaofeng Wang
Autonomous LLM agents operate as long-running processes with persistent workspaces, memory files, scheduled task state, and messaging integrations....
Tool MEDIUM
Neha Nagaraja, Hayretdin Bahsi, Carlo R. da Cunha
As large language models are integrated into autonomous robotic systems for task planning and control, compromised inputs or unsafe model outputs can...
1 weeks ago cs.CR cs.AI cs.RO
PDF
Tool MEDIUM
Kato Mivule
This paper extends the Classification Error Gauge (x-CEG) framework, originally developed for measuring the privacy-utility trade-off in tabular...
Tool MEDIUM
Mikko Lempinen, Joni Kemppainen, Niklas Raesalmi
As artificial intelligence (AI) systems are increasingly deployed across critical domains, their security vulnerabilities pose growing risks of...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Tool MEDIUM
Yuan Fang, Yiming Luo, Aimin Zhou +1 more
Ensuring the safety of large language models (LLMs) requires robust red teaming, yet the systematic synthesis of high-quality toxic data remains...
3 weeks ago cs.CL cs.AI
PDF
Tool MEDIUM
Shangkun Che, Silin Du, Ge Gao
The widespread use of Large Language Models (LLMs) in text generation has raised increasing concerns about intellectual property disputes....
4 weeks ago cs.CR cs.CL
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial