Defense LOW
Eranga Bandara, Ross Gore, Asanga Gunaratna +12 more
The rapid deployment of autonomous AI agents across enterprise, healthcare, and safety-critical environments has created a fundamental governance...
Defense HIGH
Mohamed Taoufik Kaouthar El Idrissi, Edward Zulkoski, Mohammad Hamdaqa
Code understanding models increasingly rely on pretrained language models (PLMs) and graph neural networks (GNNs), which capture complementary...
2 weeks ago cs.SE cs.LG
PDF
Defense MEDIUM
Ravikumar Balakrishnan, Sanket Mendapara
Typographic prompt injection exploits vision language models' (VLMs) ability to read text rendered in images, posing a growing threat as VLMs power...
Defense MEDIUM
Nay Myat Min, Long H. Pham, Jun Sun
Large language models deployed at runtime can misbehave in ways that clean-data validation cannot anticipate: training-time backdoors lie dormant...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Defense MEDIUM
Kaisheng Fan, Weizhe Zhang, Yishu Gao +2 more
Defending against backdoor attacks in large language models remains a critical practical challenge. Existing defenses mitigate these threats but...
2 weeks ago cs.CR cs.AI
PDF
Defense LOW
Sijia Li, Min Gao, Zongwei Wang +3 more
Sequential recommendation seeks to model the evolution of user interests by capturing temporal user intent and item-level transition patterns....
Defense HIGH
Zhaohui Geoffrey Wang
Automated code vulnerability detection is critical for software security, yet existing approaches face a fundamental trade-off between detection...
2 weeks ago cs.CR cs.LG cs.SE
PDF
Defense MEDIUM
Krishiv Agarwal, Ramneet Kaur, Colin Samplawski +6 more
Effective safety auditing of large language models (LLMs) demands tools that go beyond black-box probing and systematically uncover vulnerabilities...
2 weeks ago cs.CR cs.LG
PDF
Defense MEDIUM
Chao Pan, Yu Wu, Xin Yao
Internal Safety Collapse (ISC) is a failure mode in which frontier LLMs, when executing legitimate professional tasks whose correct completion...
2 weeks ago cs.CR cs.AI cs.LG
PDF
Defense HIGH
Ronghao Ni, Mihai Christodorescu, Limin Jia
The rapidly evolving Node$.$js ecosystem currently includes millions of packages and is a critical part of modern software supply chains, making...
2 weeks ago cs.CR cs.AI cs.SE
PDF
Defense MEDIUM
Divyesh Gabbireddy, Suman Saha
Cross-site scripting (XSS) remains a persistent web security vulnerability, especially because obfuscation can change the surface form of a malicious...
3 weeks ago cs.CR cs.LG cs.SE
PDF
Defense MEDIUM
Sarang Nambiar, Dhruv Pradhan, Ezekiel Soremekun
Pre-trained machine learning models (PTMs) are commonly provided via Model Hubs (e.g., Hugging Face) in standard formats like Pickles to facilitate...
3 weeks ago cs.CR cs.SE
PDF
Defense MEDIUM
Kun Wang, Cheng Qian, Miao Yu +6 more
Multimodal Large Language Models (MLLMs) have achieved remarkable success in cross-modal understanding and generation, yet their deployment is...
3 weeks ago cs.CR cs.AI
PDF
Defense MEDIUM
Hugo Lyons Keenan, Christopher Leckie, Sarah Erfani
We can often verify the correctness of neural network outputs using ground truth labels, but we cannot reliably determine whether the output was...
3 weeks ago cs.LG cs.CR
PDF
Defense MEDIUM
Ziyang Liu
Hosted-LLM providers have a silent-substitution incentive: advertise a stronger model while serving cheaper replies. Probe-after-return schemes such...
3 weeks ago cs.CR cs.AI
PDF
Defense LOW
Hu Wei
AI agent systems increasingly rely on reusable non-LLM engineering infrastructure that packages tool mediation, context handling, delegation, safety...
Defense MEDIUM
Dongcheng Zhang, Yiqing Jiang
Existing AI agent safety benchmarks focus on generic criminal harm (cybercrime, harassment, weapon synthesis), leaving a systematic blind spot for a...
3 weeks ago cs.CR cs.AI cs.CL
PDF
Defense MEDIUM
Ting Zhang, Yikun Li, Chengran Yang +15 more
Software vulnerabilities remain one of the most persistent threats to modern digital infrastructure. While static application security testing (SAST)...
Defense MEDIUM
Hailin Liu, Eugene Ilyushin, Jie Ni +1 more
Large language model (LLM) agents are vulnerable to prompt-injection attacks that propagate through multi-step workflows, tool interactions, and...
3 weeks ago cs.AI cs.MA
PDF
Defense LOW
Aram Ebtekar, Michael K. Cohen
Reinforcement learners can attain high reward through novel unintended strategies. We study a Bayesian mitigation for general environments: we expand...
3 weeks ago cs.LG cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial