Attack HIGH
Osama Zafar, Shaojie Zhan, Tianxi Ji +1 more
In recent years, the widespread adoption of Machine Learning as a Service (MLaaS), particularly in sensitive environments, has raised considerable...
Attack HIGH
Yannick Assogba, Jacopo Cortellazzi, Javier Abad +3 more
Jailbreak attacks remain a persistent threat to large language model safety. We propose Context-Conditioned Delta Steering (CC-Delta), an SAE-based...
1 months ago cs.CR cs.CL cs.LG
PDF
Attack HIGH
Dong Yan, Jian Liang, Ran He +1 more
Recent studies have shown that large language models (LLMs) can infer private user attributes (e.g., age, location, gender) from user-generated text...
1 months ago cs.CR cs.AI cs.CL
PDF
Attack HIGH
Sri Durga Sai Sowmya Kadali, Evangelos E. Papalexakis
Jailbreaking large language models (LLMs) has emerged as a critical security challenge with the widespread deployment of conversational AI systems....
1 months ago cs.CR cs.CL
PDF
Attack HIGH
J Alex Corll
Multi-turn prompt injection attacks distribute malicious intent across multiple conversation turns, exploiting the assumption that each turn is...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Shuyu Chang, Haiping Huang, Yanjun Zhang +3 more
Code models are increasingly adopted in software development but remain vulnerable to backdoor attacks via poisoned training data. Existing backdoor...
1 months ago cs.CR cs.SE
PDF
Attack HIGH
Qianli Wang, Boyang Ma, Minghui Xu +1 more
LLM agents often rely on Skills to describe available tools and recommended procedures. We study a hidden-comment prompt injection risk in this...
Attack HIGH
Tri Nguyen, Huy Hoang Bao Le, Lohith Srikanth Pentapalli +2 more
Detecting jailbreak attempts in clinical training large language models (LLMs) requires accurate modeling of linguistic deviations that signal unsafe...
1 months ago cs.AI cs.LG
PDF
Attack HIGH
Georgios Syros, Evan Rose, Brian Grinstead +4 more
Large language model (LLM) based web agents are increasingly deployed to automate complex online tasks by directly interacting with web sites and...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Kotekar Annapoorna Prabhu, Andrew Gan, Zahra Ghodsi
Machine learning relies on randomness as a fundamental component in various steps such as data sampling, data augmentation, weight initialization,...
1 months ago cs.CR cs.LG
PDF
Attack HIGH
Yu Yan, Sheng Sun, Shengjia Cheng +3 more
Vision-Language Models (VLMs) with multimodal reasoning capabilities are high-value attack targets, given their potential for handling complex...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Suraj Ranganath, Atharv Ramesh
AI-text detectors face a critical robustness challenge: adversarial paraphrasing attacks that preserve semantics while evading detection. We...
1 months ago cs.LG cs.AI cs.CR
PDF
Attack HIGH
Suraj Ranganath, Atharv Ramesh
AI-text detectors face a critical robustness challenge: adversarial paraphrasing attacks that preserve semantics while evading detection. We...
1 months ago cs.LG cs.AI cs.CR
PDF
Attack HIGH
Jona te Lintelo, Lichao Wu, Stjepan Picek
The rapid adoption of Mixture-of-Experts (MoE) architectures marks a major shift in the deployment of Large Language Models (LLMs). MoE LLMs improve...
Attack HIGH
Yanzhang Fu, Zizheng Guo, Jizhou Luo
Score-based query attacks pose a serious threat to deep learning models by crafting adversarial examples (AEs) using only black-box access to model...
1 months ago cs.LG cs.CR
PDF
Attack HIGH
Scott Thornton
Hybrid Retrieval-Augmented Generation (RAG) pipelines combine vector similarity search with knowledge graph expansion for multi-hop reasoning. We...
1 months ago cs.CR cs.IR cs.LG
PDF
Attack HIGH
Sahar Zargarzadeh, Mohammad Islam
The Internet of Things (IoT) has revolutionized connectivity by linking billions of devices worldwide. However, this rapid expansion has also...
1 months ago cs.CR cs.LG
PDF
Attack HIGH
Md Rafi Ur Rashid, MD Sadik Hossain Shanto, Vishnu Asutosh Dasu +1 more
Vision-Language Models (VLMs) are now a core part of modern AI. Recent work proposed several visual jailbreak attacks using single/ holistic images....
1 months ago cs.CV cs.AI
PDF
Attack HIGH
Minbeom Kim, Mihir Parmar, Phillip Wallis +5 more
AI agents equipped with tool-calling capabilities are susceptible to Indirect Prompt Injection (IPI) attacks. In this attack scenario, malicious...
1 months ago cs.CR cs.LG stat.ME
PDF
Attack HIGH
Yuhao Wang, Shengfang Zhai, Guanghao Jin +3 more
Large Language Model (LLM)-based agents employ external and internal memory systems to handle complex, goal-oriented tasks, yet this exposes them to...
1 months ago cs.CR cs.AI cs.CL
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial