Attack MEDIUM
Ruoyao Wen, Hao Li, Chaowei Xiao +1 more
Indirect prompt injection threatens LLM agents by embedding malicious instructions in external content, enabling unauthorized actions and data theft....
1 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Fengpeng Li, Kemou Li, Qizhou Wang +2 more
Concept erasure helps stop diffusion models (DMs) from generating harmful content; but current methods face robustness retention trade off....
1 months ago cs.LG cs.AI cs.CR
PDF
Attack MEDIUM
Tao Huang, Rui Wang, Xiaofei Liu +3 more
%Large vision-language models (LVLMs) have shown substantial advances in multimodal understanding and generation. However, when presented with...
Attack MEDIUM
Vishruti Kakkad, Paul Chung, Hanan Hibshi +1 more
An exponential growth of Machine Learning and its Generative AI applications brings with it significant security challenges, often referred to as...
1 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Yike Sun, Haotong Yang, Zhouchen Lin +1 more
Tokenization is fundamental to how language models represent and process text, yet the behavior of widely used BPE tokenizers has received far less...
Attack MEDIUM
Ariel Fogel, Omer Hofman, Eilon Cohen +1 more
Open-weight language models are increasingly used in production settings, raising new security challenges. One prominent threat in this context is...
1 months ago cs.CR cs.LG
PDF
Attack MEDIUM
Leo Schwinn, Moritz Ladenburger, Tim Beyer +3 more
Automated \enquote{LLM-as-a-Judge} frameworks have become the de facto standard for scalable evaluation across natural language processing. For...
1 months ago cs.CL cs.AI
PDF
Attack MEDIUM
Youngji Roh, Hyunjin Cho, Jaehyung Kim
Large Language Models (LLMs) exhibit highly anisotropic internal representations, often characterized by massive activations, a phenomenon where a...
Attack MEDIUM
Zeming Wei, Qiaosheng Zhang, Xia Hu +1 more
Large Reasoning Models (LRMs) have achieved tremendous success with their chain-of-thought (CoT) reasoning, yet also face safety issues similar to...
1 months ago cs.LG cs.AI cs.CL
PDF
Attack MEDIUM
Andrew Draganov, Tolga H. Dur, Anandmayi Bhongade +1 more
We present a data poisoning attack -- Phantom Transfer -- with the property that, even if you know precisely how the poison was placed into an...
1 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Matthew P. Lad, Louisa Conwill, Megan Levis Scheirer
With the rapid growth of Large Language Models (LLMs), criticism of their societal impact has also grown. Work in Responsible AI (RAI) has focused on...
Attack MEDIUM
Patrick Cooper, Alireza Nadali, Ashutosh Trivedi +1 more
Large language models (LLMs) are known to exhibit brittle behavior under adversarial prompts and jailbreak attacks, even after extensive alignment...
1 months ago cs.CL cs.AI cs.CR
PDF
Attack MEDIUM
Ching-Yun Ko, Pin-Yu Chen
Modern artificial intelligence (AI) models are deployed on inference engines to optimize runtime efficiency and resource allocation, particularly for...
1 months ago cs.LG cs.CL cs.PL
PDF
Attack MEDIUM
Poushali Sengupta, Shashi Raj Pandey, Sabita Maharjan +1 more
Large language models (LLMs) generate outputs by utilizing extensive context, which often includes redundant information from prompts, retrieved...
1 months ago cs.CL cs.AI stat.ML
PDF
Attack MEDIUM
Eliron Rahimi, Elad Hirshel, Rom Himelstein +3 more
Diffusion language models (DLMs) have recently emerged as a promising alternative to autoregressive (AR) models, offering parallel decoding and...
1 months ago cs.LG cs.AI
PDF
Attack MEDIUM
Xinyi Hou, Shenao Wang, Yifan Zhang +4 more
Agentic AI systems built around large language models (LLMs) are moving away from closed, single-model frameworks and toward open ecosystems that...
Attack MEDIUM
Manveer Singh Tamber, Hosna Oyarhoseini, Jimmy Lin
Research on adversarial robustness in language models is currently fragmented across applications and attacks, obscuring shared vulnerabilities. In...
1 months ago cs.CL cs.IR
PDF
Attack MEDIUM
Haitham S. Al-Sinani, Chris J. Mitchell
Wireless ethical hacking relies heavily on skilled practitioners manually interpreting reconnaissance results and executing complex, time-sensitive...
1 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Mingqian Feng, Xiaodong Liu, Weiwei Yang +3 more
Large Language Models (LLMs) are typically evaluated for safety under single-shot or low-budget adversarial prompting, which underestimates...
Attack MEDIUM
Amirhossein Taherpour, Xiaodong Wang
Federated learning (FL) enables collaborative model training while preserving data privacy, yet both centralized and decentralized approaches face...
1 months ago cs.LG cs.CR cs.DC
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial