Attack HIGH
Ravikumar Balakrishnan, Sanket Mendapara, Ankit Garg
We study typographic prompt injection attacks on vision-language models (VLMs), where adversarial text is rendered as images to bypass safety...
Attack HIGH
Yulin Chen, Tri Cao, Haoran Li +7 more
Web agents powered by vision-language models (VLMs) enable autonomous interaction with web environments by perceiving and acting on both visual and...
Attack HIGH
Qingchao Shen, Zibo Xiao, Lili Huang +3 more
Large Language Models (LLMs) are increasingly deployed across diverse domains, yet their vulnerability to jailbreak attacks, where adversarial inputs...
4 weeks ago cs.CR cs.AI cs.SE
PDF
Attack HIGH
Dominik Blain
We present COBALT-TLA, a neuro-symbolic verification loop that pairs an LLM with TLC, the TLA+ model checker, in an automated REPL. The LLM generates...
4 weeks ago cs.CR cs.LO
PDF
Attack HIGH
Gamze Kirman Tokgoz, Onat Gungor, Tajana Rosing +1 more
Time-series forecasting aims to predict future values by modeling temporal dependencies in historical observations. It is a critical component of...
4 weeks ago cs.LG cs.CR
PDF
Attack HIGH
Navid Azimi, Aditya Prakash, Yao Wang +1 more
Deep neural networks remain highly vulnerable to adversarial perturbations, limiting their reliability in security- and safety-critical applications....
1 months ago cs.CR cs.AI cs.CV
PDF
Attack HIGH
Yuanbo Xie, Yingjie Zhang, Yulin Li +5 more
Retrieval-Augmented Generation (RAG) systems augment large language models with external knowledge, yet introduce a critical security vulnerability:...
1 months ago cs.CR cs.AI cs.CL
PDF
Attack HIGH
Hanzhi Liu, Chaofan Shou, Hongbo Wen +3 more
Large language model (LLM) agents increasingly rely on third-party API routers to dispatch tool-calling requests across multiple upstream providers....
Attack HIGH
Wenpeng Xing, Moran Fang, Guangtai Wang +2 more
While Large Language Models (LLMs) have achieved remarkable performance, they remain vulnerable to jailbreak attacks that circumvent safety...
Attack HIGH
Wenkui Yang, Chao Jin, Haisu Zhu +7 more
Existing red-teaming studies on GUI agents have important limitations. Adversarial perturbations typically require white-box access, which is...
1 months ago cs.CR cs.CL cs.CV
PDF
Attack HIGH
Cheng Liu, Xiaolei Liu, Xingyu Li +2 more
Existing jailbreak defense paradigms primarily rely on static detection of prompts, outputs, or internal states, often neglecting the dynamic...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Zhiheng Li, Zongyang Ma, Yuntong Pan +8 more
Multimodal Large Language Models (MLLMs) are increasingly being deployed as automated content moderators. Within this landscape, we uncover a...
Attack HIGH
Zhiheng Li, Zongyang Ma, Yuntong Pan +8 more
Multimodal Large Language Models (MLLMs) are increasingly being deployed as automated content moderators. Within this landscape, we uncover a...
Attack HIGH
Ziye Wang, Guanyu Wang, Kailong Wang
Retrieval-Augmented Generation (RAG) significantly enhances Large Language Models (LLMs), but simultaneously exposes a critical vulnerability to...
Attack HIGH
Yizhe Zeng, Wei Zhang, Yunpeng Li +3 more
While Chain-of-Thought (CoT) prompting has become a standard paradigm for eliciting complex reasoning capabilities in Large Language Models, it...
Attack HIGH
Adrian Shuai Li, Md Ajwad Akil, Elisa Bertino
Concept drift and adversarial evasion are two major challenges for deploying machine learning-based malware detectors. While both have been studied...
Attack HIGH
Manish Bhatt, Sarthak Munshi, Vineeth Sai Narajala +4 more
We prove that no continuous, utility-preserving wrapper defense-a function $D: X\to X$ that preprocesses inputs before the model sees them-can make...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Zonghao Ying, Haowen Dai, Lianyu Hu +5 more
Modern text-to-image (T2I) models can now render legible, paragraph-length text, enabling a fundamentally new class of misuse. We identify and...
Attack HIGH
Zonghao Ying, Haowen Dai, Lianyu Hu +5 more
Modern text-to-image (T2I) models can now render legible, paragraph-length text, enabling a fundamentally new class of misuse. We identify and...
Attack HIGH
Yiyang Zhang, Chaojian Yu, Ziming Hong +4 more
Multimodal pretrained models are vulnerable to backdoor attacks, yet most existing methods rely on visual or multimodal triggers, which are...
1 months ago cs.CR cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial