Attack HIGH
Qingchao Shen, Zibo Xiao, Lili Huang +3 more
Large Language Models (LLMs) are increasingly deployed across diverse domains, yet their vulnerability to jailbreak attacks, where adversarial inputs...
4 weeks ago cs.CR cs.AI cs.SE
PDF
Attack MEDIUM
Hongru Song, Yu-An Liu, Ruqing Zhang +4 more
Retrieval-augmented generation (RAG) enhances large language model (LLM) reasoning by retrieving external documents, but also opens up new attack...
Attack HIGH
Dominik Blain
We present COBALT-TLA, a neuro-symbolic verification loop that pairs an LLM with TLC, the TLA+ model checker, in an automated REPL. The LLM generates...
4 weeks ago cs.CR cs.LO
PDF
Attack MEDIUM
Anes Abdennebi, Nadjia Kara, Laaziz Lahlou
The applications of Generative Artificial Intelligence (GenAI) and their intersections with data-driven fields, such as healthcare, finance,...
4 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Gamze Kirman Tokgoz, Onat Gungor, Tajana Rosing +1 more
Time-series forecasting aims to predict future values by modeling temporal dependencies in historical observations. It is a critical component of...
4 weeks ago cs.LG cs.CR
PDF
Attack LOW
Zhixiang Lu, Jionglong Su
Multimodal Large Language Models (MLLMs) in healthcare suffer from severe confirmation bias, often hallucinating visual details to support initial,...
Attack HIGH
Navid Azimi, Aditya Prakash, Yao Wang +1 more
Deep neural networks remain highly vulnerable to adversarial perturbations, limiting their reliability in security- and safety-critical applications....
1 months ago cs.CR cs.AI cs.CV
PDF
Attack MEDIUM
Shuhao Zhang, Yuli Chen, Jiale Han +2 more
Watermarking provides a critical safeguard for large language model (LLM) services by facilitating the detection of LLM-generated text....
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Yuanbo Xie, Yingjie Zhang, Yulin Li +5 more
Retrieval-Augmented Generation (RAG) systems augment large language models with external knowledge, yet introduce a critical security vulnerability:...
1 months ago cs.CR cs.AI cs.CL
PDF
Attack HIGH
Hanzhi Liu, Chaofan Shou, Hongbo Wen +3 more
Large language model (LLM) agents increasingly rely on third-party API routers to dispatch tool-calling requests across multiple upstream providers....
Attack MEDIUM
Nam Duong Tran, Phi Le Nguyen
Recent advances in Vision-Language Models (VLMs) have greatly enhanced the integration of visual perception and linguistic reasoning, driving rapid...
1 months ago cs.CV cs.AI
PDF
Attack MEDIUM
Nicolás E. Díaz Ferreyra, Monika Swetha Gurupathi, Zadia Codabux +2 more
Generative Artificial Intelligence (GenAI) has become a central component of many development tools (e.g., GitHub Copilot) that support software...
1 months ago cs.SE cs.CR cs.HC
PDF
Attack MEDIUM
Labani Halder, Payel Sadhukhan, Sarbani Palit
Ensuring reliability in adversarial settings necessitates treating privacy as a foundational component of data-driven systems. While differential...
1 months ago cs.CR cs.AI cs.LG
PDF
Attack HIGH
Wenpeng Xing, Moran Fang, Guangtai Wang +2 more
While Large Language Models (LLMs) have achieved remarkable performance, they remain vulnerable to jailbreak attacks that circumvent safety...
Attack HIGH
Wenkui Yang, Chao Jin, Haisu Zhu +7 more
Existing red-teaming studies on GUI agents have important limitations. Adversarial perturbations typically require white-box access, which is...
1 months ago cs.CR cs.CL cs.CV
PDF
Attack HIGH
Cheng Liu, Xiaolei Liu, Xingyu Li +2 more
Existing jailbreak defense paradigms primarily rely on static detection of prompts, outputs, or internal states, often neglecting the dynamic...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Zhiheng Li, Zongyang Ma, Yuntong Pan +8 more
Multimodal Large Language Models (MLLMs) are increasingly being deployed as automated content moderators. Within this landscape, we uncover a...
Attack HIGH
Zhiheng Li, Zongyang Ma, Yuntong Pan +8 more
Multimodal Large Language Models (MLLMs) are increasingly being deployed as automated content moderators. Within this landscape, we uncover a...
Attack MEDIUM
Simon Calderon, Niklas Johansson, Onur Günlü
Ensuring ciphertext indistinguishability is fundamental to cryptographic security, but empirically validating this property in real implementations...
1 months ago cs.CR cs.IT cs.LG
PDF
Attack HIGH
Ziye Wang, Guanyu Wang, Kailong Wang
Retrieval-Augmented Generation (RAG) significantly enhances Large Language Models (LLMs), but simultaneously exposes a critical vulnerability to...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial