Attack MEDIUM
Mingyang Liao, Yichen Wan, shuchen wu +6 more
LLM-based role-playing has rapidly improved in fidelity, yet stronger adherence to persona constraints commonly increases vulnerability to jailbreak...
Attack HIGH
Ningyuan He, Ronghong Huang, Qianqian Tang +3 more
In-context learning (ICL) has become a powerful, data-efficient paradigm for text classification using large language models. However, its robustness...
Attack MEDIUM
Wenhui Zhang, Huiyu Xu, Zhibo Wang +4 more
Recent advancements in multi-model AI systems have leveraged LLM routers to reduce computational cost while maintaining response quality by assigning...
Attack MEDIUM
Alvi Md Ishmam, Najibul Haque Sarker, Zaber Ibn Abdul Hakim +1 more
Multimodal Large Language Models (MLLMs) have achieved remarkable performance across vision-language tasks. Recent advancements allow these models to...
Attack MEDIUM
Arther Tian, Alex Ding, Frank Chen +2 more
Decentralized large language model inference networks require lightweight mechanisms to reward high quality outputs under heterogeneous latency and...
3 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Jarrod Barnes
As large language models (LLMs) improve, so do their offensive applications: frontier agents now generate working exploits for under $50 in compute...
Attack MEDIUM
Onkar Shelar, Travis Desell
Evolutionary prompt search is a practical black-box approach for red teaming large language models (LLMs), but existing methods often collapse onto a...
3 months ago cs.NE q-bio.PE
PDF
Attack HIGH
Xingwei Lin, Wenhao Lin, Sicong Cao +4 more
Multi-turn jailbreak attacks have emerged as a critical threat to Large Language Models (LLMs), bypassing safety mechanisms by progressively...
3 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Yizhong Ding
Webshells remain a primary foothold for attackers to compromise servers, particularly within PHP ecosystems. However, existing detection mechanisms...
3 months ago cs.CR cs.AI
PDF
Attack HIGH
Yuetian Chen, Kaiyuan Zhang, Yuntao Du +5 more
Diffusion Language Models (DLMs) represent a promising alternative to autoregressive language models, using bidirectional masked token prediction....
3 months ago cs.LG cs.AI
PDF
Attack HIGH
Md Tasnim Jawad, Mingyan Xiao, Yanzhao Wu
With the widespread adoption of Large Language Models (LLMs) and increasingly stringent privacy regulations, protecting data privacy in LLMs has...
Attack HIGH
Haonan Zhang, Dongxia Wang, Yi Liu +2 more
Safety-aligned LLMs suffer from two failure modes: jailbreak (answering harmful inputs) and over-refusal (declining benign queries). Existing vector...
3 months ago cs.LG cs.AI
PDF
Attack MEDIUM
Yangyang Guo, Ziwei Xu, Si Liu +2 more
This study reveals a previously unexplored vulnerability in the safety alignment of Large Language Models (LLMs). Existing aligned LLMs predominantly...
Attack MEDIUM
Sen Nie, Jie Zhang, Zhuo Wang +2 more
Vision-language models (VLMs) such as CLIP have demonstrated remarkable zero-shot generalization, yet remain highly vulnerable to adversarial...
Attack HIGH
Harsh Chaudhari, Ethan Rathbun, Hanna Foerster +5 more
Chain-of-Thought (CoT) reasoning has emerged as a powerful technique for enhancing large language models' capabilities by generating intermediate...
3 months ago cs.CR cs.LG
PDF
Attack HIGH
Gabriel Lee Jun Rong, Christos Korgialas, Dion Jia Xu Ho +3 more
Existing automated attack suites operate as static ensembles with fixed sequences, lacking strategic adaptation and semantic awareness. This paper...
Attack HIGH
Alexandra Chouldechova, A. Feder Cooper, Solon Barocas +3 more
We argue that conclusions drawn about relative system safety or attack method efficacy via AI red teaming are often not supported by evidence...
Attack HIGH
Narek Maloyan, Dmitry Namiot
The proliferation of agentic AI coding assistants, including Claude Code, GitHub Copilot, Cursor, and emerging skill-based architectures, has...
Attack HIGH
Chen Ling, Kai Hu, Hangcheng Liu +3 more
Large Vision-Language Models (LVLMs) are increasingly deployed in real-world intelligent systems for perception and reasoning in open physical...
3 months ago cs.CV cs.AI
PDF
Attack HIGH
Mohammad Zare, Pirooz Shamsinejadbabaki
Membership inference attacks (MIAs) pose a serious threat to the privacy of machine learning models by allowing adversaries to determine whether a...
3 months ago cs.CR cs.AI cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial