Attack HIGH
Dong Yan, Jian Liang, Ran He +1 more
Recent studies have shown that large language models (LLMs) can infer private user attributes (e.g., age, location, gender) from user-generated text...
1 months ago cs.CR cs.AI cs.CL
PDF
Attack HIGH
Sri Durga Sai Sowmya Kadali, Evangelos E. Papalexakis
Jailbreaking large language models (LLMs) has emerged as a critical security challenge with the widespread deployment of conversational AI systems....
1 months ago cs.CR cs.CL
PDF
Attack HIGH
J Alex Corll
Multi-turn prompt injection attacks distribute malicious intent across multiple conversation turns, exploiting the assumption that each turn is...
1 months ago cs.CR cs.AI
PDF
Defense HIGH
Samal Mukhtar, Yinghua Yao, Zhu Sun +3 more
Software vulnerability detection (SVD) is a critical challenge in modern systems. Large language models (LLMs) offer natural-language explanations...
1 months ago cs.SE cs.AI cs.CR
PDF
Attack HIGH
Shuyu Chang, Haiping Huang, Yanjun Zhang +3 more
Code models are increasingly adopted in software development but remain vulnerable to backdoor attacks via poisoned training data. Existing backdoor...
1 months ago cs.CR cs.SE
PDF
Attack HIGH
Qianli Wang, Boyang Ma, Minghui Xu +1 more
LLM agents often rely on Skills to describe available tools and recommended procedures. We study a hidden-comment prompt injection risk in this...
Survey HIGH
Peiran Wang, Xinfeng Li, Chong Xiang +5 more
The evolution of Large Language Models (LLMs) has resulted in a paradigm shift towards autonomous agents, necessitating robust security against...
1 months ago cs.CR cs.CL
PDF
Attack HIGH
Tri Nguyen, Huy Hoang Bao Le, Lohith Srikanth Pentapalli +2 more
Detecting jailbreak attempts in clinical training large language models (LLMs) requires accurate modeling of linguistic deviations that signal unsafe...
1 months ago cs.AI cs.LG
PDF
Benchmark HIGH
Adriana Alvarado Garcia, Ruyuan Wan, Ozioma C. Oguine +1 more
Recently, red teaming, with roots in security, has become a key evaluative approach to ensure the safety and reliability of Generative Artificial...
1 months ago cs.CY cs.AI cs.CL
PDF
Survey HIGH
George Tsigkourakos, Constantinos Patsakis
Static Application Security Testing (SAST) tools are integral to modern DevSecOps pipelines, yet tools like CodeQL, Semgrep, and SonarQube remain...
Tool HIGH
Hayfa Dhabhi, Kashyap Thimmaraju
Large Language Models (LLMs) deploy safety mechanisms to prevent harmful outputs, yet these defenses remain vulnerable to adversarial prompts. While...
1 months ago cs.CR cs.AI cs.CY
PDF
Benchmark HIGH
Chaeyun Kim, YongTaek Lim, Kihyun Kim +2 more
Existing red-teaming benchmarks, when adapted to new languages via direct translation, fail to capture socio-technical vulnerabilities rooted in...
1 months ago cs.CY cs.AI
PDF
Attack HIGH
Georgios Syros, Evan Rose, Brian Grinstead +4 more
Large language model (LLM) based web agents are increasingly deployed to automate complex online tasks by directly interacting with web sites and...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Kotekar Annapoorna Prabhu, Andrew Gan, Zahra Ghodsi
Machine learning relies on randomness as a fundamental component in various steps such as data sampling, data augmentation, weight initialization,...
1 months ago cs.CR cs.LG
PDF
Attack HIGH
Yu Yan, Sheng Sun, Shengjia Cheng +3 more
Vision-Language Models (VLMs) with multimodal reasoning capabilities are high-value attack targets, given their potential for handling complex...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Suraj Ranganath, Atharv Ramesh
AI-text detectors face a critical robustness challenge: adversarial paraphrasing attacks that preserve semantics while evading detection. We...
1 months ago cs.LG cs.AI cs.CR
PDF
Attack HIGH
Suraj Ranganath, Atharv Ramesh
AI-text detectors face a critical robustness challenge: adversarial paraphrasing attacks that preserve semantics while evading detection. We...
1 months ago cs.LG cs.AI cs.CR
PDF
Attack HIGH
Jona te Lintelo, Lichao Wu, Stjepan Picek
The rapid adoption of Mixture-of-Experts (MoE) architectures marks a major shift in the deployment of Large Language Models (LLMs). MoE LLMs improve...
Attack HIGH
Yanzhang Fu, Zizheng Guo, Jizhou Luo
Score-based query attacks pose a serious threat to deep learning models by crafting adversarial examples (AEs) using only black-box access to model...
1 months ago cs.LG cs.CR
PDF
Attack HIGH
Scott Thornton
Hybrid Retrieval-Augmented Generation (RAG) pipelines combine vector similarity search with knowledge graph expansion for multi-hop reasoning. We...
1 months ago cs.CR cs.IR cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial