Tool MEDIUM
Yuan Fang, Yiming Luo, Aimin Zhou +1 more
Ensuring the safety of large language models (LLMs) requires robust red teaming, yet the systematic synthesis of high-quality toxic data remains...
3 weeks ago cs.CL cs.AI
PDF
Tool LOW
Shawn, Zhong, Junxuan Liao +4 more
AI coding agents operate directly on users' filesystems, where they regularly corrupt data, delete files, and leak secrets. Current approaches force...
Tool LOW
Syed Md Mukit Rashid, Abdullah Al Ishtiaq, Kai Tu +7 more
Logical vulnerabilities in software stem from flaws in program logic rather than memory safety, which can lead to critical security failures....
4 weeks ago cs.CR cs.AI
PDF
Tool MEDIUM
Shangkun Che, Silin Du, Ge Gao
The widespread use of Large Language Models (LLMs) in text generation has raised increasing concerns about intellectual property disputes....
4 weeks ago cs.CR cs.CL
PDF
Tool HIGH
Wei Zhao, Zhe Li, Peixin Zhang +1 more
Tool-augmented Large Language Model (LLM) agents have demonstrated impressive capabilities in automating complex, multi-step real-world tasks, yet...
4 weeks ago cs.CR cs.AI
PDF
Tool HIGH
Yihao Zhang, Kai Wang, Jiangrong Wu +7 more
Large Language Models (LLMs) face prominent security risks from jailbreaking, a practice that manipulates models to bypass built-in security...
4 weeks ago cs.CR cs.AI cs.CL
PDF
Tool HIGH
Vu Tuan Truong, Long Bao Le
Large Language Models (LLMs), despite their impressive capabilities across domains, have been shown to be vulnerable to backdoor attacks. Prior...
1 months ago cs.CR cs.AI
PDF
Tool MEDIUM
Hengkai Ye, Zhechang Zhang, Jinyuan Jia +1 more
Large language models (LLMs) increasingly rely on external tools to perform time-sensitive tasks and real-world actions. While tool integration...
Tool MEDIUM
Yen-Shan Chen, Sian-Yao Huang, Cheng-Lin Yang +1 more
As large language models (LLMs) evolve from static chatbots into autonomous agents, the primary vulnerability surface shifts from final outputs to...
1 months ago cs.CR cs.AI cs.CL
PDF
Tool MEDIUM
Yinghan Hou, Zongyou Yang
OpenClaw's ClawHub marketplace hosts over 13,000 community-contributed agent skills, and between 13% and 26% of them contain security vulnerabilities...
1 months ago cs.CR cs.AI
PDF
Tool MEDIUM
Shaofei Huang, Christopher M. Poskitt, Lwin Khin Shar
Cyber-physical systems often contend with incomplete architectural documentation or outdated information resulting from legacy technologies,...
1 months ago cs.CR cs.AI
PDF
Tool MEDIUM
Anes Abdennebi, Nadjia Kara, Laaziz Lahlou +1 more
Modern Security Operations Centers struggle with alert fatigue, fragmented tooling, and limited cross-source event correlation. Challenges that...
1 months ago cs.CR cs.AI
PDF
Tool MEDIUM
Wuyang Zhang, Shichao Pei
Tool-use large language model (LLM) agents are increasingly deployed to support sensitive workflows, relying on tool calls for retrieval, external...
1 months ago cs.CR cs.AI
PDF
Tool MEDIUM
Jiling Zhou, Aisvarya Adeseye, Seppo Virtanen +2 more
Chain-of-Thought (CoT) prompting has been used to enhance the reasoning capability of LLMs. However, its reliability in security-sensitive analytical...
1 months ago cs.CR cs.AI
PDF
Tool HIGH
Zhuowen Yuan, Zhaorun Chen, Zhen Xiang +5 more
Existing research on LLM agent security mainly focuses on prompt injection and unsafe input/output behaviors. However, as agents increasingly rely on...
Tool MEDIUM
Fariha Tanjim Shifat, Hariswar Baburaj, Ce Zhou +2 more
Large language models (LLMs) are increasingly embedded in open-source software (OSS) ecosystems, creating complex interactions among natural language...
1 months ago cs.CR cs.SE
PDF
Tool MEDIUM
Jihoon Jeong
AI models of equivalent capability can exhibit fundamentally different behavioral patterns, yet no standardized instrument exists to measure these...
1 months ago cs.AI cs.CL
PDF
Tool HIGH
Anubhab Sahu, Diptisha Samanta, Reza Soosahabi
System Instructions in Large Language Models (LLMs) are commonly used to enforce safety policies, define agent behavior, and protect sensitive...
1 months ago cs.CR cs.AI
PDF
Tool HIGH
Jingning Xu, Haochen Luo, Chen Liu
Vision-language models (VLMs) are vulnerable to adversarial image perturbations. Existing works based on adversarial training against task-specific...
1 months ago cs.CV cs.MM
PDF
Tool HIGH
Aengus Lynch
Autonomous AI agents are being deployed with filesystem access, email control, and multi-step planning. This thesis contributes to four open problems...
1 months ago cs.LG cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial