Tool MEDIUM
Jiangrong Wu, Zitong Yao, Yuhong Nan +1 more
Tool-augmented LLM agents increasingly rely on multi-step, multi-tool workflows to complete real tasks. This design expands the attack surface,...
1 weeks ago cs.SE cs.CR
PDF
Attack MEDIUM
Xiangkui Cao, Jie Zhang, Meina Kan +2 more
Large Vision-Language Models (LVLMs) have shown remarkable potential across a wide array of vision-language tasks, leading to their adoption in...
Benchmark MEDIUM
Ninghui Li, Kaiyuan Zhang, Kyle Polley +1 more
This article, a lightly adapted version of Perplexity's response to NIST/CAISI Request for Information 2025-0035, details our observations and...
1 weeks ago cs.LG cs.AI cs.CR
PDF
Attack MEDIUM
Haodong Zhao, Jinming Hu, Yijie Bai +6 more
Federated Language Model (FedLM) allows a collaborative learning without sharing raw data, yet it introduces a critical vulnerability, as every...
Benchmark MEDIUM
Junjie Chu, Yiting Qu, Ye Leng +4 more
Large Language Models (LLMs) are increasingly trained to align with human values, primarily focusing on task level, i.e., refusing to execute...
1 weeks ago cs.CR cs.AI
PDF
Tool MEDIUM
Frank Li
Tool-augmented LLM agents introduce security risks that extend beyond user-input filtering, including indirect prompt injection through fetched...
Defense MEDIUM
Xinhao Deng, Yixiang Zhang, Jiaqing Wu +15 more
Autonomous Large Language Model (LLM) agents, exemplified by OpenClaw, demonstrate remarkable capabilities in executing complex, long-horizon tasks....
1 weeks ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Qizhi Chen, Chao Qi, Yihong Huang +5 more
Graph-based Retrieval-Augmented Generation (GraphRAG) constructs the Knowledge Graph (KG) from external databases to enhance the timeliness and...
1 weeks ago cs.LG cs.AI cs.CR
PDF
Defense MEDIUM
Zhiyu Xue, Zimo Qi, Guangliang Liu +2 more
Safety alignment aims to ensure that large language models (LLMs) refuse harmful requests by post-training on harmful queries paired with refusal...
Benchmark MEDIUM
Marc Damie, Murat Bilgehan Ertan, Domenico Essoussi +3 more
With their increasing capabilities, Large Language Models (LLMs) are now used across many industries. They have become useful tools for software...
2 weeks ago cs.LG cs.CL cs.CR
PDF
Tool MEDIUM
Zixun Xiong, Gaoyi Wu, Lingfeng Yao +3 more
Communication topology is a critical factor in the utility and safety of LLM-based multi-agent systems (LLM-MAS), making it a high-value intellectual...
2 weeks ago cs.CR cs.AI
PDF
Tool MEDIUM
Panagiotis Georgios Pennas, Konstantinos Papaioannou, Marco Guarnieri +1 more
Large Language Models (LLMs) rely on optimizations like Automatic Prefix Caching (APC) to accelerate inference. APC works by reusing previously...
2 weeks ago cs.CR cs.DC cs.LG
PDF
Benchmark MEDIUM
Chuan Guo, Juan Felipe Ceron Uribe, Sicheng Zhu +10 more
Instruction hierarchy (IH) defines how LLMs prioritize system, developer, user, and tool instructions under conflict, providing a concrete,...
2 weeks ago cs.AI cs.CL cs.CR
PDF
Benchmark MEDIUM
Manit Baser, Alperen Yildiz, Dinil Mon Divakaran +1 more
The static knowledge representations of large language models (LLMs) inevitably become outdated or incorrect over time. While model-editing...
Tool MEDIUM
Zhengyang Shan, Jiayun Xin, Yue Zhang +1 more
Code agents powered by large language models can execute shell commands on behalf of users, introducing severe security vulnerabilities. This paper...
Tool MEDIUM
Shriti Priya, Julian James Stephen, Arjun Natarajan
Enterprises and organizations today increasingly deploy in-house, cloud based applications and APIs for internal operations or external customers....
Attack MEDIUM
Pratyay Kumar, Miguel Antonio Guirao Aguilera, Srikathyayani Srikanteswara +2 more
Model Context Protocol (MCP) servers have rapidly emerged over the past year as a widely adopted way to enable Large Language Model (LLM) agents to...
2 weeks ago cs.CR cs.AI
PDF
Attack MEDIUM
Meenatchi Sundaram Muthu Selva Annamalai, Emiliano De Cristofaro, Peter Kairouz
As AI assistants become widely used, privacy-aware platforms like Anthropic's Clio have been introduced to generate insights from real-world AI use....
Attack MEDIUM
Jia Hu, Youcheng Sun, Pierre Olivier
Software compartmentalization breaks down an application into compartments isolated from each other: an attacker taking over a compartment will be...
Attack MEDIUM
Ali Raza, Gurang Gupta, Nikolay Matyunin +1 more
Warning: This article includes red-teaming experiments, which contain examples of compromised LLM responses that may be offensive or upsetting. Large...
2 weeks ago cs.CR cs.AI cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial