ClawSafety: "Safe" LLMs, Unsafe Agents
Bowen Wei, Yunbei Zhang, Jinhao Pan +5 more
Personal AI agents like OpenClaw run with elevated privileges on users' local machines, where a single successful prompt injection can leak...
2,583+ academic papers on AI security, attacks, and defenses
Showing 441–460 of 1,933 papers
Clear filtersBowen Wei, Yunbei Zhang, Jinhao Pan +5 more
Personal AI agents like OpenClaw run with elevated privileges on users' local machines, where a single successful prompt injection can leak...
Tiankai Yang, Jiate Li, Yi Nian +5 more
LLM-based agents increasingly operate across repeated sessions, maintaining task states to ensure continuity. In many deployments, a single agent...
Manoj Parmar
World models -- learned internal simulators of environment dynamics -- are rapidly becoming foundational to autonomous decision-making in robotics,...
Yanting Wang, Wei Zou, Runpeng Geng +1 more
Large language models (LLMs) and their applications, such as agents, are highly vulnerable to prompt injection attacks. State-of-the-art prompt...
Kıvanç Kuzey Dikici, Serdar Kara, Semih Çağlar +2 more
As Large Language Models (LLMs) for code increasingly utilize massive, often non-permissively licensed datasets, evaluating data contamination...
Saeid Jamshidi, Negar Shahabi, Foutse Khomh +2 more
Software-Defined Networking (SDN) is increasingly adopted to secure Internet-of-Things (IoT) networks due to its centralized control and programmable...
Anubhab Sahu, Diptisha Samanta, Reza Soosahabi
System Instructions in Large Language Models (LLMs) are commonly used to enforce safety policies, define agent behavior, and protect sensitive...
Jingning Xu, Haochen Luo, Chen Liu
Vision-language models (VLMs) are vulnerable to adversarial image perturbations. Existing works based on adversarial training against task-specific...
Xiaoqi Li, Shipeng Ye, Wenkai Li +1 more
Smart contract vulnerabilities can cause substantial financial losses due to the immutability of code after deployment. While existing tools detect...
Yishun Wang, Wenkai Li, Xiaoqi Li +3 more
Smart contracts are self-executing programs that manage financial transactions on blockchain networks. Developers commonly rely on third-party code...
Jiaqing Li, Zhibo Zhang, Shide Zhou +3 more
Model merging has emerged as a powerful technique for combining specialized capabilities from multiple fine-tuned LLMs without additional training...
Yao Qin, Yangyang Yan, Jinhua Pang +1 more
The integration of Large Language Models (LLMs) into life sciences has catalyzed the development of "AI Scientists." However, translating these...
Aengus Lynch
Autonomous AI agents are being deployed with filesystem access, email control, and multi-step planning. This thesis contributes to four open problems...
Yanting Wang, Jinyuan Jia
Random subspace method has wide security applications such as providing certified defenses against adversarial and backdoor attacks, and building...
Chong Xiang, Drew Zagieboylo, Shaona Ghosh +5 more
AI agents, predominantly powered by large language models (LLMs), are vulnerable to indirect prompt injection, in which malicious instructions...
Quanyan Zhu, Zhengye Han
This paper introduces a performative scenario optimization framework for decision-dependent chance-constrained problems. Unlike classical stochastic...
Xiao Qian, Shangjia Dong
Accurate prediction of evacuation behavior is critical for disaster preparedness, yet models trained in one region often fail elsewhere. Using a...
Edoardo Allegrini, Edoardo Di Paolo, Angelo Spognardi +1 more
BotVerse is a scalable, event-driven framework for high-fidelity social simulation using LLM-based agents. It addresses the ethical risks of studying...
Meiwen Ding, Song Xia, Chenqi Kong +1 more
Although multimodal large language models (MLLMs) are increasingly deployed in real-world applications, their instruction-following behavior leaves...
Yubo Cui, Xianchao Guan, Zijun Xiong +1 more
Pre-trained vision-language models (VLMs) exhibit strong zero-shot generalization but remain vulnerable to adversarial perturbations. Existing...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial