AgentTrust: Runtime Safety Evaluation and Interception for AI Agent Tool Use
Chenglin Yang
Modern AI agents execute real-world side effects through tool calls such as file operations, shell commands, HTTP requests, and database queries. A...
2,529+ academic papers on AI security, attacks, and defenses
Showing 61–66 of 66 papers
Clear filtersChenglin Yang
Modern AI agents execute real-world side effects through tool calls such as file operations, shell commands, HTTP requests, and database queries. A...
Zheng Fang, Xiaosen Wang, Shenyi Zhang +2 more
Jailbreak attacks on audio language models (ALMs) optimize audio perturbations to elicit unsafe generations, and they typically update the entire...
Jan Dolejš, Martin Jureček, Róbert Lórencz
Modern malware detection pipelines rely on continuous data ingestion and machine learning to counter the high volume of novel threats. This work...
Kaifeng He, Xiaojun Zhang, Peiliang Cai +7 more
Large language models (LLMs) frequently generate defective outputs in code generation tasks, ranging from logical bugs to security vulnerabilities....
Hanum Ko, Sangheum Yeon, Jong Hwan Ko +1 more
As DRAM scales in density and adopts 3D integration, raw fault rates increase and multi-bit errors are no longer rare. Such errors can severely...
Zekun Fei, Zihao Wang, Weijie Liu +4 more
Mixture-of-Experts (MoE) architectures have emerged as a leading paradigm for scaling large language models through sparse, routing-based...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial