Securing AI Agent Execution
Christoph Bühler, Matteo Biagiola, Luca Di Grazia +1 more
Large Language Models (LLMs) have evolved into AI agents that interact with external tools and environments to perform complex tasks. The Model...
2,560+ academic papers on AI security, attacks, and defenses
Showing 1021–1040 of 1,220 papers
Clear filtersChristoph Bühler, Matteo Biagiola, Luca Di Grazia +1 more
Large Language Models (LLMs) have evolved into AI agents that interact with external tools and environments to perform complex tasks. The Model...
Divyanshu Kumar, Nitin Aravind Birur, Tanay Baswa +2 more
Frontier Large Language Models (LLMs) pose unprecedented dual-use risks through the potential proliferation of chemical, biological, radiological,...
Nils Philipp Walter, Chawin Sitawarin, Jamie Hayes +2 more
Large Language Models (LLMs) are increasingly deployed in agentic systems that interact with an external environment; this makes them susceptible to...
Li An, Yujian Liu, Yepeng Liu +3 more
Watermarking has emerged as a promising solution for tracing and authenticating text generated by large language models (LLMs). A common approach to...
Soham Hans, Stacy Marsella, Sophia Hirschmann +1 more
Understanding adversarial behavior in cybersecurity has traditionally relied on high-level intelligence reports and manual interpretation of attack...
Austin Jia, Avaneesh Ramesh, Zain Shamsi +2 more
Retrieval-Augmented Generation (RAG) has emerged as the dominant architectural pattern to operationalize Large Language Model (LLM) usage in Cyber...
Ronghao Ni, Aidan Z. H. Yang, Min-Chien Hsu +5 more
Program analysis tools often produce large volumes of candidate vulnerability reports that require costly manual review, creating a practical...
Alyssa Gerhart, Balaji Iyangar
Adversarial attacks pose a severe risk to AI systems used in healthcare, capable of misleading models into dangerous misclassifications that can...
Daniel Gilkarov, Ran Dubin
Pretrained deep learning model sharing holds tremendous value for researchers and enterprises alike. It allows them to apply deep learning by...
Yulong Chen, Yadong Liu, Jiawen Zhang +3 more
Large Language Models (LLMs), despite advances in safety alignment, remain vulnerable to jailbreak attacks designed to circumvent protective...
Tushar Nayan, Ziqi Zhang, Ruimin Sun
With the increasing deployment of Large Language Models (LLMs) on mobile and edge platforms, securing them against model extraction attacks has...
Hanbin Hong, Ashish Kundu, Ali Payani +2 more
Randomized smoothing has become essential for achieving certified adversarial robustness in machine learning models. However, current methods...
Zhonghao Zhan, Amir Al Sadi, Krinos Li +1 more
In this work, we study security of Model Context Protocol (MCP) agent toolchains and their applications in smart homes. We introduce AegisMCP, a...
Chengcan Wu, Zhixin Zhang, Mingqian Xu +2 more
Large Language Model (LLM)-based Multi-Agent Systems (MAS) have become a popular paradigm of AI applications. However, trustworthiness issues in MAS...
Petar Radanliev
Problem Space: AI Vulnerabilities and Quantum Threats Generative AI vulnerabilities: model inversion, data poisoning, adversarial inputs. Quantum...
Thomas Wang, Haowen Li
As large language models (LLMs) are increasingly integrated into real-world applications, ensuring their safety, robustness, and privacy compliance...
Alexander Nemecek, Zebin Yun, Zahra Rahmani +4 more
As large language models (LLMs) become progressively more embedded in clinical decision-support, documentation, and patient-information systems,...
Marco Alecci, Jordan Samhi, Tegawendé F. Bissyandé +1 more
Mobile apps often embed authentication secrets, such as API keys, tokens, and client IDs, to integrate with cloud services. However, developers often...
Giovanni De Muri, Mark Vero, Robin Staab +1 more
LLMs are often used by downstream users as teacher models for knowledge distillation, compressing their capabilities into memory-efficient models....
Oleksandr Adamov, Anders Carlsson
This paper explores the challenges of cyberattack attribution, specifically APTs, applying the case study approach for the WhisperGate cyber...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial