The Autonomy Tax: Defense Training Breaks LLM Agents
Shawn Li, Yue Zhao
Large language model (LLM) agents increasingly rely on external tools (file operations, API calls, database transactions) to autonomously complete...
2,583+ academic papers on AI security, attacks, and defenses
Showing 541–560 of 1,930 papers
Clear filtersShawn Li, Yue Zhao
Large language model (LLM) agents increasingly rely on external tools (file operations, API calls, database transactions) to autonomously complete...
Toan Tran, Olivera Kotevska, Li Xiong
Membership inference attacks (MIAs), which enable adversaries to determine whether specific data points were part of a model's training dataset, have...
Rohan Siva, Kai Cheung, Lichi Li +1 more
Modern machine learning systems rely on complex data engineering workflows to extract, transform, and load (ELT) data into production pipelines....
Zou Qiang
Large language models (LLMs) demonstrate strong generative capabilities but remain vulnerable to hallucination and unreliable reasoning under...
Aravind Krishnan, Karolina Stańczak, Dietrich Klakow
As Spoken Language Models (SLMs) integrate speech and text modalities, they inherit the safety vulnerabilities of their LLM backbone and an expanded...
Sheng Liu, Panos Papadimitratos
FL has emerged as a transformative paradigm for ITS, notably camera-based Road Condition Classification (RCC). However, by enabling collaboration,...
Carlos Hinojosa, Clemens Grange, Bernard Ghanem
Vision-language models (VLMs) are increasingly deployed in real-world and embodied settings where safety decisions depend on visual context. However,...
Pranay Anchuri, Matteo Campanelli, Paul Cesaretti +4 more
When large AI models are deployed as cloud-based services, clients have no guarantee that responses are correct or were produced by the intended...
Zikang Ding, Junhao Li, Suling Wu +3 more
Model watermarking utilizes internal representations to protect the ownership of large language models (LLMs). However, these features inevitably...
Dimitris Mitropoulos, Nikolaos Alexopoulos, Georgios Alexopoulos +1 more
Security code reviews increasingly rely on systems integrating Large Language Models (LLMs), ranging from interactive assistants to autonomous agents...
Mohammadhossein Homaei, Iman Khazrak, Rubén Molano +2 more
Industrial Cyber-Physical Systems (ICPS) face growing threats from cyber-attacks that exploit sensor and control vulnerabilities. Digital Twin (DT)...
Jiahao Zhang, Yilong Wang, Suhang Wang
Graph neural networks (GNNs) are widely used for learning from graph-structured data in domains such as social networks, recommender systems, and...
Md Takrim Ul Alam, Akif Islam, Mohd Ruhul Ameen +2 more
Large language models (LLMs) deployed behind APIs and retrieval-augmented generation (RAG) stacks are vulnerable to prompt injection attacks that may...
Alvin Rajkomar, Pavan Sudarshan, Angela Lai +1 more
Background: Clinical trials rely on transparent inclusion criteria to ensure generalizability. In contrast, benchmarks validating health-related...
Saket Sanjeev Chaturvedi, Joshua Bergerson, Tanwi Mallick
As large language models (LLMs) evolve into autonomous "AI scientists," they promise transformative advances but introduce novel vulnerabilities,...
Xavier Cadet, Aditya Vikram Singh, Harsh Mamania +6 more
Investigating cybersecurity incidents requires collecting and analyzing evidence from multiple log sources, including intrusion detection alerts,...
Xavier Cadet, Aditya Vikram Singh, Harsh Mamania +6 more
Investigating cybersecurity incidents requires collecting and analyzing evidence from multiple log sources, including intrusion detection alerts,...
Iakovos-Christos Zarkadis, Christos Douligeris
Supervised detection of network attacks has always been a critical part of network intrusion detection systems (NIDS). Nowadays, in a pivotal time...
Haocheng Li, Juepeng Zheng, Shuangxi Miao +4 more
Multimodal remote sensing semantic segmentation enhances scene interpretation by exploiting complementary physical cues from heterogeneous data....
Wanjun Du, Zifeng Yuan, Tingting Chen +3 more
Existing vision-language models (VLMs) have demonstrated impressive performance in reasoning-based segmentation. However, current benchmarks are...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial