AI Security Research

2,589+ academic papers on AI security, attacks, and defenses

Total
2,589
Attack
998
Benchmark
740
Defense
355
Tool
276
Survey
147

Showing 881–900 of 1,931 papers

Clear filters
Attack HIGH

TFL: Targeted Bit-Flip Attack on Large Language Model

Jingkai Guo, Chaitali Chakrabarti, Deliang Fan

Large language models (LLMs) are increasingly deployed in safety and security critical applications, raising concerns about their robustness to model...

2 months ago cs.CR cs.CL cs.LG PDF
Defense MEDIUM

Fail-Closed Alignment for Large Language Models

Zachary Coalson, Beth Sohler, Aiden Gabriel +1 more

We identify a structural weakness in current large language model (LLM) alignment: modern refusal mechanisms are fail-open. While existing approaches...

2 months ago cs.LG cs.CR PDF
Defense MEDIUM

NeST: Neuron Selective Tuning for LLM Safety

Sasha Behrouzi, Lichao Wu, Mohamadreza Rostami +1 more

Safety alignment is essential for the responsible deployment of large language models (LLMs). Yet, existing approaches often rely on heavyweight...

2 months ago cs.CR cs.LG PDF
Benchmark MEDIUM

Large-scale online deanonymization with LLMs

Simon Lermen, Daniel Paleka, Joshua Swanson +3 more

We show that large language models can be used to perform at-scale deanonymization. With full Internet access, our agent can re-identify Hacker News...

2 months ago cs.CR cs.AI cs.LG PDF
Attack MEDIUM

Policy Compiler for Secure Agentic Systems

Nils Palumbo, Sarthak Choudhary, Jihye Choi +2 more

LLM-based agents are increasingly being deployed in contexts requiring complex authorization policies: customer service protocols, approval...

2 months ago cs.CR cs.AI cs.MA PDF

Track AI security vulnerabilities in real time

Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.

Start 14-Day Free Trial