AI Security Research

2,529+ academic papers on AI security, attacks, and defenses

Total
2,529
Attack
969
Benchmark
729
Defense
345
Tool
272
Survey
142

Showing 101–120 of 345 papers

Clear filters
Defense MEDIUM

Fail-Closed Alignment for Large Language Models

Zachary Coalson, Beth Sohler, Aiden Gabriel +1 more

We identify a structural weakness in current large language model (LLM) alignment: modern refusal mechanisms are fail-open. While existing approaches...

2 months ago cs.LG cs.CR PDF
Defense MEDIUM

NeST: Neuron Selective Tuning for LLM Safety

Sasha Behrouzi, Lichao Wu, Mohamadreza Rostami +1 more

Safety alignment is essential for the responsible deployment of large language models (LLMs). Yet, existing approaches often rely on heavyweight...

2 months ago cs.CR cs.LG PDF

Track AI security vulnerabilities in real time

Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.

Start 14-Day Free Trial