AI Security Research

2,560+ academic papers on AI security, attacks, and defenses

Total
2,560
Attack
982
Benchmark
736
Defense
350
Tool
275
Survey
144

Showing 321–340 of 725 papers

Clear filters
Attack HIGH

Boundary Point Jailbreaking of Black-Box LLMs

Xander Davies, Giorgi Giglemiani, Edmund Lau +3 more

Frontier LLMs are safeguarded against attempts to extract harmful information via adversarial prompts known as "jailbreaks". Recently, defenders have...

2 months ago cs.LG PDF

Track AI security vulnerabilities in real time

Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.

Start 14-Day Free Trial