Tool MEDIUM
Joel Rorseth, Parke Godfrey, Lukasz Golab +2 more
This paper demonstrates RUBEN, an interactive tool for discovering minimal rules to explain the outputs of retrieval-augmented large language models...
Benchmark MEDIUM
Pedro Conde, Henrique Branquinho, Valerio Mazzone +3 more
AI pentesting agents are increasingly credible as offensive security systems, but current benchmarks still provide limited guidance on which will...
Yesterday cs.AI cs.CR
PDF
Benchmark MEDIUM
Saba Pourhanifeh, AbdulAziz AbdulGhaffar, Ashraf Matrawy
Large Language Models(LLMs) are increasingly explored for cybersecurity applications such as vulnerability detection. In the domain of threat...
Yesterday cs.CR cs.AI
PDF
Survey MEDIUM
Johann Knechtel, Ozgur Sinanoglu, Ramesh Karri
The integration of Large Language Models (LLMs) into Electronic Design Automation (EDA) and hardware security is rapidly reshaping the semiconductor...
Yesterday cs.CR cs.AR cs.LG
PDF
Defense MEDIUM
Krishak Aneja, Manas Mittal, Anmol Goel +2 more
Fine-tuning Large Language Models (LLMs) on benign narrow data can sometimes induce broad harmful behaviors, a vulnerability termed emergent...
Yesterday cs.CL cs.AI
PDF
Benchmark MEDIUM
Qinghua Mao, Xi Lin, Jinze Gu +3 more
Large language models (LLMs) increasingly rely on knowledge editing to support knowledge-intensive reasoning, but this flexibility also introduces...
Yesterday cs.AI cs.CR
PDF
Benchmark MEDIUM
Xia Hu, Zhenrui Yue, Brian Potetz +4 more
As current Multimodal Large Language Models rapidly saturate canonical visual reasoning benchmarks, a key question emerges: do these strong scores...
Yesterday cs.CV cs.AI
PDF
Attack MEDIUM
Ben Kereopa-Yorke, Guillermo Diaz, Holly Wright +3 more
We define Oracle Poisoning, an attack class in which an adversary corrupts a structured knowledge graph that AI agents query at runtime via tool-use...
2 days ago cs.CR cs.AI
PDF
Attack MEDIUM
Li Lixing
Modern large language models (LLMs) rely on system prompts to establish behavioral constraints and safety rules. Standard causal self-attention...
Benchmark MEDIUM
Huy Hoang Ha, Benoit Favre, Francois Portet
Large language models (LLMs) have saturated standard medical benchmarks that test factual recall, yet their ability to perform higher-order...
2 days ago cs.CL cs.AI
PDF
Benchmark MEDIUM
Jingshen Zhang, Bo Wang, Yanlin Fu +4 more
In this paper, we study an emergent self-debiasing mechanisms against stereotypical content in Large Language Models (LLMs). Unlike traditional...
Tool MEDIUM
Michael A. Riegler, Inga Strümke
We present swarm-attack, an open-source adversarial testing framework in which multiple lightweight LLM agents coordinate through shared memory,...
2 days ago cs.CR cs.AI cs.LG
PDF
Benchmark MEDIUM
Yilin Zhang, Yingkai Hua, Chunyu Wei +2 more
Vision-language model (VLM) based web agents demonstrate impressive autonomous GUI interaction but remain vulnerable to deceptive interface elements....
2 days ago cs.AI cs.CR
PDF
Attack MEDIUM
Isaac David, Arthur Gervais
Security updates create a short but important window in which defenders and attackers can compare vulnerable and patched software. Yet in many...
5 days ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Di Lu, Bo Zhang, Xiyuan Li +5 more
Self-hosted computer-use agents (SHCUAs), such as OpenClaw, combine natural-language interaction with direct access to host-side resources, including...
Tool MEDIUM
Chengjie Wang, Jingzheng Wu, Xiang Ling +2 more
Large language models (LLMs) are now largely involved in software development workflows, and the code they generate routinely includes third-party...
5 days ago cs.SE cs.AI
PDF
Defense MEDIUM
Leo Linqian Gan, Jeffery Wu, Longyuan Ge +6 more
Autonomous LLM agents face a critical security risk known as workflow hijacking, where attackers subtly alter tool and skill invocations. Existing...
Defense MEDIUM
Guoxin Lu, Letian Sha, Qing Wang +4 more
The safety alignment of Large Language Models (LLMs) remains vulnerable to Harmful Fine-tuning (HFT). While existing defenses impose constraints on...
5 days ago cs.CR cs.AI cs.CL
PDF
Benchmark MEDIUM
Qinfeng Li, Yuntai Bao, Jianghui Hu +5 more
LLM agents rely on prompts to implement task-specific capabilities based on foundation LLMs, making agent prompts valuable intellectual property....
5 days ago cs.CR cs.AI
PDF
Survey MEDIUM
Bonan Ruan, Yeqi Fu, Chuqi Zhang +3 more
GitHub Continuous Integration (CI) workflows increasingly integrate Large Language Models (LLMs) to automate review, triage, content generation, and...
5 days ago cs.CR cs.SE
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial