Tool HIGH
Md. Mehedi Hasan, Ziaur Rahman, Rafid Mostafiz +1 more
This paper presents a real-time modular defense system named Sentra-Guard. The system detects and mitigates jailbreak and prompt injection attacks...
5 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Julia Bazinska, Max Mathys, Francesco Casucci +4 more
AI agents powered by large language models (LLMs) are being deployed at scale, yet we lack a systematic understanding of how the choice of backbone...
5 months ago cs.CR cs.AI cs.LG
PDF
Attack HIGH
Dongyi Liu, Jiangtong Li, Dawei Cheng +1 more
Graph Neural Networks(GNNs) are vulnerable to backdoor attacks, where adversaries implant malicious triggers to manipulate model predictions....
5 months ago cs.CR cs.LG
PDF
Benchmark MEDIUM
Hao Zheng, Zirui Pang, Ling li +5 more
Advances in Multimodal Large Language Models (MLLMs) intensify concerns about data privacy, making Machine Unlearning (MU), the selective removal of...
5 months ago cs.AI cs.CL
PDF
Attack MEDIUM
Devon A. Kelly, Christiana Chamon
Wide-bandgap (WBG) technologies offer unprecedented improvements in power system efficiency, size, and performance, but also introduce unique sensor...
5 months ago cs.CR cs.LG eess.SY
PDF
Attack HIGH
Anum Paracha, Junaid Arshad, Mohamed Ben Farah +1 more
Data poisoning attacks are a potential threat to machine learning (ML) models, aiming to manipulate training datasets to disrupt their performance....
5 months ago cs.CR cs.LG
PDF
Benchmark LOW
Wenxuan Bao, Ruxi Deng, Jingrui He
Pretrained vision-language models such as CLIP achieve strong zero-shot generalization but remain vulnerable to distribution shifts caused by input...
5 months ago cs.CV cs.LG
PDF
Attack HIGH
Pavlos Ntais
Large language models (LLMs) remain vulnerable to sophisticated prompt engineering attacks that exploit contextual framing to bypass safety...
5 months ago cs.CR cs.AI cs.CL
PDF
Attack MEDIUM
Sarah Ball, Niki Hasrati, Alexander Robey +4 more
Discrete optimization-based jailbreaking attacks on large language models aim to generate short, nonsensical suffixes that, when appended onto input...
5 months ago cs.CL cs.AI
PDF
Attack HIGH
Havva Alizadeh Noughabi, Julien Serbanescu, Fattane Zarrinkalam +1 more
Despite recent advances, Large Language Models remain vulnerable to jailbreak attacks that bypass alignment safeguards and elicit harmful outputs....
5 months ago cs.CL cs.AI
PDF
Attack HIGH
Kieu Dang, Phung Lai, NhatHai Phan +3 more
Large language models (LLMs) demonstrate remarkable capabilities across various tasks. However, their deployment introduces significant risks related...
Attack HIGH
Mahavir Dabas, Tran Huynh, Nikhil Reddy Billa +8 more
Large language models remain vulnerable to jailbreak attacks that bypass safety guardrails to elicit harmful outputs. Defending against novel...
Tool MEDIUM
Adetayo Adebimpe, Helmut Neukirchen, Thomas Welsh
Honeypots are decoy systems used for gathering valuable threat intelligence or diverting attackers away from production systems. Maximising attacker...
5 months ago cs.CR cs.CL cs.LG
PDF
Benchmark MEDIUM
Mojtaba Eshghie, Gabriele Morello, Matteo Lauretano +2 more
Smart contract vulnerabilities cost billions of dollars annually, yet existing automated analysis tools fail to generate deployable defenses. We...
5 months ago cs.CR cs.SE
PDF
Defense MEDIUM
Lu Liu, Wuqi Zhang, Lili Wei +3 more
Decentralized Finance (DeFi) smart contracts manage billions of dollars, making them a prime target for exploits. Price manipulation vulnerabilities,...
5 months ago cs.CR cs.SE
PDF
Benchmark MEDIUM
Christoph Bühler, Matteo Biagiola, Luca Di Grazia +1 more
Large Language Models (LLMs) have evolved into AI agents that interact with external tools and environments to perform complex tasks. The Model...
5 months ago cs.CR cs.AI cs.SE
PDF
Attack HIGH
Xingwei Zhong, Kar Wai Fok, Vrizlynn L. L. Thing
Multimodal large language models (MLLMs) comprise of both visual and textual modalities to process vision language tasks. However, MLLMs are...
Attack HIGH
Mingrui Liu, Sixiao Zhang, Cheng Long +1 more
As Large Language Models (LLMs) become integral to computing infrastructure, safety alignment serves as the primary security control preventing the...
Attack HIGH
Yukun Jiang, Mingjie Li, Michael Backes +1 more
Despite their superior performance on a wide range of domains, large language models (LLMs) remain vulnerable to misuse for generating harmful...
Benchmark MEDIUM
Divyanshu Kumar, Nitin Aravind Birur, Tanay Baswa +2 more
Frontier Large Language Models (LLMs) pose unprecedented dual-use risks through the potential proliferation of chemical, biological, radiological,...
5 months ago cs.CR cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial