Attack HIGH
Hicham Eddoubi, Umar Faruk Abdullahi, Fadi Hassan
Large Language Models (LLMs) have seen widespread adoption across multiple domains, creating an urgent need for robust safety alignment mechanisms....
Attack MEDIUM
Matthew P. Lad, Louisa Conwill, Megan Levis Scheirer
With the rapid growth of Large Language Models (LLMs), criticism of their societal impact has also grown. Work in Responsible AI (RAI) has focused on...
Attack LOW
Blake Bullwinkel, Giorgio Severi, Keegan Hines +3 more
Detecting whether a model has been poisoned is a longstanding problem in AI security. In this work, we present a practical scanner for identifying...
3 months ago cs.CR cs.AI
PDF
Attack HIGH
Xiaozuo Shen, Yifei Cai, Rui Ning +2 more
The widespread adoption of Vision Transformers (ViTs) elevates supply-chain risk on third-party model hubs, where an adversary can implant backdoors...
Attack HIGH
Nirab Hossain, Pablo Moriano
Modern vehicles rely on electronic control units (ECUs) interconnected through the Controller Area Network (CAN), making in-vehicle communication a...
3 months ago cs.CR cs.AI cs.LG
PDF
Attack MEDIUM
Patrick Cooper, Alireza Nadali, Ashutosh Trivedi +1 more
Large language models (LLMs) are known to exhibit brittle behavior under adversarial prompts and jailbreak attacks, even after extensive alignment...
3 months ago cs.CL cs.AI cs.CR
PDF
Attack HIGH
Samuel Nellessen, Tal Kachman
The evolution of large language models into autonomous agents introduces adversarial failures that exploit legitimate tool privileges, transforming...
3 months ago cs.LG cs.AI cs.CR
PDF
Attack MEDIUM
Ching-Yun Ko, Pin-Yu Chen
Modern artificial intelligence (AI) models are deployed on inference engines to optimize runtime efficiency and resource allocation, particularly for...
3 months ago cs.LG cs.CL cs.PL
PDF
Attack HIGH
Pengfei He, Ash Fox, Lesly Miculicich +7 more
Large language models (LLMs) have shown promise in assisting cybersecurity tasks, yet existing approaches struggle with automatic vulnerability...
3 months ago cs.LG cs.CR
PDF
Attack HIGH
Jiayao Wang, Yang Song, Zhendong Zhao +5 more
Federated self-supervised learning (FSSL) enables collaborative training of self-supervised representation models without sharing raw unlabeled data....
Attack HIGH
Mingrui Liu, Sixiao Zhang, Cheng Long +1 more
Large Language Models (LLMs) are increasingly vulnerable to Prompt Injection (PI) attacks, where adversarial instructions hidden within retrieved...
3 months ago cs.CR cs.AI cs.LG
PDF
Attack LOW
Pengyu Li, Lingling Zhang, Zhitao Gao +5 more
While Large Language Models (LLMs) have achieved remarkable capabilities, they unintentionally memorize sensitive data, posing critical privacy and...
3 months ago cs.LG cs.CL
PDF
Attack HIGH
Seyed Mohammad Hadi Hosseini, Amir Najafi, Mahdieh Soleymani Baghshah
Bandit algorithms have recently emerged as a powerful tool for evaluating machine learning models, including generative image models and large...
3 months ago cs.LG cs.AI
PDF
Attack HIGH
Haobo Wang, Weiqi Luo, Xiaojun Jia +1 more
Large vision-language models (VLMs) are vulnerable to transfer-based adversarial perturbations, enabling attackers to optimize on surrogate models...
Attack HIGH
Xiaoyu Wen, Zhida He, Han Qi +7 more
Ensuring robust safety alignment is crucial for Large Language Models (LLMs), yet existing defenses often lag behind evolving adversarial attacks due...
3 months ago cs.AI cs.CL cs.LG
PDF
Attack MEDIUM
Poushali Sengupta, Shashi Raj Pandey, Sabita Maharjan +1 more
Large language models (LLMs) generate outputs by utilizing extensive context, which often includes redundant information from prompts, retrieved...
3 months ago cs.CL cs.AI stat.ML
PDF
Attack MEDIUM
Eliron Rahimi, Elad Hirshel, Rom Himelstein +3 more
Diffusion language models (DLMs) have recently emerged as a promising alternative to autoregressive (AR) models, offering parallel decoding and...
3 months ago cs.LG cs.AI
PDF
Attack HIGH
Ziyue Wang, Jiangshan Yu, Kaihua Qin +3 more
Decentralized Finance (DeFi) has turned blockchains into financial infrastructure, allowing anyone to trade, lend, and build protocols without...
3 months ago cs.CR cs.AI
PDF
Attack HIGH
Terry Yue Zhuo, Yangruibo Ding, Wenbo Guo +1 more
For over a decade, cybersecurity has relied on human labor scarcity to limit attackers to high-value targets manually or generic automated attacks at...
3 months ago cs.CR cs.AI cs.CY
PDF
Attack MEDIUM
Xinyi Hou, Shenao Wang, Yifan Zhang +4 more
Agentic AI systems built around large language models (LLMs) are moving away from closed, single-model frameworks and toward open ecosystems that...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial