Attack HIGH
Jiayao Wang, Mohammad Maruf Hasan, Yiping Zhang +5 more
Self-Supervised Learning (SSL) has emerged as a significant paradigm in representation learning thanks to its ability to learn without extensive...
Attack MEDIUM
Shuyi Zhou, Zeen Song, Wenwen Qiang +4 more
Large Language Models remain vulnerable to adversarial prefix attacks (e.g., ``Sure, here is'') despite robust standard safety. We diagnose this...
Attack HIGH
Huw Day, Adrianna Jezierska, Jessica Woodgate
Large Language Models have intensified the scale and strategic manipulation of political discourse on social media, leading to conflict escalation....
2 months ago cs.HC cs.AI
PDF
Attack MEDIUM
Guoxin Shi, Haoyu Wang, Zaihui Yang +2 more
Adversarial behavior plays a central role in aligning large language models with human values. However, existing alignment methods largely rely on...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Duoxun Tang, Dasen Dai, Jiyao Wang +3 more
Video-LLMs are increasingly deployed in safety-critical applications but are vulnerable to Energy-Latency Attacks (ELAs) that exhaust computational...
2 months ago cs.CV cs.AI
PDF
Attack HIGH
Xinyu Huang, Qiang Yang, Leming Shen +2 more
Embodied Large Language Models (LLMs) enable AI agents to interact with the physical world through natural language instructions and actions....
Attack HIGH
Jiayao Wang, Yiping Zhang, Mohammad Maruf Hasan +5 more
Self-supervised diffusion models learn high-quality visual representations via latent space denoising. However, their representation layer poses a...
2 months ago cs.CR cs.LG
PDF
Attack MEDIUM
Martin Odersky, Yaoyu Zhao, Yichen Xu +2 more
AI agents that interact with the real world through tool calls pose fundamental safety challenges: agents might leak private information, cause...
2 months ago cs.AI cs.PL
PDF
Attack HIGH
Oluseyi Olukola, Nick Rahimi
Machine learning based network intrusion detection systems are vulnerable to adversarial attacks that degrade classification performance under both...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Hsin Lin, Yan-Lun Chen, Ren-Hung Hwang +1 more
Backdoor attacks pose a critical threat to the security of deep neural networks, yet existing efforts on universal backdoors often rely on visually...
2 months ago cs.CR cs.CV cs.LG
PDF
Attack HIGH
Yilian Liu, Xiaojun Jia, Guoshun Nan +6 more
Multimodal Large Language Models (MLLMs) have achieved remarkable performance but remain vulnerable to jailbreak attacks that can induce harmful...
2 months ago cs.CV cs.AI cs.CR
PDF
Attack HIGH
Swapnil Parekh
Image captioning models are encoder-decoder architectures trained on large-scale image-text datasets, making them susceptible to adversarial attacks....
2 months ago cs.CV cs.AI
PDF
Attack MEDIUM
Jingyuan Xie, Wenjie Wang, Ji Wu +1 more
Supervised fine-tuning (SFT) is essential for the development of medical large language models (LLMs), yet prior poisoning studies have mainly...
2 months ago cs.CR cs.AI cs.LG
PDF
Attack HIGH
Linxi Jiang, Zhijie Liu, Haotian Luo +1 more
Browser-use agents are widely used for everyday tasks. They enable automated interaction with web pages through structured DOM based interfaces or...
2 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Qianxun Xu, Chenxi Song, Yujun Cai +1 more
Recent advances in text-to-video diffusion models have enabled high-fidelity and temporally coherent videos synthesis. However, current models are...
Attack MEDIUM
Qianxun Xu, Chenxi Song, Yujun Cai +1 more
Recent advances in text-to-video diffusion models have enabled high-fidelity and temporally coherent videos synthesis. However, current models are...
Attack HIGH
Kennedy Edemacu, Mohammad Mahdi Shokri
Retrieval-augmented generation (RAG) has emerged as a powerful paradigm for enhancing multimodal large language models by grounding their responses...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Xun Huang, Simeng Qin, Xiaoshuang Jia +6 more
As Large Language Models (LLMs) are increasingly used, their security risks have drawn increasing attention. Existing research reveals that LLMs are...
2 months ago cs.AI cs.CR
PDF
Attack HIGH
Tian Zhang, Yiwei Xu, Juan Wang +8 more
Large language model (LLM) agents increasingly rely on external tools and retrieval systems to autonomously complete complex tasks. However, this...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Marcus Graves
We introduce Reverse CAPTCHA, an evaluation framework that tests whether large language models follow invisible Unicode-encoded instructions embedded...
2 months ago cs.CR cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial