Attack MEDIUM
K. J. Kevin Feng, Tae Soo Kim, Rock Yuren Pang +3 more
AI agents that take actions in their environment autonomously over extended time horizons require robust governance interventions to curb their...
5 months ago cs.CY cs.AI
PDF
Attack HIGH
Haoyu Shen, Weimin Lyu, Haotian Xu +1 more
Vision-Language Models (VLMs) have achieved impressive progress in multimodal text generation, yet their rapid adoption raises increasing concerns...
5 months ago cs.CR cs.AI
PDF
Attack HIGH
Mohammad M Maheri, Xavier Cadet, Peter Chin +1 more
Approximate machine unlearning aims to efficiently remove the influence of specific data points from a trained model, offering a practical...
5 months ago cs.LG cs.AI cs.CR
PDF
Attack MEDIUM
Tong Wu, Weibin Wu, Zibin Zheng
Equipped with various tools and knowledge, GPTs, one kind of customized AI agents based on OpenAI's large language models, have illustrated great...
5 months ago cs.CR cs.SE
PDF
Attack MEDIUM
Zeng Wang, Minghao Shao, Akashdeep Saha +4 more
Graph neural networks (GNNs) have shown promise in hardware security by learning structural motifs from netlist graphs. However, this reliance on...
5 months ago cs.CR cs.AI
PDF
Attack HIGH
Richard J. Young
Large Language Model (LLM) safety guardrail models have emerged as a primary defense mechanism against harmful content generation, yet their...
Attack HIGH
Tianyu Zhang, Zihang Xi, Jingyu Hua +1 more
In the realm of black-box jailbreak attacks on large language models (LLMs), the feasibility of constructing a narrow safety proxy, a lightweight...
5 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Herman Errico, Jiquan Ngiam, Shanita Sojan
The Model Context Protocol (MCP) replaces static, developer-controlled API integrations with more dynamic, user-driven agent systems, which also...
Attack HIGH
Kaiyuan Zhang, Mark Tenenholtz, Kyle Polley +3 more
The integration of artificial intelligence (AI) agents into web browsers introduces security challenges that go beyond traditional web application...
5 months ago cs.LG cs.AI cs.CR
PDF
Attack HIGH
Jakub Hoscilowicz, Artur Janicki
We introduce the Adversarial Confusion Attack, a new class of threats against multimodal large language models (MLLMs). Unlike jailbreaks or targeted...
Attack MEDIUM
Sidahmed Benabderrahmane, James Cheney, Talal Rahwan
Advanced Persistent Threats (APTs) pose a significant challenge in cybersecurity due to their stealthy and long-term nature. Modern supervised...
5 months ago cs.LG cs.AI cs.CR
PDF
Attack HIGH
Sen Nie, Jie Zhang, Jianxin Yan +2 more
Adversarial attacks have evolved from simply disrupting predictions on conventional task-specific models to the more complex goal of manipulating...
Attack MEDIUM
Steven Peh
Large Language Models (LLMs) remain vulnerable to prompt injection attacks, representing the most significant security threat in production...
5 months ago cs.CR cs.AI
PDF
Attack HIGH
Yingjia Shang, Yi Liu, Huimin Wang +4 more
With the rapid advancement of retrieval-augmented vision-language models, multimodal medical retrieval-augmented generation (MMed-RAG) systems are...
5 months ago cs.CR cs.AI cs.LG
PDF
Attack HIGH
Md Akil Raihan Iftee, Syed Md. Ahnaf Hasan, Amin Ahsan Ali +3 more
Test-time personalization in federated learning enables models at clients to adjust online to local domain shifts, enhancing robustness and...
5 months ago cs.CR cs.CV
PDF
Attack HIGH
Xurui Li, Kaisong Song, Rui Zhu +2 more
Large Language Models (LLMs) have developed rapidly in web services, delivering unprecedented capabilities while amplifying societal risks. Existing...
5 months ago cs.CR cs.AI
PDF
Attack HIGH
Yixin Wu, Rui Wen, Chi Cui +2 more
Inference attacks have been widely studied and offer a systematic risk assessment of ML services; however, their implementation and the attack...
5 months ago cs.CR cs.AI
PDF
Attack HIGH
Ryan Wong, Hosea David Yu Fei Ng, Dhananjai Sharma +2 more
Large Language Models (LLMs) remain susceptible to jailbreak exploits that bypass safety filters and induce harmful or unethical behavior. This work...
5 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Adarsh Kumarappan, Ayushi Mehrotra
The SmoothLLM defense provides a certification guarantee against jailbreaking attacks, but it relies on a strict "k-unstable" assumption that rarely...
5 months ago cs.LG cs.AI
PDF
Attack HIGH
Adarsh Kumarappan, Ananya Mujoo
Multi-turn conversational attacks, which leverage psychological principles like Foot-in-the-Door (FITD), where a small initial request paves the way...
5 months ago cs.LG cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial