Attack HIGH
Santhosh KumarRavindran
The rapid adoption of large language models (LLMs) in enterprise systems exposes vulnerabilities to prompt injection attacks, strategic deception,...
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Buyun Liang, Liangzu Peng, Jinqi Luo +3 more
Large Language Models (LLMs) are increasingly deployed in high-risk domains. However, state-of-the-art LLMs often exhibit hallucinations, raising...
7 months ago cs.CL cs.AI cs.CR
PDF
Attack HIGH
Yu Cui, Sicheng Pan, Yifei Liu +2 more
Large language models (LLMs) have been widely deployed in Conversational AIs (CAIs), while exposing privacy and security threats. Recent research...
Attack HIGH
Yanjie Li, Yiming Cao, Dong Wang +1 more
Multimodal agents built on large vision-language models (LVLMs) are increasingly deployed in open-world settings but remain highly vulnerable to...
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Xiangxiang Chen, Peixin Zhang, Jun Sun +2 more
Model quantization is a popular technique for deploying deep learning models on resource-constrained environments. However, it may also introduce...
7 months ago cs.CR cs.AI cs.LG
PDF
Attack HIGH
Yulin Chen, Haoran Li, Yuan Sui +2 more
With the development of technology, large language models (LLMs) have dominated the downstream natural language processing (NLP) tasks. However,...
Attack HIGH
Rabeya Amin Jhuma, Mostafa Mohaimen Akand Faisal
This study explored how in-context learning (ICL) in large language models can be disrupted by data poisoning attacks in the setting of public health...
7 months ago cs.LG cs.CL cs.CR
PDF
Attack HIGH
Maraz Mia, Mir Mehedi A. Pritom
Explainable Artificial Intelligence (XAI) has aided machine learning (ML) researchers with the power of scrutinizing the decisions of the black-box...
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Javad Rafiei Asl, Sidhant Narula, Mohammad Ghasemigol +2 more
Large Language Models (LLMs) have revolutionized natural language processing but remain vulnerable to jailbreak attacks, especially multi-turn...
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Sanket Badhe
We present LegalSim, a modular multi-agent simulation of adversarial legal proceedings that explores how AI systems can exploit procedural weaknesses...
7 months ago cs.MA cs.AI cs.CR
PDF
Attack HIGH
Xinzhe Huang, Wenjing Hu, Tianhang Zheng +5 more
Existing gradient-based jailbreak attacks on Large Language Models (LLMs) typically optimize adversarial suffixes to align the LLM output with...
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Yu He, Yifei Chen, Yiming Li +5 more
In recent years, RAG has emerged as a key paradigm for enhancing large language models (LLMs). By integrating externally retrieved information, RAG...
Attack HIGH
Zhixin Xie, Xurui Song, Jun Luo
Despite substantial efforts in safety alignment, recent research indicates that Large Language Models (LLMs) remain highly susceptible to jailbreak...
Attack HIGH
Chinthana Wimalasuriya, Spyros Tragoudas
Adversarial attacks present a significant threat to modern machine learning systems. Yet, existing detection methods often lack the ability to detect...
7 months ago cs.CR cs.CV cs.LG
PDF
Attack HIGH
Zhaorun Chen, Xun Liu, Mintong Kang +4 more
As vision-language models (VLMs) gain prominence, their multimodal interfaces also introduce new safety vulnerabilities, making the safety evaluation...
7 months ago cs.AI cs.LG
PDF
Benchmark HIGH
Chengquan Guo, Chulin Xie, Yu Yang +6 more
Code agents have gained widespread adoption due to their strong code generation capabilities and integration with code interpreters, enabling dynamic...
Tool HIGH
Jonathan Sneh, Ruomei Yan, Jialin Yu +6 more
As LLMs increasingly power agents that interact with external tools, tool use has become an essential mechanism for extending their capabilities....
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Ruohao Guo, Afshin Oroojlooy, Roshan Sridhar +3 more
Despite recent rapid progress in AI safety, current large language models remain vulnerable to adversarial attacks in multi-turn interaction...
7 months ago cs.LG cs.AI cs.CL
PDF
Attack HIGH
Kedong Xiu, Churui Zeng, Tianhang Zheng +6 more
Existing gradient-based jailbreak attacks typically optimize an adversarial suffix to induce a fixed affirmative response, e.g., ``Sure, here...
7 months ago cs.CR cs.AI
PDF
Attack HIGH
Milad Nasr, Yanick Fratantonio, Luca Invernizzi +7 more
As deep learning models become widely deployed as components within larger production systems, their individual shortcomings can create system-level...
7 months ago cs.CR cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial