Attack HIGH
Songze Li, Ruishi He, Xiaojun Jia +2 more
Large Language Models (LLMs) face a significant threat from multi-turn jailbreak attacks, where adversaries progressively steer conversations to...
4 months ago cs.CR cs.LG
PDF
Attack HIGH
Badhan Chandra Das, Md Tasnim Jawad, Joaquin Molto +2 more
In recent years, the security vulnerabilities of Multi-modal Large Language Models (MLLMs) have become a serious concern in the Generative Artificial...
4 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Mizuki Sakai, Mizuki Yokoyama, Wakaba Tateishi +1 more
Large language models (LLMs) are increasingly used as autonomous agents in strategic and social interactions. Although recent studies suggest that...
Attack HIGH
Zhiyuan Chang, Mingyang Li, Yuekai Huang +6 more
Large language model (LLM)-integrated applications have become increasingly prevalent, yet face critical security vulnerabilities from prompt...
4 months ago cs.AI cs.CR
PDF
Attack HIGH
Hoagy Cunningham, Jerry Wei, Zihan Wang +26 more
We introduce enhanced Constitutional Classifiers that deliver production-grade jailbreak robustness with dramatically reduced computational costs and...
4 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Mohamed Nabeel, Oleksii Starov
According to Gartner, more than 70% of organizations will have integrated AI models into their workflows by the end of 2025. In order to reduce cost...
Attack MEDIUM
San Kim, Gary Geunbae Lee
Large Language Models (LLMs) have greatly advanced Natural Language Processing (NLP), particularly through instruction tuning, which enables broad...
4 months ago cs.CL cs.AI
PDF
Attack HIGH
Ahmad Mohammad Saber, Saeed Jafari, Zhengmao Ouyang +3 more
This paper presents a large language model (LLM)-based framework that adapts and fine-tunes compact LLMs for detecting cyberattacks on transformer...
4 months ago cs.CR cs.LG eess.SP
PDF
Attack HIGH
Iago Alves Brito, Walcy Santos Rezende Rios, Julia Soares Dollis +2 more
Current safety evaluations of large language models (LLMs) create a dangerous illusion of universality, aggregating "Identity Hate" into scalar...
4 months ago cs.CL cs.AI
PDF
Attack HIGH
Yu Yan, Sheng Sun, Mingfeng Li +6 more
Recently, people have suffered from LLM hallucination and have become increasingly aware of the reliability gap of LLMs in open and...
Attack HIGH
Siyuan Li, Xi Lin, Jun Wu +5 more
Jailbreak attacks pose significant threats to large language models (LLMs), enabling attackers to bypass safeguards. However, existing reactive...
4 months ago cs.CR cs.AI
PDF
Attack HIGH
Ji Guo, Wenbo Jiang, Yansong Lin +7 more
Vision-Language-Action (VLA) models are widely deployed in safety-critical embodied AI applications such as robotics. However, their complex...
4 months ago cs.CR cs.LG
PDF
Attack HIGH
Hang Fu, Wanli Peng, Yinghan Zhou +3 more
The widespread adoption of Large Language Model (LLM) in commercial and research settings has intensified the need for robust intellectual property...
Attack HIGH
Binh Nguyen, Thai Le
Audio Language Models (ALMs) offer a promising shift towards explainable audio deepfake detections (ADDs), moving beyond \textit{black-box}...
4 months ago cs.CL cs.SD eess.AS
PDF
Attack HIGH
Xiao Lin, Philip Li, Zhichen Zeng +6 more
Despite rich safety alignment strategies, large language models (LLMs) remain highly susceptible to jailbreak attacks, which compromise safety...
4 months ago cs.LG cs.AI cs.IR
PDF
Attack HIGH
Zhakshylyk Nurlanov, Frank R. Schmidt, Florian Bernard
As Large Language Models (LLMs) are increasingly deployed in safety-critical domains, rigorously evaluating their robustness against adversarial...
4 months ago cs.LG cs.AI cs.CR
PDF
Attack MEDIUM
Bocheng Chen, Xi Chen, Han Zi +5 more
Identifying specific moral errors in an input and generating appropriate corrections require moral sensitivity in large language models (LLMs), which...
Attack HIGH
Xi Wang, Songlei Jian, Shasha Li +5 more
Despite extensive safety alignment, Large Language Models (LLMs) often fail against jailbreak attacks. While machine unlearning has emerged as a...
4 months ago cs.CR cs.AI
PDF
Attack HIGH
Yuetian Chen, Yuntao Du, Kaiyuan Zhang +4 more
Most membership inference attacks (MIAs) against Large Language Models (LLMs) rely on global signals, like average loss, to identify training data....
4 months ago cs.CL cs.AI cs.CR
PDF
Attack HIGH
Dinghong Song, Zhiwei Xu, Hai Wan +3 more
Model quantization is critical for deploying large language models (LLMs) on resource-constrained hardware, yet recent work has revealed severe...
4 months ago cs.CR cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial