Benchmark HIGH
Zejian Chen, Chaozhuo Li, Chao Li +3 more
This paper provides a systematic survey of jailbreak attacks and defenses on Large Language Models (LLMs) and Vision-Language Models (VLMs),...
Benchmark LOW
Shidong Cao, Hongzhan Lin, Yuxuan Gu +2 more
Chain-of-Thought (CoT) reasoning improves multi-step mathematical problem solving in large language models but remains vulnerable to exposure bias...
Defense MEDIUM
Di Wu, Yanyan Zhao, Xin Lu +2 more
Defending against jailbreak attacks is crucial for the safe deployment of Large Language Models (LLMs). Recent research has attempted to improve...
2 months ago cs.AI cs.CL
PDF
Tool LOW
Rasmus Erlemann, Charles Colyer Morris, Sanjyot Sathe
The emergence of large-scale quantum computing threatens widely deployed public-key cryptographic systems, creating an urgent need for...
Attack HIGH
Zhakshylyk Nurlanov, Frank R. Schmidt, Florian Bernard
As Large Language Models (LLMs) are increasingly deployed in safety-critical domains, rigorously evaluating their robustness against adversarial...
2 months ago cs.LG cs.AI cs.CR
PDF
Tool HIGH
Xiangdong Hu, Yangyang Jiang, Qin Hu +1 more
Multimodal Large Language Models (MLLMs) have become widely deployed, yet their safety alignment remains fragile under adversarial inputs. Previous...
Tool MEDIUM
Sofie Goethals, Foster Provost, João Sedoc
As generative AI systems become integrated into real-world applications, organizations increasingly need to be able to understand and interpret their...
2 months ago cs.LG cs.AI cs.CL
PDF
Benchmark HIGH
Xiangzhe Yuan, Zhenhao Zhang, Haoming Tang +1 more
As LLMs gain persuasive agentic capabilities through extended dialogues, they introduce novel risks in multi-turn conversational scams that...
Benchmark LOW
Nick Pepper, Adam Keane, Amy Hodgkin +11 more
This paper presents the first probabilistic Digital Twin of operational en route airspace, developed for the London Area Control Centre. The Digital...
Attack MEDIUM
Bocheng Chen, Xi Chen, Han Zi +5 more
Identifying specific moral errors in an input and generating appropriate corrections require moral sensitivity in large language models (LLMs), which...
Attack HIGH
Xi Wang, Songlei Jian, Shasha Li +5 more
Despite extensive safety alignment, Large Language Models (LLMs) often fail against jailbreak attacks. While machine unlearning has emerged as a...
2 months ago cs.CR cs.AI
PDF
Benchmark LOW
Chengcheng Feng, Haojie Yin, Yucheng Jin +1 more
Comic-based visual question answering (CVQA) poses distinct challenges to multimodal large language models (MLLMs) due to its reliance on symbolic...
2 months ago cs.CV cs.AI
PDF
Attack HIGH
Yuetian Chen, Yuntao Du, Kaiyuan Zhang +4 more
Most membership inference attacks (MIAs) against Large Language Models (LLMs) rely on global signals, like average loss, to identify training data....
2 months ago cs.CL cs.AI cs.CR
PDF
Survey MEDIUM
Gaurav Sarraf, Vibhor Pal
Cloud computing has changed online communities in three dimensions, which are scalability, adaptability and reduced overhead. But there are serious...
Attack HIGH
Dinghong Song, Zhiwei Xu, Hai Wan +3 more
Model quantization is critical for deploying large language models (LLMs) on resource-constrained hardware, yet recent work has revealed severe...
2 months ago cs.CR cs.LG
PDF
Tool LOW
Gongao Zhang, Haijiang Zeng, Lu Jiang
Financial institutions and regulators require systems that integrate heterogeneous data to assess risks from stock fluctuations to systemic...
2 months ago cs.LG q-fin.RM q-fin.ST
PDF
Attack HIGH
Scott Thornton
Large language models remain vulnerable to jailbreak attacks, and single-layer defenses often trade security for usability. We present TRYLOCK, the...
2 months ago cs.CR cs.LG
PDF
Attack MEDIUM
Ahmed Ahmed, A. Feder Cooper, Sanmi Koyejo +1 more
Many unresolved legal questions over LLMs and copyright center on memorization: whether specific training data have been encoded in the model's...
2 months ago cs.CL cs.AI cs.LG
PDF
Attack HIGH
Devang Kulshreshtha, Hang Su, Chinmay Hegde +1 more
Most jailbreak methods achieve high attack success rates (ASR) but require attacker LLMs to craft adversarial queries and/or demand high query...
Tool MEDIUM
Md Ajoad Hasan, Dipayan Saha, Khan Thamid Hasan +5 more
The growing complexity of modern system-on-chip (SoC) and IP designs is making security assurance difficult day by day. One of the fundamental steps...
2 months ago cs.CR cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial