Survey MEDIUM
Sahaya Jestus Lazer, Kshitiz Aryal, Maanak Gupta +1 more
Agentic AI marks an important transition from single-step generative models to systems capable of reasoning, planning, acting, and adapting over...
2 months ago cs.CR cs.AI
PDF
Benchmark LOW
Jacob Ede Levine, Yun Lyan Luo, Sai Chandra Kosaraju
The design of reliable, valid, and diverse molecules is fundamental to modern drug discovery, as improved molecular generation supports efficient...
2 months ago cs.LG cs.AI
PDF
Benchmark LOW
Atharv Naphade
Retrieval-Augmented Generation (RAG) is the prevailing paradigm for grounding Large Language Models (LLMs), yet the mechanisms governing how models...
2 months ago cs.AI cs.LG
PDF
Attack MEDIUM
San Kim, Gary Geunbae Lee
Large Language Models (LLMs) have greatly advanced Natural Language Processing (NLP), particularly through instruction tuning, which enables broad...
2 months ago cs.CL cs.AI
PDF
Attack HIGH
Ahmad Mohammad Saber, Saeed Jafari, Zhengmao Ouyang +3 more
This paper presents a large language model (LLM)-based framework that adapts and fine-tunes compact LLMs for detecting cyberattacks on transformer...
2 months ago cs.CR cs.LG eess.SP
PDF
Attack HIGH
Iago Alves Brito, Walcy Santos Rezende Rios, Julia Soares Dollis +2 more
Current safety evaluations of large language models (LLMs) create a dangerous illusion of universality, aggregating "Identity Hate" into scalar...
2 months ago cs.CL cs.AI
PDF
Attack HIGH
Yu Yan, Sheng Sun, Mingfeng Li +6 more
Recently, people have suffered from LLM hallucination and have become increasingly aware of the reliability gap of LLMs in open and...
Benchmark LOW
Xinyue Lou, Jinan Xu, Jingyi Yin +8 more
As Multimodal Large Language Models (MLLMs) become an indispensable assistant in human life, the unsafe content generated by MLLMs poses a danger to...
Attack HIGH
Siyuan Li, Xi Lin, Jun Wu +5 more
Jailbreak attacks pose significant threats to large language models (LLMs), enabling attackers to bypass safeguards. However, existing reactive...
2 months ago cs.CR cs.AI
PDF
Survey MEDIUM
Andreea-Elena Bodea, Stephen Meisenbacher, Alexandra Klymenko +1 more
The continued promise of Large Language Models (LLMs), particularly in their natural language understanding and generation capabilities, has driven a...
2 months ago cs.CR cs.CL
PDF
Benchmark LOW
Haeun Jang, Hwan Chang, Hwanhee Lee
The deployment of Large Vision-Language Models (LVLMs) for real-world document question answering is often constrained by dynamic, user-defined...
Benchmark MEDIUM
Xiaoyu Xu, Minxin Du, Zitong Li +6 more
Although machine unlearning is essential for removing private, harmful, or copyrighted content from LLMs, current benchmarks often fail to faithfully...
2 months ago cs.CL cs.AI cs.CR
PDF
Defense MEDIUM
Xing Li, Hui-Ling Zhen, Lihao Yin +3 more
This paper presents a comprehensive empirical study on the safety alignment capabilities. We evaluate what matters for safety alignment in LLMs and...
2 months ago cs.CL cs.AI cs.CR
PDF
Benchmark MEDIUM
Dinesh Srivasthav P, Ashok Urlana, Rahul Mishra +2 more
Machine unlearning aims to selectively remove the influence of specific training samples to satisfy privacy regulations such as the GDPR's 'Right to...
2 months ago cs.CR cs.AI cs.CL
PDF
Benchmark LOW
Jin Wang, Liang Lin, Kaiwen Luo +8 more
While Audio Large Language Models (ALLMs) have achieved remarkable progress in understanding and generation, their potential privacy implications...
Attack HIGH
Ji Guo, Wenbo Jiang, Yansong Lin +7 more
Vision-Language-Action (VLA) models are widely deployed in safety-critical embodied AI applications such as robotics. However, their complex...
2 months ago cs.CR cs.LG
PDF
Benchmark HIGH
Quy-Anh Dang, Chris Ngo, Truong-Son Hy
As large language models (LLMs) become integral to safety-critical applications, ensuring their robustness against adversarial prompts is paramount....
Attack HIGH
Hang Fu, Wanli Peng, Yinghan Zhou +3 more
The widespread adoption of Large Language Model (LLM) in commercial and research settings has intensified the need for robust intellectual property...
Attack HIGH
Binh Nguyen, Thai Le
Audio Language Models (ALMs) offer a promising shift towards explainable audio deepfake detections (ADDs), moving beyond \textit{black-box}...
2 months ago cs.CL cs.SD eess.AS
PDF
Attack HIGH
Xiao Lin, Philip Li, Zhichen Zeng +6 more
Despite rich safety alignment strategies, large language models (LLMs) remain highly susceptible to jailbreak attacks, which compromise safety...
2 months ago cs.LG cs.AI cs.IR
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial