Attack HIGH
Yuhao Wang, Shengfang Zhai, Guanghao Jin +3 more
Large Language Model (LLM)-based agents employ external and internal memory systems to handle complex, goal-oriented tasks, yet this exposes them to...
3 months ago cs.CR cs.AI cs.CL
PDF
Attack MEDIUM
Zhiyu Sun, Minrui Luo, Yu Wang +2 more
Large language models (LLMs) are pretrained on corpora containing trillions of tokens and, therefore, inevitably memorize sensitive information....
3 months ago cs.CR cs.AI cs.CL
PDF
Attack MEDIUM
Ruoyao Wen, Hao Li, Chaowei Xiao +1 more
Indirect prompt injection threatens LLM agents by embedding malicious instructions in external content, enabling unauthorized actions and data theft....
3 months ago cs.CR cs.AI
PDF
Attack HIGH
Abdullah Arafat Miah, Kevin Vu, Yu Bi
Spiking Neural Networks (SNNs) are energy-efficient counterparts of Deep Neural Networks (DNNs) with high biological plausibility, as information is...
3 months ago cs.CR cs.AI
PDF
Attack HIGH
Shang Liu, Hanyu Pei, Zeyan Liu
Large Language Models(LLMs) have been successful in numerous fields. Alignment has usually been applied to prevent them from harmful purposes....
3 months ago cs.CR cs.AI
PDF
Attack HIGH
Zhuoheng Li, Ying Chen
Multimodal large language models (MLLMs) have advanced the capabilities to interpret and act on visual input in 3D environments, empowering diverse...
3 months ago cs.CV cs.AI
PDF
Attack HIGH
Mingqian Feng, Xiaodong Liu, Weiwei Yang +4 more
Multi-turn jailbreaks capture the real threat model for safety-aligned chatbots, where single-turn attacks are merely a special case. Yet existing...
Attack HIGH
Yassine Chagna, Antal Goldschmidt
This project explores large language models (LLMs) for anomaly detection across heterogeneous log sources. Traditional intrusion detection systems...
3 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Fengpeng Li, Kemou Li, Qizhou Wang +2 more
Concept erasure helps stop diffusion models (DMs) from generating harmful content; but current methods face robustness retention trade off....
3 months ago cs.LG cs.AI cs.CR
PDF
Attack HIGH
Ying Song, Balaji Palanisamy
Graph-structured data underpin a wide spectrum of modern applications. However, complex graph topologies and homophilic patterns can facilitate...
3 months ago cs.CR cs.LG
PDF
Attack HIGH
Mengyao Du, Han Fang, Haokai Ma +4 more
Suffix-based jailbreak attacks append an adversarial suffix, i.e., a short token sequence, to steer aligned LLMs into unsafe outputs. Since suffixes...
Attack HIGH
Haipeng Li, Rongxuan Peng, Anwei Luo +3 more
The rapid advancement of AI-Generated Content (AIGC) technologies poses significant challenges for authenticity assessment. However, existing...
3 months ago cs.CV cs.CR
PDF
Attack HIGH
Minkyoo Song, Jaehan Kim, Myungchul Kang +3 more
Graph-based retrieval-augmented generation (Graph RAG) is increasingly deployed to support LLM applications by augmenting user queries with...
Attack HIGH
Sung-Hoon Yoon, Ruizhi Qian, Minda Zhao +2 more
Large Language Models (LLMs) have become integral to many domains, making their safety a critical priority. Prior jailbreaking research has explored...
3 months ago cs.CL cs.AI cs.CR
PDF
Attack LOW
Mona Rajhans, Vishal Khawarey
Machine learning (ML) models are increasingly deployed in cybersecurity applications such as phishing detection and network intrusion prevention....
3 months ago cs.CR cs.AI cs.LG
PDF
Attack HIGH
Xin Chen, Jie Zhang, Florian Tramèr
Prompt injection is one of the most critical vulnerabilities in LLM agents; yet, effective automated attacks remain largely unexplored from an...
3 months ago cs.LG cs.AI
PDF
Attack MEDIUM
Tao Huang, Rui Wang, Xiaofei Liu +3 more
%Large vision-language models (LVLMs) have shown substantial advances in multimodal understanding and generation. However, when presented with...
Attack HIGH
Takashi Koide, Hiroki Nakano, Daiki Chiba
Phishing sites continue to grow in volume and sophistication. Recent work leverages large language models (LLMs) to analyze URLs, HTML, and rendered...
Attack HIGH
Yao Zhou, Zeen Song, Wenwen Qiang +4 more
Safety alignment mechanisms in Large Language Models (LLMs) often operate as latent internal states, obscuring the model's inherent capabilities....
Attack HIGH
Zihan Wang, Hongwei Li, Rui Zhang +2 more
Chat template is a common technique used in the training and inference stages of Large Language Models (LLMs). It can transform input and output data...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial