Survey LOW
Brecht Verbeken, Joke Van den Broeck, Inge De Cleyn +4 more
Purpose: Higher education institutions face increasing pressure to audit course designs for generative AI (GenAI) integration. This paper presents an...
Benchmark MEDIUM
Arjun Chakraborty, Sandra Ho, Adam Cook +1 more
CTI-REALM (Cyber Threat Real World Evaluation and LLM Benchmarking) is a benchmark designed to evaluate AI agents' ability to interpret cyber threat...
Benchmark LOW
Ziyu Liu, Shengyuan Ding, Xinyu Fang +7 more
Vision-to-code tasks require models to reconstruct structured visual inputs, such as charts, tables, and SVGs, into executable or structured...
2 months ago cs.CV cs.AI
PDF
Attack MEDIUM
Md. Abdul Awal, Mrigank Rochan, Chanchal K. Roy
Large language models for code have achieved strong performance across diverse software analytics tasks, yet their real-world adoption remains...
Attack MEDIUM
Jianwei Li, Jung-Eun Kim
Backdoor attacks pose severe security threats to large language models (LLMs), where a model behaves normally under benign inputs but produces...
2 months ago cs.CR cs.AI cs.LG
PDF
Attack HIGH
Chenlong Yin, Runpeng Geng, Yanting Wang +1 more
Prompt injection poses serious security risks to real-world LLM applications, particularly autonomous agents. Although many defenses have been...
2 months ago cs.LG cs.CR
PDF
Defense MEDIUM
Matthew Butler, Yi Fan, Christos Faloutsos
The proposed method (FraudFox) provides solutions to adversarial attacks in a resource constrained environment. We focus on questions like the...
2 months ago cs.CR cs.LG
PDF
Benchmark MEDIUM
Zhifang Zhang, Bojun Yang, Shuo He +5 more
Despite the strong multimodal performance, large vision-language models (LVLMs) are vulnerable during fine-tuning to backdoor attacks, where...
2 months ago cs.CV cs.CR
PDF
Attack HIGH
Zheng Gao, Yifan Yang, Xiaoyu Li +4 more
Watermarking the initial noise of diffusion models has emerged as a promising approach for image provenance, but content-independent noise patterns...
2 months ago cs.CV cs.CR cs.LG
PDF
Attack HIGH
Sihao Ding
We introduce Colluding LoRA (CoLoRA), an attack in which each adapter appears benign and plausibly functional in isolation, yet their linear...
2 months ago cs.CR cs.LG
PDF
Defense MEDIUM
Zonghao Ying, Xiao Yang, Siyang Wu +7 more
The rapid evolution of Large Language Models (LLMs) into autonomous, tool-calling agents has fundamentally altered the cybersecurity landscape....
Tool MEDIUM
Jiangrong Wu, Zitong Yao, Yuhong Nan +1 more
Tool-augmented LLM agents increasingly rely on multi-step, multi-tool workflows to complete real tasks. This design expands the attack surface,...
2 months ago cs.SE cs.CR
PDF
Attack MEDIUM
Xiangkui Cao, Jie Zhang, Meina Kan +2 more
Large Vision-Language Models (LVLMs) have shown remarkable potential across a wide array of vision-language tasks, leading to their adoption in...
Attack HIGH
Darren Cheng, Wen-Kwang Tsao
Prompt injection remains one of the most practical attack vectors against LLM-integrated applications. We replicate the Microsoft LLMail-Inject...
2 months ago cs.CR cs.AI
PDF
Benchmark HIGH
Siddharth Srikanth, Freddie Liang, Sophie Hsu +9 more
Vision-Language-Action (VLA) models have significant potential to enable general-purpose robotic systems for a range of vision-language tasks....
2 months ago cs.RO cs.AI cs.CL
PDF
Attack HIGH
Xinhai Wang, Shaopeng Fu, Shu Yang +3 more
Suffix jailbreak attacks serve as a systematic method for red-teaming Large Language Models (LLMs) but suffer from prohibitive computational costs,...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Davi Bonetto
State Space Models (SSMs) such as Mamba achieve linear-time sequence processing through input-dependent recurrence, but this mechanism introduces a...
2 months ago cs.LG cs.CR
PDF
Benchmark MEDIUM
Ninghui Li, Kaiyuan Zhang, Kyle Polley +1 more
This article, a lightly adapted version of Perplexity's response to NIST/CAISI Request for Information 2025-0035, details our observations and...
2 months ago cs.LG cs.AI cs.CR
PDF
Attack HIGH
Alexandre Le Mercier, Thomas Demeester, Chris Develder
State space models (SSMs) like Mamba have gained significant traction as efficient alternatives to Transformers, achieving linear complexity while...
Attack MEDIUM
Haodong Zhao, Jinming Hu, Yijie Bai +6 more
Federated Language Model (FedLM) allows a collaborative learning without sharing raw data, yet it introduces a critical vulnerability, as every...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial