Attack HIGH
Xinhe Wang, Katia Sycara, Yaqi Xie
Large (vision-)language models exhibit remarkable capability but remain highly susceptible to jailbreaking. Existing safety training approaches aim...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Benchmark MEDIUM
Eungyu Woo, Yooshin Kim, Wonje Heo +1 more
Industrial Control Systems (ICS) integrate computing, physical processes, and communication to operate critical infrastructures such as power grids,...
Defense LOW
Sijia Li, Min Gao, Zongwei Wang +3 more
Sequential recommendation seeks to model the evolution of user interests by capturing temporal user intent and item-level transition patterns....
Survey MEDIUM
Jiaqi Li, Yang Zhao, Bin Sun +3 more
Autonomous AI agents deployed on platforms such as OpenClaw face prompt injection, memory poisoning, supply-chain attacks, and social engineering,...
2 weeks ago cs.CR cs.AI
PDF
Benchmark HIGH
Priyal Deep, Shane Emmons, Amy Fox +3 more
LLM-powered applications routinely embed secrets in system prompts, yet models can be tricked into revealing them. We built an adaptive attacker that...
2 weeks ago cs.CR cs.AI
PDF
Tool MEDIUM
Kato Mivule
This paper extends the Classification Error Gauge (x-CEG) framework, originally developed for measuring the privacy-utility trade-off in tabular...
Benchmark MEDIUM
Qi Li, Bo Yin, Weiqi Huang +6 more
Vision-Language-Action (VLA) models are emerging as a unified substrate for embodied intelligence. This shift raises a new class of safety...
Attack HIGH
Yu Cui, Ruiqing Yue, Hang Fu +6 more
With the wide adoption of personal AI assistants such as OpenClaw, privacy leakage in user interaction contexts with large language model (LLM)...
Attack LOW
Rong Xiang
Recent evidence suggests that frontier AI systems can exhibit agentic misalignment, generating and executing harmful actions derived from internally...
2 weeks ago cs.AI cs.CR
PDF
Tool HIGH
Yuchuan Zhao, Tong Chen, Junliang Yu +3 more
Large language model-powered sequential recommender systems (LLM-SRSs) have recently demonstrated remarkable performance, enabling recommendations...
Survey MEDIUM
Jialiang Wang, Yuchen Liu, Hang Xu +7 more
The volume of scientific submissions continues to climb, outpacing the capacity of qualified human referees and stretching editorial timelines. At...
Benchmark LOW
Pegah Khayatan, Jayneel Parekh, Arnaud Dapogny +3 more
Despite impressive progress in capabilities of large vision-language models (LVLMs), these systems remain vulnerable to hallucinations, i.e., outputs...
2 weeks ago cs.CV cs.AI cs.CL
PDF
Attack HIGH
Naheed Rayhan, Sohely Jahan
Large language models (LLMs) are increasingly integrated into sensitive workflows, raising the stakes for adversarial robustness and safety. This...
2 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Zihan Wang, Rui Zhang, Yu Liu +4 more
LLM agents increasingly rely on skills to encapsulate reusable capabilities via progressively disclosed instructions. High-quality skills inject...
Attack HIGH
Jiali Wei, Ming Fan, Guoheng Sun +3 more
The growing application of large language models (LLMs) in safety-critical domains has raised urgent concerns about their security. Many recent...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Tool HIGH
Run Hao, Zhuoran Tan
Model Context Protocol (MCP) is increasingly adopted for tool-integrated LLM agents, but its multi-layer design and third-party server ecosystem...
Other LOW
Arthur Douillard, Keith Rush, Yani Donchev +14 more
Modern large-scale language model pre-training relies heavily on the single program multiple data (SPMD) paradigm, which requires tight coupling...
Benchmark MEDIUM
Yuchen Shi, Xin Guo, Huajie Chen +3 more
Poisoning-based backdoor attacks pose significant threats to deep neural networks by embedding triggers in training data, causing models to...
2 weeks ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Vishal Rajput
We prove that empirical risk minimisation (ERM) imposes a necessary geometric constraint on learned representations: any encoder that minimises...
2 weeks ago cs.LG cs.AI cs.CV
PDF
Benchmark LOW
Yongcan Yu, Lingxiao He, Jian Liang +5 more
Test-time reinforcement learning (TTRL) always adapts models at inference time via pseudo-labeling, leaving it vulnerable to spurious optimization...
2 weeks ago cs.LG cs.AI cs.CL
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial