Defense MEDIUM
Nay Myat Min, Long H. Pham, Jun Sun
Large language models deployed at runtime can misbehave in ways that clean-data validation cannot anticipate: training-time backdoors lie dormant...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Benchmark MEDIUM
Pablo Mateo-Torrejón, Alfonso Sánchez-Macián
The rapid integration of Large Language Models (LLMs) into Multi-Agent Systems (MAS) has significantly enhanced their collaborative problem-solving...
2 weeks ago cs.CR cs.AI cs.MA
PDF
Survey MEDIUM
Zihan Liu, Yizhen Wang, Rui Wang +2 more
Fine-tuning unlocks large language models (LLMs) for specialized applications, but its high computational cost often puts it out of reach for...
2 weeks ago cs.CR cs.CL cs.DC
PDF
Attack MEDIUM
Mengnan Zhao, Lihe Zhang, Tianhang Zheng +2 more
Fast Adversarial Training (FAT) has attracted significant attention due to its efficiency in enhancing neural network robustness against adversarial...
2 weeks ago cs.LG cs.AI cs.CR
PDF
Tool LOW
Zheng Wu, Yi Hua, Zhaoyuan Huang +8 more
The evolution of Multimodal Large Language Models (MLLMs) has shifted the focus from text generation to active behavioral execution, particularly via...
Benchmark MEDIUM
Zijun Feng, Yuming Feng, Yu Wang +4 more
Cross-chain bridges, the critical infrastructure of the multi-chain ecosystem, have become a primary target for attackers, resulting in over $2.8...
Attack MEDIUM
Mengnan Zhao, Lihe Zhang, Bo Wang +3 more
Fast Adversarial Training (FAT) has proven effective in enhancing model robustness by encouraging networks to learn perturbation-invariant...
2 weeks ago cs.LG cs.CR
PDF
Benchmark MEDIUM
Víctor Mayoral-Vilches, María Sanz-Gómez, Francesco Balassone +6 more
As LLM-driven agents advance in cybersecurity, Jeopardy CTF benchmarks are approaching saturation and cyber ranges, the natural next evaluation...
Defense MEDIUM
Kaisheng Fan, Weizhe Zhang, Yishu Gao +2 more
Defending against backdoor attacks in large language models remains a critical practical challenge. Existing defenses mitigate these threats but...
2 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Zonghao Ying, Haozheng Wang, Jiangfan Liu +5 more
Large Language Model (LLM) agents are increasingly used to automate complex workflows, but integrating untrusted external data with privileged...
Tool LOW
Jeffrey Wong, Antoine Creux
Create an idea, prototype it, evaluate if users like it, then learn. It is the circle of business. If AI can operate in all parts of the circle, it...
2 weeks ago cs.SE cs.MS stat.AP
PDF
Attack HIGH
Xinhe Wang, Katia Sycara, Yaqi Xie
Large (vision-)language models exhibit remarkable capability but remain highly susceptible to jailbreaking. Existing safety training approaches aim...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Benchmark MEDIUM
Eungyu Woo, Yooshin Kim, Wonje Heo +1 more
Industrial Control Systems (ICS) integrate computing, physical processes, and communication to operate critical infrastructures such as power grids,...
Defense LOW
Sijia Li, Min Gao, Zongwei Wang +3 more
Sequential recommendation seeks to model the evolution of user interests by capturing temporal user intent and item-level transition patterns....
Survey MEDIUM
Jiaqi Li, Yang Zhao, Bin Sun +3 more
Autonomous AI agents deployed on platforms such as OpenClaw face prompt injection, memory poisoning, supply-chain attacks, and social engineering,...
2 weeks ago cs.CR cs.AI
PDF
Benchmark HIGH
Priyal Deep, Shane Emmons, Amy Fox +3 more
LLM-powered applications routinely embed secrets in system prompts, yet models can be tricked into revealing them. We built an adaptive attacker that...
2 weeks ago cs.CR cs.AI
PDF
Tool MEDIUM
Kato Mivule
This paper extends the Classification Error Gauge (x-CEG) framework, originally developed for measuring the privacy-utility trade-off in tabular...
Benchmark MEDIUM
Qi Li, Bo Yin, Weiqi Huang +6 more
Vision-Language-Action (VLA) models are emerging as a unified substrate for embodied intelligence. This shift raises a new class of safety...
Attack HIGH
Yu Cui, Ruiqing Yue, Hang Fu +6 more
With the wide adoption of personal AI assistants such as OpenClaw, privacy leakage in user interaction contexts with large language model (LLM)...
Attack LOW
Rong Xiang
Recent evidence suggests that frontier AI systems can exhibit agentic misalignment, generating and executing harmful actions derived from internally...
2 weeks ago cs.AI cs.CR
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial