Benchmark MEDIUM
Zijun Feng, Yuming Feng, Yu Wang +4 more
Cross-chain bridges, the critical infrastructure of the multi-chain ecosystem, have become a primary target for attackers, resulting in over $2.8...
Attack MEDIUM
Mengnan Zhao, Lihe Zhang, Bo Wang +3 more
Fast Adversarial Training (FAT) has proven effective in enhancing model robustness by encouraging networks to learn perturbation-invariant...
2 weeks ago cs.LG cs.CR
PDF
Benchmark MEDIUM
Víctor Mayoral-Vilches, María Sanz-Gómez, Francesco Balassone +6 more
As LLM-driven agents advance in cybersecurity, Jeopardy CTF benchmarks are approaching saturation and cyber ranges, the natural next evaluation...
Defense MEDIUM
Kaisheng Fan, Weizhe Zhang, Yishu Gao +2 more
Defending against backdoor attacks in large language models remains a critical practical challenge. Existing defenses mitigate these threats but...
2 weeks ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Eungyu Woo, Yooshin Kim, Wonje Heo +1 more
Industrial Control Systems (ICS) integrate computing, physical processes, and communication to operate critical infrastructures such as power grids,...
Survey MEDIUM
Jiaqi Li, Yang Zhao, Bin Sun +3 more
Autonomous AI agents deployed on platforms such as OpenClaw face prompt injection, memory poisoning, supply-chain attacks, and social engineering,...
2 weeks ago cs.CR cs.AI
PDF
Tool MEDIUM
Kato Mivule
This paper extends the Classification Error Gauge (x-CEG) framework, originally developed for measuring the privacy-utility trade-off in tabular...
Benchmark MEDIUM
Qi Li, Bo Yin, Weiqi Huang +6 more
Vision-Language-Action (VLA) models are emerging as a unified substrate for embodied intelligence. This shift raises a new class of safety...
Survey MEDIUM
Jialiang Wang, Yuchen Liu, Hang Xu +7 more
The volume of scientific submissions continues to climb, outpacing the capacity of qualified human referees and stretching editorial timelines. At...
Benchmark MEDIUM
Yuchen Shi, Xin Guo, Huajie Chen +3 more
Poisoning-based backdoor attacks pose significant threats to deep neural networks by embedding triggers in training data, causing models to...
2 weeks ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Vishal Rajput
We prove that empirical risk minimisation (ERM) imposes a necessary geometric constraint on learned representations: any encoder that minimises...
2 weeks ago cs.LG cs.AI cs.CV
PDF
Attack MEDIUM
Irti Haq, Belén Saldías
As state-of-the-art Large Language Models (LLMs) have become ubiquitous, ensuring equitable performance across diverse demographics is critical....
2 weeks ago cs.CY cs.AI cs.CL
PDF
Benchmark MEDIUM
Ari Azarafrooz
AI-agent guardrails are memoryless: each message is judged in isolation, so an adversary who spreads a single attack across dozens of sessions slips...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Benchmark MEDIUM
Mohammad Farhad, Shuvalaxmi Dass
Software security relies on effective vulnerability detection and patching, yet determining whether a patch fully eliminates risk remains an...
2 weeks ago cs.SE cs.CR
PDF
Tool MEDIUM
Mikko Lempinen, Joni Kemppainen, Niklas Raesalmi
As artificial intelligence (AI) systems are increasingly deployed across critical domains, their security vulnerabilities pose growing risks of...
2 weeks ago cs.CR cs.AI cs.CL
PDF
Defense MEDIUM
Krishiv Agarwal, Ramneet Kaur, Colin Samplawski +6 more
Effective safety auditing of large language models (LLMs) demands tools that go beyond black-box probing and systematically uncover vulnerabilities...
2 weeks ago cs.CR cs.LG
PDF
Benchmark MEDIUM
Hoang Nguyen, Lu Wang, Marta Gaia Bras
Freight brokerages negotiate thousands of carrier rates daily under dynamic pricing conditions where models frequently revise targets...
2 weeks ago cs.MA cs.AI cs.CL
PDF
Attack MEDIUM
Abhijit Talluri
Adversarial robustness evaluation underpins every claim of trustworthy ML deployment, yet the field suffers from fragmented protocols and undetected...
2 weeks ago cs.CR cs.LG
PDF
Other MEDIUM
Yuhang Wu, Qinyuan Liu, Qiuyang Zhao +1 more
Currently, Large Language Models (LLMs) feature a diversified architectural landscape, including traditional Transformer, GateDeltaNet, and Mamba....
2 weeks ago cs.CL cs.AI
PDF
Defense MEDIUM
Chao Pan, Yu Wu, Xin Yao
Internal Safety Collapse (ISC) is a failure mode in which frontier LLMs, when executing legitimate professional tasks whose correct completion...
2 weeks ago cs.CR cs.AI cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial