Benchmark MEDIUM
Djiré Albérick Euraste, Kaboré Abdoul Kader, Jordan Samhi +3 more
The lack of transparency about code datasets used to train large language models (LLMs) makes it difficult to detect, evaluate, and mitigate data...
Survey MEDIUM
Yi Ting Shen, Kentaroh Toyoda, Alex Leung
The rapid proliferation of Model Context Protocol (MCP)-based agentic systems has introduced a new category of security threats that existing...
3 weeks ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Xixun Lin, Yang Liu, Yancheng Chen +9 more
The performance of large language model (LLM) agents depends critically on the execution harness, the system layer that orchestrates tool use,...
4 weeks ago cs.CR cs.AI
PDF
Defense LOW
Aram Ebtekar, Michael K. Cohen
Reinforcement learners can attain high reward through novel unintended strategies. We study a Bayesian mitigation for general environments: we expand...
4 weeks ago cs.LG cs.AI
PDF
Defense MEDIUM
Xiaohua Wang, Muzhao Tian, Yuqi Zeng +20 more
Reinforcement Learning from Human Feedback (RLHF) and related alignment paradigms have become central to steering large language models (LLMs) and...
Tool LOW
Shawn, Zhong, Junxuan Liao +4 more
AI coding agents operate directly on users' filesystems, where they regularly corrupt data, delete files, and leak secrets. Current approaches force...
Benchmark LOW
Eun Woo Im, Dhruv Madhwal, Vivek Gupta
Vision-Language Models demonstrate remarkable capabilities but often struggle with compositional reasoning, exhibiting vulnerabilities regarding word...
Attack HIGH
Andrii Vakhnovskyi
The United States designates Food and Agriculture as one of sixteen critical infrastructure sectors, yet no mandatory cybersecurity requirements...
4 weeks ago cs.CR eess.SY
PDF
Defense MEDIUM
Sujan Ghimire, Parsa Mirfasihi, Muhtasim Alam Chowdhury +6 more
The globalization of integrated circuit (IC) design and manufacturing has increased the exposure of hardware intellectual property (IP) to untrusted...
Benchmark MEDIUM
Prajas Wadekar, Venkata Sai Pranav Bachina, Kunal Bhosikar +2 more
3D Gaussian Splatting (3DGS) has recently enabled highly photorealistic 3D reconstruction from casually captured multi-view images. However, this...
4 weeks ago cs.CV cs.CR cs.LG
PDF
Tool LOW
Syed Md Mukit Rashid, Abdullah Al Ishtiaq, Kai Tu +7 more
Logical vulnerabilities in software stem from flaws in program logic rather than memory safety, which can lead to critical security failures....
4 weeks ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Joel Fokou
Autonomous AI agents are rapidly transitioning from experimental tools to operational infrastructure, with projections that 80% of enterprise...
4 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Yingying Zhao, Chengyin Hu, Qike Zhang +7 more
Vision-Language Models (VLMs) have shown remarkable performance, yet their security remains insufficiently understood. Existing adversarial studies...
Attack MEDIUM
Shaopeng Fu, Di Wang
Adversarial training (AT) is an effective defense for large language models (LLMs) against jailbreak attacks, but performing AT on LLMs is costly. To...
4 weeks ago cs.LG cs.CR stat.ML
PDF
Attack MEDIUM
Anasuya Chattopadhyay, Daniel Reti, Hans D. Schotten
Cloud networks increasingly rely on machine learning based Network Intrusion Detection Systems to defend against evolving cyber threats. However,...
4 weeks ago cs.LG cs.CR
PDF
Attack HIGH
Jianhao Chen, Haoyang Chen, Hanjie Zhao +2 more
The rapid evolution of Vision-Language Models (VLMs) has catalyzed unprecedented capabilities in artificial intelligence; however, this continuous...
4 weeks ago cs.AI cs.MM
PDF
Attack MEDIUM
Vladimir A. Mazin, Mikhail A. Zorin, Dmitrii S. Korzh +3 more
Passwords still remain a dominant authentication method, yet their security is routinely subverted by predictable user choices and large-scale...
4 weeks ago cs.CR cs.AI
PDF
Attack HIGH
Junyu Ren, Xingjian Pan, Wensheng Gan +1 more
Prompt injection has emerged as a critical security threat to large language models (LLMs), yet existing studies predominantly focus on...
Benchmark MEDIUM
Miit Daga, Swarna Priya Ramu
Organisations increasingly outsource privacy-sensitive data transformations to cloud providers, yet no practical mechanism lets the data owner verify...
4 weeks ago cs.CR cs.DB cs.LG
PDF
Attack HIGH
Ravikumar Balakrishnan, Sanket Mendapara, Ankit Garg
We study typographic prompt injection attacks on vision-language models (VLMs), where adversarial text is rendered as images to bypass safety...
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial