Benchmark LOW
Jiangnan Fang, Cheng-Tse Liu, Hanieh Deilamsalehy +5 more
Large language model (LLM) judges have often been used alongside traditional, algorithm-based metrics for tasks like summarization because they...
Survey LOW
Cen Zhang, Younggi Park, Fabian Fleischer +20 more
DARPA's AI Cyber Challenge (AIxCC, 2023--2025) is the largest competition to date for building fully autonomous cyber reasoning systems (CRSs) that...
1 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Sai Puppala, Ismail Hossain, Md Jahangir Alam +5 more
Large language models are increasingly deployed as *deep agents* that plan, maintain persistent state, and invoke external tools, shifting safety...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Yuhao Wang, Shengfang Zhai, Guanghao Jin +3 more
Large Language Model (LLM)-based agents employ external and internal memory systems to handle complex, goal-oriented tasks, yet this exposes them to...
1 months ago cs.CR cs.AI cs.CL
PDF
Attack MEDIUM
Zhiyu Sun, Minrui Luo, Yu Wang +2 more
Large language models (LLMs) are pretrained on corpora containing trillions of tokens and, therefore, inevitably memorize sensitive information....
1 months ago cs.CR cs.AI cs.CL
PDF
Benchmark HIGH
Tianyi Wu, Mingzhe Du, Yue Liu +4 more
Large language models (LLMs) are increasingly used in software development, yet their tendency to generate insecure code remains a major barrier to...
1 months ago cs.CR cs.AI cs.CL
PDF
Attack MEDIUM
Ruoyao Wen, Hao Li, Chaowei Xiao +1 more
Indirect prompt injection threatens LLM agents by embedding malicious instructions in external content, enabling unauthorized actions and data theft....
1 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Kunal Pai, Parth Shah, Harshil Patel
AI agents are increasingly deployed in production, yet their security evaluations remain bottlenecked by manual red-teaming or static benchmarks that...
1 months ago cs.AI cs.MA
PDF
Defense LOW
Gautam Siddharth Kashyap, Mark Dras, Usman Naseem
Large Language Models (LLMs) need to be in accordance with human values-being helpful, harmless, and honest (HHH)-is important for safe deployment....
Benchmark MEDIUM
Xiang Li, Pin-Yu Chen, Wenqi Wei
With the rapid advancement and adoption of Audio Large Language Models (ALLMs), voice agents are now being deployed in high-stakes domains such as...
1 months ago cs.CR cs.MA
PDF
Defense MEDIUM
Yunbei Zhang, Kai Mei, Ming Liu +5 more
We present the first large-scale empirical study of Moltbook, an AI-only social platform where 27,269 agents produced 137,485 posts and 345,580...
1 months ago cs.SI cs.AI
PDF
Tool MEDIUM
Juefei Pu, Xingyu Li, Zhengchuan Liang +5 more
Autonomous large language model (LLM) based systems have recently shown promising results across a range of cybersecurity tasks. However, there is no...
1 months ago cs.CR cs.SE
PDF
Benchmark MEDIUM
Qi Sun, Ahmed Abdo, Luis Burbano +4 more
Autonomous Vehicles (AVs), especially vision-based AVs, are rapidly being deployed without human operators. As AVs operate in safety-critical...
1 months ago cs.CR cs.LG
PDF
Attack HIGH
Abdullah Arafat Miah, Kevin Vu, Yu Bi
Spiking Neural Networks (SNNs) are energy-efficient counterparts of Deep Neural Networks (DNNs) with high biological plausibility, as information is...
1 months ago cs.CR cs.AI
PDF
Attack HIGH
Shang Liu, Hanyu Pei, Zeyan Liu
Large Language Models(LLMs) have been successful in numerous fields. Alignment has usually been applied to prevent them from harmful purposes....
1 months ago cs.CR cs.AI
PDF
Tool MEDIUM
Saad Hossain, Tom Tseng, Punya Syon Pandey +8 more
As increasingly capable open-weight large language models (LLMs) are deployed, improving their tamper resistance against unsafe modifications,...
1 months ago cs.CR cs.AI
PDF
Defense MEDIUM
Chen Chen, Yuchen Sun, Jiaxin Gao +4 more
Large language models (LLMs) are increasingly deployed in security-sensitive applications, yet remain vulnerable to backdoor attacks. However,...
Attack HIGH
Zhuoheng Li, Ying Chen
Multimodal large language models (MLLMs) have advanced the capabilities to interpret and act on visual input in 3D environments, empowering diverse...
1 months ago cs.CV cs.AI
PDF
Attack HIGH
Mingqian Feng, Xiaodong Liu, Weiwei Yang +4 more
Multi-turn jailbreaks capture the real threat model for safety-aligned chatbots, where single-turn attacks are merely a special case. Yet existing...
Attack HIGH
Yassine Chagna, Antal Goldschmidt
This project explores large language models (LLMs) for anomaly detection across heterogeneous log sources. Traditional intrusion detection systems...
1 months ago cs.CR cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial