Benchmark MEDIUM
Sai Puppala, Ismail Hossain, Md Jahangir Alam +5 more
Large language models are increasingly deployed as *deep agents* that plan, maintain persistent state, and invoke external tools, shifting safety...
1 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Zhiyu Sun, Minrui Luo, Yu Wang +2 more
Large language models (LLMs) are pretrained on corpora containing trillions of tokens and, therefore, inevitably memorize sensitive information....
1 months ago cs.CR cs.AI cs.CL
PDF
Attack MEDIUM
Ruoyao Wen, Hao Li, Chaowei Xiao +1 more
Indirect prompt injection threatens LLM agents by embedding malicious instructions in external content, enabling unauthorized actions and data theft....
1 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Kunal Pai, Parth Shah, Harshil Patel
AI agents are increasingly deployed in production, yet their security evaluations remain bottlenecked by manual red-teaming or static benchmarks that...
1 months ago cs.AI cs.MA
PDF
Benchmark MEDIUM
Xiang Li, Pin-Yu Chen, Wenqi Wei
With the rapid advancement and adoption of Audio Large Language Models (ALLMs), voice agents are now being deployed in high-stakes domains such as...
1 months ago cs.CR cs.MA
PDF
Defense MEDIUM
Yunbei Zhang, Kai Mei, Ming Liu +5 more
We present the first large-scale empirical study of Moltbook, an AI-only social platform where 27,269 agents produced 137,485 posts and 345,580...
1 months ago cs.SI cs.AI
PDF
Tool MEDIUM
Juefei Pu, Xingyu Li, Zhengchuan Liang +5 more
Autonomous large language model (LLM) based systems have recently shown promising results across a range of cybersecurity tasks. However, there is no...
1 months ago cs.CR cs.SE
PDF
Benchmark MEDIUM
Qi Sun, Ahmed Abdo, Luis Burbano +4 more
Autonomous Vehicles (AVs), especially vision-based AVs, are rapidly being deployed without human operators. As AVs operate in safety-critical...
1 months ago cs.CR cs.LG
PDF
Tool MEDIUM
Saad Hossain, Tom Tseng, Punya Syon Pandey +8 more
As increasingly capable open-weight large language models (LLMs) are deployed, improving their tamper resistance against unsafe modifications,...
1 months ago cs.CR cs.AI
PDF
Defense MEDIUM
Chen Chen, Yuchen Sun, Jiaxin Gao +4 more
Large language models (LLMs) are increasingly deployed in security-sensitive applications, yet remain vulnerable to backdoor attacks. However,...
Attack MEDIUM
Fengpeng Li, Kemou Li, Qizhou Wang +2 more
Concept erasure helps stop diffusion models (DMs) from generating harmful content; but current methods face robustness retention trade off....
1 months ago cs.LG cs.AI cs.CR
PDF
Survey MEDIUM
Yunlong Lyu, Yixuan Tang, Peng Chen +4 more
Modern AI-integrated IDEs are shifting from passive code completion to proactive Next Edit Suggestions (NES). Unlike traditional autocompletion, NES...
1 months ago cs.CR cs.HC
PDF
Benchmark MEDIUM
Haoyang Hu, Zhejun Jiang, Yueming Lyu +3 more
Retrieval-augmented generation (RAG) is increasingly deployed in real-world applications, where its reference-grounded design makes outputs appear...
1 months ago cs.CR cs.LG
PDF
Benchmark MEDIUM
Yi Liu, Zhihao Chen, Yanjun Zhang +5 more
Third-party agent skills extend LLM-based agents with instruction files and executable code that run on users' machines. Skills execute with user...
1 months ago cs.CR cs.AI cs.CL
PDF
Defense MEDIUM
Hema Karnam Surendrababu, Nithin Nagaraj
Machine Learning (ML) models, including Large Language Models (LLMs), are characterized by a range of system-level attributes such as security and...
Tool MEDIUM
Guowei Guan, Yurong Hao, Jiaming Zhang +6 more
Multimodal large language models (MLLMs) are pushing recommender systems (RecSys) toward content-grounded retrieval and ranking via cross-modal...
Benchmark MEDIUM
Navita Goyal, Hal Daumé
Model steering, which involves intervening on hidden representations at inference time, has emerged as a lightweight alternative to finetuning for...
1 months ago cs.LG cs.AI cs.CL
PDF
Benchmark MEDIUM
José Ramón Pareja Monturiol, Juliette Sinnott, Roger G. Melko +1 more
Machine learning in clinical settings must balance predictive accuracy, interpretability, and privacy. Models such as logistic regression (LR) offer...
1 months ago cs.LG cs.CR quant-ph
PDF
Attack MEDIUM
Tao Huang, Rui Wang, Xiaofei Liu +3 more
%Large vision-language models (LVLMs) have shown substantial advances in multimodal understanding and generation. However, when presented with...
Defense MEDIUM
Rohan Subramanian Thomas, Shikhar Shiromani, Abdullah Chaudhry +4 more
Prompt design significantly impacts the moral competence and safety alignment of large language models (LLMs), yet empirical comparisons remain...
1 months ago cs.AI cs.CL
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial