Attack MEDIUM
Kaiwen Wang, Xiaolin Chang, Yuehan Dong +1 more
Secure comparison is a fundamental primitive in multi-party computation, supporting privacy-preserving applications such as machine learning and data...
Attack HIGH
Nadav Kadvil, Malak Fares, Ayellet Tal
Large Vision-Language Models (LVLMs) can be vulnerable to adversarial images that subtly bias their outputs toward plausible yet incorrect responses....
Attack HIGH
Xiaochong Jiang, Shiqi Yang, Wenting Yang +2 more
Agentic systems built on large language models (LLMs) extend beyond text generation to autonomously retrieve information and invoke tools. This...
2 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Lei Ba, Qinbin Li, Songze Li
LLM-based code interpreter agents are increasingly deployed in critical workflows, yet their robustness against risks introduced by their code...
Tool HIGH
Xingyu Shen, Tommy Duong, Xiaodong An +6 more
Age estimation systems are increasingly deployed as gatekeepers for age-restricted online content, yet their robustness to cosmetic modifications has...
2 months ago cs.CV cs.CR cs.LG
PDF
Benchmark MEDIUM
Jingwei Shi, Xinxiang Yin, Jing Huang +2 more
The evaluation of Large Language Models (LLMs) for code generation relies heavily on the quality and robustness of test cases. However, existing...
2 months ago cs.SE cs.AI cs.CR
PDF
Survey HIGH
Kunal Mukherjee
Trusted Execution Environments (TEEs) (e.g., Intel SGX and ArmTrustZone) aim to protect sensitive computation from a compromised operating system,...
2 months ago cs.CR cs.AI
PDF
Attack HIGH
Amirhossein Farzam, Majid Behabahani, Mani Malek +2 more
Large language models (LLMs) remain vulnerable to jailbreak prompts that are fluent and semantically coherent, and therefore difficult to detect with...
Attack HIGH
Charles Ye, Jasmine Cui, Dylan Hadfield-Menell
Language models remain vulnerable to prompt injection attacks despite extensive safety training. We trace this failure to role confusion: models...
2 months ago cs.CL cs.AI cs.CR
PDF
Attack HIGH
Charles Ye, Jasmine Cui, Dylan Hadfield-Menell
Language models remain vulnerable to prompt injection attacks despite extensive safety training. We trace this failure to role confusion: models...
2 months ago cs.CL cs.AI cs.CR
PDF
Tool MEDIUM
Florin Adrian Chitan
The proliferation of autonomous AI agents capable of executing real-world actions - filesystem operations, API calls, database modifications,...
2 months ago cs.AI cs.CR
PDF
Attack HIGH
Sieun Kim, Yeeun Jo, Sungmin Na +5 more
Red-teaming, where adversarial prompts are crafted to expose harmful behaviors and assess risks, offers a dynamic approach to surfacing underlying...
Survey MEDIUM
Kiarash Ahi, Vaibhav Agrawal, Saeed Valizadeh
Large Language Models (LLMs) & Generative AI are transforming cybersecurity, enabling both advanced defenses and new attacks. Organizations now use...
Tool MEDIUM
Emmanuel Bamidele
Long-running LLM agents require persistent memory to preserve state across interactions, yet most deployed systems manage memory with age-based...
2 months ago cs.DC cs.AI cs.LG
PDF
Attack HIGH
Shenyang Chen, Liuwan Zhu
Standard evaluations of backdoor attacks on text-to-image (T2I) models primarily measure trigger activation and visual fidelity. We challenge this...
2 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Abdullah Caglar Oksuz, Anisa Halimi, Erman Ayday
Membership inference attacks (MIAs) threaten the privacy of machine learning models by revealing whether a specific data point was used during...
2 months ago cs.LG cs.CR
PDF
Defense MEDIUM
Chun Yan Ryan Kan, Tommy Tran, Vedant Yadav +4 more
Defending LLMs against adversarial jailbreak attacks remains an open challenge. Existing defenses rely on binary classifiers that fail when...
2 months ago cs.CR cs.AI cs.CL
PDF
Attack HIGH
Zafir Shamsi, Nikhil Chekuru, Zachary Guzman +1 more
Large Language Models (LLMs) are increasingly integrated into high-stakes applications, making robust safety guarantees a central practical and...
2 months ago cs.CL cs.AI
PDF
Benchmark LOW
Martin Bertran, Riccardo Fogliato, Zhiwei Steven Wu
Empirical conclusions depend not only on data but on analytic decisions made throughout the research process. Many-analyst studies have quantified...
2 months ago cs.AI cs.LG
PDF
Benchmark HIGH
Mirae Kim, Seonghun Jeong, Youngjun Kwak
Jailbreaking poses a significant risk to the deployment of Large Language Models (LLMs) and Vision Language Models (VLMs). VLMs are particularly...
2 months ago cs.CL cs.AI cs.DB
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial