I Can't Believe It's Not a Valid Exploit
Derin Gezgin, Amartya Das, Shinhae Kim +3 more
Recently Large Language Models (LLMs) have been used in security vulnerability detection tasks including generating proof-of-concept (PoC) exploits....
2,077+ academic papers on AI security, attacks, and defenses
Showing 221–240 of 795 papers
Clear filtersDerin Gezgin, Amartya Das, Shinhae Kim +3 more
Recently Large Language Models (LLMs) have been used in security vulnerability detection tasks including generating proof-of-concept (PoC) exploits....
Hoang Long Do, Nasrin Sohrabi, Muneeb Ul Hassan
Large language models (LLMs) have been widely adopted in modern software development lifecycles, where they are increasingly used to automate and...
Shutong Fan, Lan Zhang, Xiaoyong Yuan
Most adversarial threats in artificial intelligence target the computational behavior of models rather than the humans who rely on them. Yet modern...
Xilong Wang, Yinuo Liu, Zhun Wang +2 more
Prompt injection attacks manipulate webpage content to cause web agents to execute attacker-specified tasks instead of the user's intended ones....
Andrew Draganov, Tolga H. Dur, Anandmayi Bhongade +1 more
We present a data poisoning attack -- Phantom Transfer -- with the property that, even if you know precisely how the poison was placed into an...
Chen Xiong, Zhiyuan He, Pin-Yu Chen +2 more
Activation steering is a practical post-training model alignment technique to enhance the utility of Large Language Models (LLMs). Prior to deploying...
Mengxuan Wang, Yuxin Chen, Gang Xu +3 more
Vision language models (VLMs) extend the reasoning capabilities of large language models (LLMs) to cross-modal settings, yet remain highly vulnerable...
Hicham Eddoubi, Umar Faruk Abdullahi, Fadi Hassan
Large Language Models (LLMs) have seen widespread adoption across multiple domains, creating an urgent need for robust safety alignment mechanisms....
Matthew P. Lad, Louisa Conwill, Megan Levis Scheirer
With the rapid growth of Large Language Models (LLMs), criticism of their societal impact has also grown. Work in Responsible AI (RAI) has focused on...
Blake Bullwinkel, Giorgio Severi, Keegan Hines +3 more
Detecting whether a model has been poisoned is a longstanding problem in AI security. In this work, we present a practical scanner for identifying...
Xiaozuo Shen, Yifei Cai, Rui Ning +2 more
The widespread adoption of Vision Transformers (ViTs) elevates supply-chain risk on third-party model hubs, where an adversary can implant backdoors...
Nirab Hossain, Pablo Moriano
Modern vehicles rely on electronic control units (ECUs) interconnected through the Controller Area Network (CAN), making in-vehicle communication a...
Patrick Cooper, Alireza Nadali, Ashutosh Trivedi +1 more
Large language models (LLMs) are known to exhibit brittle behavior under adversarial prompts and jailbreak attacks, even after extensive alignment...
Samuel Nellessen, Tal Kachman
The evolution of large language models into autonomous agents introduces adversarial failures that exploit legitimate tool privileges, transforming...
Ching-Yun Ko, Pin-Yu Chen
Modern artificial intelligence (AI) models are deployed on inference engines to optimize runtime efficiency and resource allocation, particularly for...
Pengfei He, Ash Fox, Lesly Miculicich +7 more
Large language models (LLMs) have shown promise in assisting cybersecurity tasks, yet existing approaches struggle with automatic vulnerability...
Jiayao Wang, Yang Song, Zhendong Zhao +5 more
Federated self-supervised learning (FSSL) enables collaborative training of self-supervised representation models without sharing raw unlabeled data....
Mingrui Liu, Sixiao Zhang, Cheng Long +1 more
Large Language Models (LLMs) are increasingly vulnerable to Prompt Injection (PI) attacks, where adversarial instructions hidden within retrieved...
Pengyu Li, Lingling Zhang, Zhitao Gao +5 more
While Large Language Models (LLMs) have achieved remarkable capabilities, they unintentionally memorize sensitive data, posing critical privacy and...
Seyed Mohammad Hadi Hosseini, Amir Najafi, Mahdieh Soleymani Baghshah
Bandit algorithms have recently emerged as a powerful tool for evaluating machine learning models, including generative image models and large...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial