Benchmark MEDIUM
Yixuan Liu, Xinlei Li, Yi Li
Phishing attacks in Web3 ecosystems are increasingly sophisticated, exploiting deceptive contract logic, malicious frontend scripts, and token...
Attack MEDIUM
Yushi Yang, Shreyansh Padarha, Andrew Lee +1 more
Agentic reinforcement learning (RL) trains large language models to autonomously call tools during reasoning, with search as the most common...
Tool MEDIUM
Rishi Jha, Harold Triedman, Justin Wagle +1 more
Control-flow hijacking attacks manipulate orchestration mechanisms in multi-agent systems into performing unsafe actions that compromise the system...
6 months ago cs.LG cs.CR eess.SY
PDF
Defense MEDIUM
Runlin Lei, Lu Yi, Mingguo He +4 more
While Graph Neural Networks (GNNs) and Large Language Models (LLMs) are powerful approaches for learning on Text-Attributed Graphs (TAGs), a...
Attack MEDIUM
Elias Hossain, Swayamjit Saha, Somshubhra Roy +1 more
Even when prompts and parameters are secured, transformer language models remain vulnerable because their key-value (KV) cache during inference...
6 months ago cs.CR cs.AI
PDF
Defense MEDIUM
Qiusi Zhan, Angeline Budiman-Chan, Abdelrahman Zayed +3 more
Large language model (LLM) based search agents iteratively generate queries, retrieve external information, and reason to answer open-domain...
Defense MEDIUM
Qiusi Zhan, Angeline Budiman-Chan, Abdelrahman Zayed +3 more
Large language model (LLM) based search agents iteratively generate queries, retrieve external information, and reason to answer open-domain...
Defense MEDIUM
Bo-Han Feng, Chien-Feng Liu, Yu-Hsuan Li Liang +9 more
Large audio-language models (LALMs) extend text-based LLMs with auditory understanding, offering new opportunities for multimodal applications. While...
6 months ago cs.SD cs.AI cs.CL
PDF
Tool MEDIUM
Yue Liu, Zhenchang Xing, Shidong Pan +1 more
In recent years, the AI wave has grown rapidly in software development. Even novice developers can now design and generate complex...
6 months ago cs.SE cs.CR
PDF
Attack MEDIUM
Jie Zhang, Meng Ding, Yang Liu +2 more
We present a novel approach for attacking black-box large language models (LLMs) by exploiting their ability to express confidence in natural...
6 months ago cs.CR cs.LG
PDF
Attack MEDIUM
Asmita Mohanty, Gezheng Kang, Lei Gao +1 more
Large Language Models (LLMs) have demonstrated strong performance across diverse tasks, but fine-tuning them typically relies on cloud-based,...
6 months ago cs.CR cs.LG
PDF
Benchmark MEDIUM
Shivam Ratnakar, Sanjay Raghavendra
Integration of Large Language Models with search/retrieval engines has become ubiquitous, yet these systems harbor a critical vulnerability that...
6 months ago cs.CL cs.AI
PDF
Tool MEDIUM
Xiaofan Li, Xing Gao
The Model Context Protocol (MCP) is an emerging open standard that enables AI-powered applications to interact with external tools through structured...
6 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
David Peer, Sebastian Stabinger
Large Language Models (LLMs) have demonstrated impressive capabilities, yet their deployment in high-stakes domains is hindered by inherent...
6 months ago cs.CL cs.AI
PDF
Benchmark MEDIUM
Shuai Li, Kejiang Chen, Jun Jiang +5 more
Large Language Models (LLMs) have demonstrated remarkable capabilities, but their training requires extensive data and computational resources,...
Attack MEDIUM
Sarah Egler, John Schulman, Nicholas Carlini
Large Language Model (LLM) providers expose fine-tuning APIs that let end users fine-tune their frontier LLMs. Unfortunately, it has been shown that...
6 months ago cs.CR cs.AI
PDF
Defense MEDIUM
Yang Feng, Xudong Pan
Malicious agents pose significant threats to the reliability and decision-making capabilities of Multi-Agent Systems (MAS) powered by Large Language...
6 months ago cs.CR cs.AI
PDF
Defense MEDIUM
Eduard Andrei Cristea, Petter Molnes, Jingyue Li
Malicious software attacks are having an increasingly significant economic impact. Commercial malware detection software can be costly, and tools...
6 months ago cs.CR cs.SE
PDF
Defense MEDIUM
Yuexiao Liu, Lijun Li, Xingjun Wang +1 more
Recent advancements in Reinforcement Learning with Verifiable Rewards (RLVR) have gained significant attention due to their objective and verifiable...
Survey MEDIUM
Hanbin Hong, Shuya Feng, Nima Naderloui +6 more
Large Language Models (LLMs) have rapidly become integral to real-world applications, powering services across diverse sectors. However, their...
6 months ago cs.CR cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial