SHIELD: Classifier-Guided Prompting for Robust and Safer LVLMs
Juan Ren, Mark Dras, Usman Naseem
Large Vision-Language Models (LVLMs) unlock powerful multimodal reasoning but also expand the attack surface, particularly through adversarial inputs...
2,077+ academic papers on AI security, attacks, and defenses
Showing 301–320 of 355 papers
Clear filtersJuan Ren, Mark Dras, Usman Naseem
Large Vision-Language Models (LVLMs) unlock powerful multimodal reasoning but also expand the attack surface, particularly through adversarial inputs...
João A. Leite, Arnav Arora, Silvia Gargova +5 more
Large Language Models (LLMs) can generate human-like disinformation, yet their ability to personalise such content across languages and demographics...
Blazej Manczak, Eric Lin, Francisco Eiras +2 more
Large language models (LLMs) are rapidly transitioning into medical clinical use, yet their reliability under realistic, multi-turn interactions...
Lipeng He, Vasisht Duddu, N. Asokan
Chatbot providers (e.g., OpenAI) rely on tiered subscription schemes to generate revenue, offering basic models for free users, and advanced models...
Shuo Chen, Zonggen Li, Zhen Han +7 more
Deep Research (DR) agents built on Large Language Models (LLMs) can perform complex, multi-step research by decomposing tasks, retrieving online...
Dominik Schwarz
The security of Large Language Model (LLM) applications is fundamentally challenged by "form-first" attacks like prompt injection and jailbreaking,...
Sarah Ball, Andreas Haupt
Generative models are increasingly paired with safety classifiers that filter harmful or undesirable outputs. A common strategy is to fine-tune the...
Jiayu Ding, Lei Cui, Li Dong +2 more
Recent advances in Large Language Models (LLMs) show that extending the length of reasoning chains significantly improves performance on complex...
Mohan Zhang, Yihua Zhang, Jinghan Jia +3 more
Modern large reasoning models (LRMs) exhibit impressive multi-step problem-solving via chain-of-thought (CoT) reasoning. However, this iterative...
Shaolun Liu, Sina Marefat, Omar Tsai +4 more
GraphQL's flexible query model and nested data dependencies expose APIs to complex, context-dependent vulnerabilities that are difficult to uncover...
Zonghao Ying, Yangguang Shao, Jianle Gan +9 more
Large vision-language model (LVLM)-based web agents are emerging as powerful tools for automating complex online tasks. However, when deployed in...
Ines Altemir Marinas, Anastasiia Kucherenko, Alexander Sternfeld +1 more
The performance of Large Language Models (LLMs) is determined by their training data. Despite the proliferation of open-weight LLMs, access to LLM...
Yongding Tao, Tian Wang, Yihong Dong +4 more
Data contamination poses a significant threat to the reliable evaluation of Large Language Models (LLMs). This issue arises when benchmark samples...
Xiaonan Si, Meilin Zhu, Simeng Qin +7 more
Retrieval-augmented generation (RAG) systems enhance large language models (LLMs) with external knowledge but are vulnerable to corpus poisoning and...
Debeshee Das, Luca Beurer-Kellner, Marc Fischer +1 more
The increasing adoption of LLM agents with access to numerous tools and sensitive data significantly widens the attack surface for indirect prompt...
Eric Hanchen Jiang, Weixuan Ou, Run Liu +8 more
Safety alignment of large language models currently faces a central challenge: existing alignment techniques often prioritize mitigating responses to...
Shen Dong, Mingxuan Zhang, Pengfei He +4 more
Large Language Model (LLM)-based Multi-Agent Systems (MAS) have emerged as a powerful paradigm for tackling complex, multi-step tasks across diverse...
Riku Mochizuki, Shusuke Komatsu, Souta Noguchi +1 more
We analyze answers generated by generative engines (GEs) from the perspectives of citation publishers and the content-injection barrier, defined as...
Zhiyuan Wei, Xiaoxuan Yang, Jing Sun +1 more
The increasing complexity of modern software systems exacerbates the prevalence of security vulnerabilities, posing risks of severe breaches and...
Weidi Luo, Qiming Zhang, Tianyu Lu +9 more
Computer-use agent (CUA) frameworks, powered by large language models (LLMs) or multimodal LLMs (MLLMs), are rapidly maturing as assistants that can...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial