Understanding the Dynamics of Demonstration Conflict in In-Context Learning
Difan Jiao, Di Wang, Lijie Hu
In-context learning enables large language models to perform novel tasks through few-shot demonstrations. However, demonstrations per se can...
2,077+ academic papers on AI security, attacks, and defenses
Showing 241–260 of 341 papers
Clear filtersDifan Jiao, Di Wang, Lijie Hu
In-context learning enables large language models to perform novel tasks through few-shot demonstrations. However, demonstrations per se can...
Achyutha Menon, Magnus Saebo, Tyler Crosse +3 more
The accelerating adoption of language models (LMs) as agents for deployment in long-context tasks motivates a thorough understanding of goal drift:...
Aradhye Agarwal, Gurdit Siyan, Yash Pandya +3 more
Agentic language models operate in a fundamentally different safety regime than chat models: they must plan, call tools, and execute long-horizon...
Romina Omidi, Yun Dong, Binghui Wang
Google's SynthID-Text, the first ever production-ready generative watermark system for large language model, designs a novel Tournament-based method...
Yuhang Li, Yajie Wang, Xiangyun Tang +3 more
Secure aggregation is a foundational building block of privacy-preserving learning, yet achieving robustness under adversarial behavior remains...
Zhi Xu, Jiaqi Li, Xiaotong Zhang +2 more
Large language models (LLMs) have achieved remarkable success across diverse applications but remain vulnerable to jailbreak attacks, where attackers...
Pearl Mody, Mihir Panchal, Rishit Kar +2 more
Large language model (LLM) agents are increasingly deployed in long running workflows, where they must preserve user and task state across many...
Edouard Lansiaux
Federated Learning (FL) enables collaborative training of medical AI models across hospitals without centralizing patient data. However, the exchange...
Peter Horvath, Ilia Shumailov, Lukasz Chmielewski +2 more
The multi-million dollar investment required for modern machine learning (ML) has made large ML models a prime target for theft. In response, the...
Jiayao Wang, Mohammad Maruf Hasan, Yiping Zhang +5 more
Self-Supervised Learning (SSL) has emerged as a significant paradigm in representation learning thanks to its ability to learn without extensive...
Junjie Chu, Xinyue Shen, Ye Leng +3 more
The rapid growth of research in LLM safety makes it hard to track all advances. Benchmarks are therefore crucial for capturing key trends and...
Hongduan Tian, Xiao Feng, Ziyuan Zhao +3 more
Large language models (LLMs) have recently demonstrated impressive capabilities in reasoning tasks. Currently, mainstream LLM reasoning frameworks...
Shuyi Zhou, Zeen Song, Wenwen Qiang +4 more
Large Language Models remain vulnerable to adversarial prefix attacks (e.g., ``Sure, here is'') despite robust standard safety. We diagnose this...
Zixuan Xu, Tiancheng He, Huahui Yi +7 more
Vision-language models remain susceptible to multimodal jailbreaks and over-refusal because safety hinges on both visual evidence and user intent,...
Minseok Choi, Dongjin Kim, Seungbin Yang +5 more
With the growing deployment of large language models (LLMs) in real-world applications, establishing robust safety guardrails to moderate their...
Zhongxi Wang, Yueqian Lin, Jingyang Zhang +2 more
Safety evaluation and red-teaming of large language models remain predominantly text-centric, and existing frameworks lack the infrastructure to...
Bhanu Pallakonda, Mikkel Hindsbo, Sina Ehsani +1 more
The proliferation of open-weight Large Language Models (LLMs) has democratized agentic AI, yet fine-tuned weights are frequently shared and adopted...
Sami Abuzakuk, Lucas Crijns, Anne-Marie Kermarrec +2 more
Infrastructure as code (IaC) tools automate cloud provisioning but verifying that deployed systems remain consistent with the IaC specifications...
Tatiana Chakravorti, Pranav Narayanan Venkit, Sourojit Ghosh +1 more
Generative AI tools are increasingly entering academic peer review workflows, raising questions about fairness, accountability, and the legitimacy of...
Nancy Lau, Louis Sloot, Jyoutir Raj +6 more
Large language models (LLMs) are increasingly being deployed as software engineering agents that autonomously contribute to repositories. A major...
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act), and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial