Zizhao Wang, Dingcheng Li, Vaishakh Keshava +4 more
Large Language Model (LLM) agents can leverage tools such as Google Search to complete complex tasks. However, this tool usage introduces the risk of...
Abrar Shahid, Ibteeker Mahir Ishum, AKM Tahmidul Haque +2 more
This paper presents a controlled study of adversarial reinforcement learning in network security through a custom OpenAI Gym environment that models...
Deterministic pseudo random number generators (PRNGs) used in generative artificial intelligence (GAI) models produce predictable patterns vulnerable...
7 months ago cs.LG cond-mat.mtrl-sci physics.data-an
PDF
As large language models (LLMs) advance, ensuring AI safety and alignment is paramount. One popular approach is prompt guards, lightweight mechanisms...
Large language models have gained widespread prominence, yet their vulnerability to prompt injection and other adversarial attacks remains a critical...
Generative models can generate photorealistic images at scale. This raises urgent concerns about the ability to detect synthetically generated images...
Firas Ben Hmida, Abderrahmen Amich, Ata Kaboudi +1 more
Deep neural networks (DNNs) are increasingly being deployed in high-stakes applications, from self-driving cars to biometric authentication. However,...
Marco Zimmerli, Andreas Plesner, Till Aczel +1 more
Deep neural networks remain vulnerable to adversarial examples despite advances in architectures and training paradigms. We investigate how training...
Large language models (LLMs) have become increasingly popular due to their ability to interact with unstructured content. As such, LLMs are now a key...