Large Language Models (LLMs) have achieved remarkable success but remain highly susceptible to jailbreak attacks, in which adversarial prompts coerce...
LLM agents with tool access can discover and exploit security vulnerabilities. This is known. What is not known is which features of a system prompt...
Shouqiao Wang, Marcello Politi, Samuele Marro +1 more
As agentic systems move into real-world deployments, their decisions increasingly depend on external inputs such as retrieved content, tool outputs,...
Dimitris Mitropoulos, Nikolaos Alexopoulos, Georgios Alexopoulos +1 more
Security code reviews increasingly rely on systems integrating Large Language Models (LLMs), ranging from interactive assistants to autonomous agents...
AI-assisted code review is widely used to detect vulnerabilities before production release. Prior work shows that adversarial prompt manipulation can...
Pedro H. Barcha Correia, Ryan W. Achjian, Diego E. G. Caetano de Oliveira +5 more
The rapid advancement and widespread adoption of generative artificial intelligence (GenAI) and large language models (LLMs) has been accompanied by...
Strahinja Janjusevic, Anna Baron Garcia, Sohrob Kazerounian
Generative AI is reshaping offensive cybersecurity by enabling autonomous red team agents that can plan, execute, and adapt during penetration tests....