Tracking Capabilities for Safer Agents
challenges: agents might leak private information, cause unintended side effects, or be manipulated through prompt injection. To address these challenges, we propose to put the agent in a programming-language
From Goals to Aspects, Revisited: An NFR Pattern Language for Agentic AI Systems
patterns address agent-specific crosscutting concerns absent from traditional AOP literature: tool-scope sandboxing, prompt injection detection, token budget management, and action audit trails. We extend the V-graph model
LiaisonAgent: An Multi-Agent Framework for Autonomous Risk Investigation and Governance
Furthermore, the system exhibits significant resilience against out-of-distribution noise and adversarial prompt injections, while achieving a 92.7% reduction in manual investigation overhead
Reverse CAPTCHA: Evaluating LLM Susceptibility to Invisible Unicode Instruction Injection
statistically significant (p < 0.05, Bonferroni-corrected). These results highlight an underexplored attack surface for prompt injection via invisible Unicode payloads
SoK: Agentic Skills -- Beyond Tool Use in LLM Agents
analyze the security and governance implications of skill-based agents, covering supply-chain risks, prompt injection via skill payloads, and trust-tiered execution, grounded by a case study
AMV-L: Lifecycle-Managed Agent Memory for Tail-Latency Control in Long-Running LLM Systems
running workloads against two baselines: TTL and an LRU working-set policy, with fixed prompt-injection caps. Relative to TTL, AMV-L improves throughput by 3.1x and reduces latency
Policy Compiler for Secure Agentic Systems
specific restructuring required. We evaluate PCAS on three case studies: information flow policies for prompt injection defense, approval workflows in a multi-agent pharmacovigilance system, and organizational policies for customer
OMNI-LEAK: Orchestrator Multi-Agent Network Induced Data Leakage
OMNI-LEAK, that compromises several agents to leak sensitive data through a single indirect prompt injection, even in the presence of data access control. We report the susceptibility of frontier
Peak + Accumulation: A Proxy-Level Scoring Formula for Multi-Turn LLM Attack Detection
Multi-turn prompt injection attacks distribute malicious intent across multiple conversation turns, exploiting the assumption that each turn is evaluated independently. While single-turn detection has been extensively studied
Blind Gods and Broken Screens: Architecting a Secure, Intent-Centric Mobile Agent Operating System
Action Execution - revealing critical flaws such as fake App identity, visual spoofing, indirect prompt injection, and unauthorized privilege escalation stemming from a reliance on unstructured visual data. To address these
Autonomous Action Runtime Management(AARM):A System Specification for Securing AI-Driven Actions at Runtime
records tamper-evident receipts for forensic reconstruction. We formalize a threat model addressing prompt injection, confused deputy attacks, data exfiltration, and intent drift. We introduce an action classification framework distinguishing
When Actions Go Off-Task: Detecting and Correcting Misaligned Actions in Computer-Use Agents
user's original intent. Such misaligned actions may arise from external attacks (e.g., indirect prompt injection) or from internal limitations (e.g., erroneous reasoning). They not only expose CUAs to safety
When the Model Said 'No Comment', We Knew Helpfulness Was Dead, Honesty Was Alive, and Safety Was Terrified
experts. To resolve this, we propose AlignX, a two-stage framework. Stage 1 uses prompt-injected fine-tuning to extract axis-specific task features, mitigating catastrophic forgetting. Stage 2 deploys
Agents in the Wild: Safety, Society, and the Illusion of Sociality on Moltbook
content touches safety-related themes; social engineering (31.9% of attacks) far outperforms prompt injection (3.7%), and adversarial posts receive 6x higher engagement than normal content. (3) The Illusion of Sociality
vLLM Hook v0: A Plug-in for Programming Model Internals on vLLM
core functions of vLLM Hook, in version 0, we demonstrate 3 use cases including prompt injection detection, enhanced retrieval-augmented retrieval (RAG), and activation steering. Finally, we welcome the community
Human Society-Inspired Approaches to Agentic AI Security: The 4C Framework
Although recent work has strengthened defenses against model and pipeline level vulnerabilities such as prompt injection, data poisoning, and tool misuse, these system centric approaches may fail to capture risks
SMCP: Secure Model Context Protocol
security and privacy challenges. These include risks such as unauthorized access, tool poisoning, prompt injection, privilege escalation, and supply chain attacks, any of which can impact different parts
Machine-Assisted Grading of Nationwide School-Leaving Essay Exams with LLMs and Statistical NLP
raters and tends to fall within the human scoring range. We also evaluate bias, prompt injection risks, and LLMs as essay writers. These findings demonstrate that a principled, rubric-driven
From Biased Chatbots to Biased Agents: Examining Role Assignment Effects on LLM Agent Robustness
shifts appear across task types and model architectures, indicating that persona conditioning and simple prompt injections can distort an agent's decision-making reliability. Our findings reveal an overlooked vulnerability
MirrorGuard: Toward Secure Computer-Use Agents via Simulation-to-Real Reasoning Correction
perform complex tasks. This autonomy introduces serious security risks: malicious instructions or visual prompt injections can trigger unsafe reasoning and cause harmful system-level actions. Existing defenses, such as detection