Paper 2510.13893v2

Guarding the Guardrails: A Taxonomy-Driven Approach to Jailbreak Detection

poisoning. Second, we analyzed the data collected from the challenge to examine the prevalence and success rates of different attack types, providing insights into how specific jailbreak strategies exploit model

high relevance survey
Paper 2510.14381v2

Are My Optimized Prompts Compromised? Exploring Vulnerabilities of LLM-based Optimizers

systematic analysis of poisoning risks in LLM-based prompt optimization. Using HarmBench, we find systems are substantially more vulnerable to manipulated feedback than to query poisoning alone: feedback-based attacks

medium relevance attack
Paper 2509.21011v1

Automatic Red Teaming LLM-based Agents with Model Context Protocol Tools

large language models (LLMs) has led to the wide application of LLM-based agents in various domains. To standardize interactions between LLM-based agents and their environments, model context protocol

high relevance tool
Paper 2512.10637v2

Adaptive Intrusion Detection System Leveraging Dynamic Neural Models with Adversarial Learning for 5G/6G Networks

network security by providing robust, real-time threat detection and response capabilities. Unlike conventional models, which require costly retraining to update knowledge, the proposed framework integrates incremental learning algorithms, reducing

medium relevance attack
Paper 2511.14989v2

Critical Evaluation of Quantum Machine Learning for Adversarial Robustness

three threat models-black-box, gray-box, and white-box. We implement representative attacks in each category, including label-flipping for black-box, QUID encoder-level data poisoning for gray

medium relevance benchmark
Paper 2511.12936v1

Privacy-Preserving Federated Learning from Partial Decryption Verifiable Threshold Multi-Client Functional Encryption

cooperate to train the model without directly exchanging their own private data, but the gradient leakage problem still threatens the privacy security and model integrity. Although the existing scheme uses

medium relevance benchmark
Paper 2512.09742v1

Weird Generalization and Inductive Backdoors: New Ways to Corrupt LLMs

contexts can dramatically shift behavior outside those contexts. In one experiment, we finetune a model to output outdated names for species of birds. This causes it to behave

medium relevance benchmark
Paper 2603.02849v1

DSBA: Dynamic Stealthy Backdoor Attack with Collaborative Optimization in Self-Supervised Learning

generalization capabilities, and its potential for privacy preservation. However, recent research reveals that SSL models are also vulnerable to backdoor attacks. Existing backdoor attack methods in the SSL context commonly

high relevance attack
Paper 2602.19555v1

Agentic AI as a Cybersecurity Attack Surface: Threats, Exploits, and Defenses in Runtime Supply Chains

Agentic systems built on large language models (LLMs) extend beyond text generation to autonomously retrieve information and invoke tools. This runtime execution model shifts the attack surface from build-time

high relevance attack
Paper 2601.11207v1

LoRA as Oracle

Existing defenses for backdoor detection and membership inference typically require access to clean reference models, extensive retraining, or strong assumptions about the attack mechanism. In this work, we introduce

medium relevance attack
Paper 2603.07835v1

DistillGuard: Evaluating Defenses Against LLM Knowledge Distillation

Knowledge distillation from proprietary LLM APIs poses a growing threat to model providers, yet defenses against this attack remain fragmented and unevaluated. We present DistillGuard, a framework for systematically evaluating

medium relevance defense
Paper 2511.06212v1

RAG-targeted Adversarial Attack on LLM-based Threat Detection and Mitigation Framework

Artificial Intelligence has become a valuable solution in securing IoT networks, with Large Language Models (LLMs) enabling automated attack behavior analysis and mitigation suggestion in Network Intrusion Detection Systems (NIDS

high relevance tool
Paper 2512.19297v1

Causal-Guided Detoxify Backdoor Attack of Open-Weight LoRA Models

Backdoor Attack (CBA), a novel backdoor attack framework specifically designed for open-weight LoRA models. CBA operates without access to original training data and achieves high stealth through

high relevance attack
Paper 2603.03108v1

RAIN: Secure and Robust Aggregation under Shuffle Model of Differential Privacy

achieving robustness under adversarial behavior remains challenging. Modern systems increasingly adopt the shuffle model of differential privacy (Shuffle-DP) to locally perturb client updates and globally anonymize them via shuffling

medium relevance benchmark
Paper 2512.13501v1

Behavior-Aware and Generalizable Defense Against Black-Box Adversarial Attacks for ML-Based IDS

often fall short in practice. Most are tailored to specific attack types, require internal model access, or rely on static mechanisms that fail to generalize across evolving attack strategies. Furthermore

high relevance attack
Paper 2512.23307v1

RobustMask: Certified Robustness against Adversarial Neural Ranking Attack via Randomized Masking

Neural ranking models have achieved remarkable progress and are now widely deployed in real-world applications such as Retrieval-Augmented Generation (RAG). However, like other neural architectures, they remain vulnerable

high relevance attack
Paper 2603.21654v1

Towards Secure Retrieval-Augmented Generation: A Comprehensive Review of Threats, Defenses and Benchmarks

Augmented Generation (RAG) significantly mitigates the hallucinations and domain knowledge deficiency in large language models by incorporating external knowledge bases. However, the multi-module architecture of RAG introduces complex system

medium relevance survey
Paper 2602.02615v1

TinyGuard:A lightweight Byzantine Defense for Resource-Constrained Federated Learning via Statistical Update Fingerprints

label poisoning. Against adaptive white-box adversaries, Pareto frontier analysis across four orders of magnitude confirms that attackers cannot simultaneously evade detection and achieve effective poisoning, features we term statistical

medium relevance defense
Paper 2602.17973v1

PenTiDef: Enhancing Privacy and Robustness in Decentralized Federated Intrusion Detection Systems against Poisoning Attacks

Systems (IDS) introduces new challenges related to data privacy, centralized coordination, and susceptibility to poisoning attacks. While significant research has focused on protecting traditional FL-IDS with centralized aggregation servers

high relevance tool
Paper 2512.15799v1

Cybercrime and Computer Forensics in Epoch of Artificial Intelligence in India

while Machine Learning offers high accuracy in pattern recognition, it introduces vulnerabilities regarding data poisoning and algorithmic bias. Findings highlight a critical tension between the Act's data minimization principles

low relevance benchmark
Previous Page 8 of 10 Next