Attack MEDIUM
Han Liu, Shanghao Shi, Yevgeniy Vorobeychik +2 more
Low-Rank Adaptation (LoRA), which leverages the insight that model updates typically reside in a low-dimensional space, has significantly improved...
1 weeks ago cs.LG cs.CR
PDF
Attack MEDIUM
David Fernandez, Pedram MohajerAnsari, Amir Salarpour +1 more
Vision-language models (VLMs) are increasingly used in autonomous driving because they combine visual perception with language-based reasoning,...
1 weeks ago cs.CV cs.CR cs.LG
PDF
Attack MEDIUM
Mahshid Rezakhani, Nowfel Mashnoor, Kimia Azar +1 more
As large language models (LLMs) are increasingly fine-tuned for hardware tasks like RTL code generation, the scarcity of high-quality datasets often...
1 weeks ago cs.CR cs.AR
PDF
Attack MEDIUM
Wenhao Lan, Shan Li, Junbin Yang +2 more
Safety-aligned language models must refuse harmful requests without collapsing into broad over-refusal, but the training-time mechanisms behind this...
1 weeks ago cs.LG cs.CL cs.CR
PDF
Attack MEDIUM
Hanna Foerster, Ilia Shumailov, Cheng Zhang +3 more
Dynamic quantization emerged as a practical approach to increase the utilization and efficiency of the machine learning serving flow. Unlike static...
1 weeks ago cs.CR cs.LG
PDF
Attack MEDIUM
Mengnan Zhao, Lihe Zhang, Tianhang Zheng +2 more
Fast Adversarial Training (FAT) has attracted significant attention due to its efficiency in enhancing neural network robustness against adversarial...
2 weeks ago cs.LG cs.AI cs.CR
PDF
Attack MEDIUM
Mengnan Zhao, Lihe Zhang, Bo Wang +3 more
Fast Adversarial Training (FAT) has proven effective in enhancing model robustness by encouraging networks to learn perturbation-invariant...
2 weeks ago cs.LG cs.CR
PDF
Attack MEDIUM
Irti Haq, Belén Saldías
As state-of-the-art Large Language Models (LLMs) have become ubiquitous, ensuring equitable performance across diverse demographics is critical....
2 weeks ago cs.CY cs.AI cs.CL
PDF
Attack MEDIUM
Abhijit Talluri
Adversarial robustness evaluation underpins every claim of trustworthy ML deployment, yet the field suffers from fragmented protocols and undetected...
2 weeks ago cs.CR cs.LG
PDF
Attack MEDIUM
Ruixuan Liu, David Evans, Li Xiong
Indistinguishability properties such as differential privacy bounds or low empirically measured membership inference are widely treated as proxies to...
3 weeks ago cs.CR cs.CL cs.LG
PDF
Attack MEDIUM
Jianming Tong, Hanshen Xiao, Krishna Kumar Nair +5 more
Multi-user virtual reality enables immersive interaction. However, rendering avatars for numerous participants on each headset incurs prohibitive...
3 weeks ago cs.CR cs.AR cs.CV
PDF
Attack MEDIUM
Xuanli He, Bilgehan Sel, Faizan Ali +3 more
Large Language Models (LLMs) are increasingly exposed to adaptive jailbreaking, particularly in high-stakes Chemical, Biological, Radiological, and...
3 weeks ago cs.CL cs.CR
PDF
Attack MEDIUM
Firas Ben Hmida, Philemon Hailemariam, Kashif Ali Khan +1 more
Deep neural networks (DNNs) remain largely opaque at inference time, limiting our ability to detect and diagnose malicious input manipulations such...
Attack MEDIUM
Pavel Chizhov, Egor Bogomolov, Ivan P. Yamshchikov
Efficiency and safety of Large Language Models (LLMs), among other factors, rely on the quality of tokenization. A good tokenizer not only improves...
Attack MEDIUM
Shaopeng Fu, Di Wang
Adversarial training (AT) is an effective defense for large language models (LLMs) against jailbreak attacks, but performing AT on LLMs is costly. To...
4 weeks ago cs.LG cs.CR stat.ML
PDF
Attack MEDIUM
Anasuya Chattopadhyay, Daniel Reti, Hans D. Schotten
Cloud networks increasingly rely on machine learning based Network Intrusion Detection Systems to defend against evolving cyber threats. However,...
4 weeks ago cs.LG cs.CR
PDF
Attack MEDIUM
Vladimir A. Mazin, Mikhail A. Zorin, Dmitrii S. Korzh +3 more
Passwords still remain a dominant authentication method, yet their security is routinely subverted by predictable user choices and large-scale...
4 weeks ago cs.CR cs.AI
PDF
Attack MEDIUM
Hongru Song, Yu-An Liu, Ruqing Zhang +4 more
Retrieval-augmented generation (RAG) enhances large language model (LLM) reasoning by retrieving external documents, but also opens up new attack...
Attack MEDIUM
Anes Abdennebi, Nadjia Kara, Laaziz Lahlou
The applications of Generative Artificial Intelligence (GenAI) and their intersections with data-driven fields, such as healthcare, finance,...
4 weeks ago cs.CR cs.AI
PDF
Attack MEDIUM
Shuhao Zhang, Yuli Chen, Jiale Han +2 more
Watermarking provides a critical safeguard for large language model (LLM) services by facilitating the detection of LLM-generated text....
1 months ago cs.CR cs.AI
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial