Survey MEDIUM
Chongyu Fan, Changsheng Wang, Yancheng Huang +2 more
Machine unlearning for large language models (LLMs) aims to remove undesired data, knowledge, and behaviors (e.g., for safety, privacy, or copyright)...
7 months ago cs.LG cs.CL
PDF
Benchmark MEDIUM
Shen Dong, Mingxuan Zhang, Pengfei He +4 more
Large Language Model (LLM)-based Multi-Agent Systems (MAS) have emerged as a powerful paradigm for tackling complex, multi-step tasks across diverse...
Tool MEDIUM
Muris Sladić, Veronica Valeros, Carlos Catania +1 more
There are very few SotA deception systems based on Large Language Models. The existing ones are limited only to simulating one type of service,...
7 months ago cs.CR cs.AI cs.CL
PDF
Benchmark MEDIUM
Riku Mochizuki, Shusuke Komatsu, Souta Noguchi +1 more
We analyze answers generated by generative engines (GEs) from the perspectives of citation publishers and the content-injection barrier, defined as...
7 months ago cs.CR cs.CL cs.IR
PDF
Attack MEDIUM
Tavish McDonald, Bo Lei, Stanislav Fort +2 more
Models are susceptible to adversarially out-of-distribution (OOD) data despite large training-compute investments into their robustification. Zaremba...
Attack MEDIUM
Tiancheng Xing, Jerry Li, Yixuan Du +1 more
Large language models (LLMs) are increasingly used as rerankers in information retrieval, yet their ranking behavior can be steered by small,...
7 months ago cs.CL cs.AI cs.IR
PDF
Benchmark MEDIUM
Zhiyuan Wei, Xiaoxuan Yang, Jing Sun +1 more
The increasing complexity of modern software systems exacerbates the prevalence of security vulnerabilities, posing risks of severe breaches and...
7 months ago cs.CR cs.AI
PDF
Benchmark MEDIUM
Weidi Luo, Qiming Zhang, Tianyu Lu +9 more
Computer-use agent (CUA) frameworks, powered by large language models (LLMs) or multimodal LLMs (MLLMs), are rapidly maturing as assistants that can...
Defense MEDIUM
Thusitha Dayaratne, Ngoc Duy Pham, Viet Vo +5 more
The quality and experience of mobile communication have significantly improved with the introduction of 5G, and these improvements are expected to...
7 months ago cs.CR cs.ET cs.LG
PDF
Benchmark MEDIUM
Ali Naseh, Anshuman Suri, Yuefeng Peng +3 more
Generative AI leaderboards are central to evaluating model capabilities, but remain vulnerable to manipulation. Among key adversarial objectives is...
7 months ago cs.LG cs.CR
PDF
Benchmark MEDIUM
Shadi Rahimian, Mario Fritz
Single nucleotide polymorphism (SNP) datasets are fundamental to genetic studies but pose significant privacy risks when shared. The correlation of...
7 months ago cs.LG cs.CR q-bio.GN
PDF
Benchmark MEDIUM
Mary Llewellyn, Annie Gray, Josh Collyer +1 more
Before adopting a new large language model (LLM) architecture, it is critical to understand vulnerabilities accurately. Existing evaluations can be...
7 months ago cs.CR cs.AI cs.CL
PDF
Tool MEDIUM
Yasod Ginige, Akila Niroshan, Sajal Jain +1 more
Penetration testing and vulnerability assessment are essential industry practices for safeguarding computer systems. As cyber threats grow in scale...
7 months ago cs.CR cs.AI
PDF
Survey MEDIUM
Cade Houston Kennedy, Amr Hilal, Morteza Momeni
With the growth of digital financial systems, robust security and privacy have become a concern for financial institutions. Even though traditional...
7 months ago cs.CR cs.AI
PDF
Attack MEDIUM
Zizhao Wang, Dingcheng Li, Vaishakh Keshava +4 more
Large Language Model (LLM) agents can leverage tools such as Google Search to complete complex tasks. However, this tool usage introduces the risk of...
7 months ago cs.LG cs.AI cs.CL
PDF
Benchmark MEDIUM
Yongan Yu, Xianda Du, Qingchen Hu +7 more
Historical archives on weather events are collections of enduring primary source records that offer rich, untapped narratives of how societies have...
7 months ago cs.CL cs.AI
PDF
Benchmark MEDIUM
Ruoxing Yang
Large language models (LLMs) such as ChatGPT have evolved into powerful and ubiquitous tools. Fine-tuning on small datasets allows LLMs to acquire...
7 months ago cs.LG cs.AI cs.CR
PDF
Benchmark MEDIUM
Punya Syon Pandey, Hai Son Le, Devansh Bhardwaj +2 more
Large language models (LLMs) are increasingly deployed in contexts where their failures can have direct sociopolitical consequences. Yet, existing...
7 months ago cs.CL cs.AI cs.LG
PDF
Defense MEDIUM
Shuai Zhao, Xinyi Wu, Shiqian Zhao +4 more
During fine-tuning, large language models (LLMs) are increasingly vulnerable to data-poisoning backdoor attacks, which compromise their reliability...
7 months ago cs.CR cs.AI cs.CL
PDF
Defense MEDIUM
Anindya Sundar Das, Kangjie Chen, Monowar Bhuyan
Pre-trained language models have achieved remarkable success across a wide range of natural language processing (NLP) tasks, particularly when...
7 months ago cs.CL cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial