Tool MEDIUM
Dalal Alharthi, Ivan Roberto Kawaminami Garcia
Large Language Models (LLMs) have gained prominence in domains including cloud security and forensics. Yet cloud forensic investigations still rely...
5 months ago cs.CR cs.AI cs.LG
PDF
Attack MEDIUM
Dalal Alharthi, Ivan Roberto Kawaminami Garcia
Large language models have gained widespread prominence, yet their vulnerability to prompt injection and other adversarial attacks remains a critical...
5 months ago cs.CR cs.AI cs.LG
PDF
Benchmark HIGH
Haoran Xi, Minghao Shao, Brendan Dolan-Gavitt +2 more
Large language models show promise for vulnerability discovery, yet prevailing methods inspect code in isolation, struggle with long contexts, and...
5 months ago cs.SE cs.CR cs.LG
PDF
Attack MEDIUM
Samar Fares, Nurbek Tastan, Noor Hussein +1 more
Generative models can generate photorealistic images at scale. This raises urgent concerns about the ability to detect synthetically generated images...
5 months ago cs.CV cs.CR cs.LG
PDF
Benchmark MEDIUM
Ehsan Aghaei, Sarthak Jain, Prashanth Arun +1 more
Effective analysis of cybersecurity and threat intelligence data demands language models that can interpret specialized terminology, complex document...
5 months ago cs.CR cs.AI cs.LG
PDF
Attack MEDIUM
Luis Burbano, Diego Ortiz, Qi Sun +5 more
Embodied Artificial Intelligence (AI) promises to handle edge cases in robotic vehicle systems where data is scarce by using common-sense reasoning...
5 months ago cs.CR cs.AI cs.LG
PDF
Tool LOW
João Vitorino, Eva Maia, Isabel Praça +1 more
Due to the susceptibility of Artificial Intelligence (AI) to data perturbations and adversarial examples, it is crucial to perform a thorough...
5 months ago cs.LG cs.CR
PDF
Attack MEDIUM
Anshul Nasery, Edoardo Contente, Alkin Kaz +2 more
Model fingerprinting has emerged as a promising paradigm for claiming model ownership. However, robustness evaluations of these schemes have mostly...
5 months ago cs.CR cs.AI cs.LG
PDF
Benchmark MEDIUM
Matheus Vinicius da Silva de Oliveira, Jonathan de Andrade Silva, Awdren de Lima Fontao
Large Language Models (LLMs) are widely used across multiple domains but continue to raise concerns regarding security and fairness. Beyond known...
5 months ago cs.AI cs.IR cs.LG
PDF
Other LOW
Zheng Zhang, Ziwei Shan, Kaitao Song +2 more
Process Reward Models (PRMs) have emerged as a promising approach to enhance the reasoning capabilities of large language models (LLMs) by guiding...
Attack MEDIUM
Firas Ben Hmida, Abderrahmen Amich, Ata Kaboudi +1 more
Deep neural networks (DNNs) are increasingly being deployed in high-stakes applications, from self-driving cars to biometric authentication. However,...
5 months ago cs.CR cs.LG
PDF
Benchmark LOW
Seiji Maekawa, Jackson Hassell, Pouya Pezeshkpour +2 more
Existing benchmarks for tool-augmented language models (TaLMs) lack fine-grained control over task difficulty and remain vulnerable to data...
5 months ago cs.CL cs.PL
PDF
Benchmark LOW
Yao Tong, Haonan Wang, Siquan Li +2 more
Fingerprinting Large Language Models (LLMs) is essential for provenance verification and model attribution. Existing methods typically extract...
5 months ago cs.CR cs.AI cs.CL
PDF
Attack LOW
Shuai Shao, Qihan Ren, Chen Qian +8 more
Advances in Large Language Models (LLMs) have enabled a new class of self-evolving agents that autonomously improve through interaction with the...
5 months ago cs.AI cs.CL cs.LG
PDF
Attack HIGH
Qinjian Zhao, Jiaqi Wang, Zhiqiang Gao +3 more
Large Language Models (LLMs) have achieved impressive performance across diverse natural language processing tasks, but their growing power also...
Benchmark LOW
Joel Dyer, Daniel Jarne Ornia, Nicholas Bishop +2 more
Evaluating the safety of frontier AI systems is an increasingly important concern, helping to measure the capabilities of such models and identify...
5 months ago cs.LG cs.AI stat.ML
PDF
Benchmark LOW
Yixu Wang, Xin Wang, Yang Yao +4 more
The rapid integration of Large Language Models (LLMs) into high-stakes domains necessitates reliable safety and compliance evaluation. However,...
Attack HIGH
Xiaobao Wang, Ruoxiao Sun, Yujun Zhang +4 more
Graph Neural Networks (GNNs) have demonstrated strong performance across tasks such as node classification, link prediction, and graph...
5 months ago cs.LG cs.CR
PDF
Attack MEDIUM
Marco Zimmerli, Andreas Plesner, Till Aczel +1 more
Deep neural networks remain vulnerable to adversarial examples despite advances in architectures and training paradigms. We investigate how training...
5 months ago cs.CV cs.AI cs.CR
PDF
Attack MEDIUM
Dennis Jacob, Emad Alghamdi, Zhanhao Hu +2 more
Large language models (LLMs) have become increasingly popular due to their ability to interact with unstructured content. As such, LLMs are now a key...
5 months ago cs.CR cs.LG
PDF
Track AI security vulnerabilities in real time
Get breaking CVE alerts, compliance reports (ISO 42001, EU AI Act),
and CISO risk assessments for your AI/ML stack.
Start 14-Day Free Trial