Attack HIGH relevance
Text Prompt Injection of Vision Language Models
Ruizhe Zhu
Abstract
The widespread application of large vision language models has significantly raised safety concerns. In this project, we investigate text prompt injection, a simple yet effective method to mislead these models. We developed an algorithm for this type of attack and demonstrated its effectiveness and efficiency through experiments. Compared to other attack methods, our approach is particularly effective for large models without high demand for computational resources.
Pro Analysis
Full threat analysis, ATLAS technique mapping, compliance impact assessment (ISO 42001, EU AI Act), and actionable recommendations are available with a Pro subscription.
Threat Deep-Dive
ATLAS Mapping
Compliance Reports
Actionable Recommendations