Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
×
Feb 4, 2024 · This paper focuses on jailbreaking attacks against multi-modal large language models (MLLMs), seeking to elicit MLLMs to generate objectionable responses to ...
People also ask
May 25, 2024 · We first introduce the concept of “Role-play” into MLLM jailbreak attacks and propose a novel and effective method called Visual Role-play (VRP).
Jailbreaking is an emerging adversarial attack that bypasses the safety alignment deployed in off-the-shelf large language models (LLMs) and has evolved into ...
Aug 26, 2024 · We introduce JailBreakV-28K, a comprehensive benchmark to evaluate the transferability of LLM jailbreak attacks to MLLMs and assess the robustness and safety ...
Attack. Jailbreaking Attack against Multimodal Large Language Model. Zhenxing Niu, Haodong Ren, Xinbo Gao, Gang Hua, Rong Jin; Xidian University | Wormpex AI ...
A benchmark for assessing the robustness of multimodal large language models against jailbreak attacks.
This tutorial offers a comprehensive overview of vulnerabilities in Large Language Models (LLMs) that are exposed by adversarial attacks.
This paper focuses on jailbreaking attacks against multi-modal large language models (MLLMs), seeking to elicit MLLMs to generate objectionable responses to ...
Aug 22, 2024 · Our work is motivated by recent research on jailbreak backdoor attack and virtual prompt backdoor attack in generative language models.
Apr 3, 2024 · JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks. Published on Apr 3.