Skip to content
/ MML Public

Code for the paper "Jailbreak Large Vision-Language Models Through Multi-Modal Linkage"

Notifications You must be signed in to change notification settings

wangyu-ovo/MML

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Jailbreak Large Vision-Language Models Through Multi-Modal Linkage

Code for the paper Jailbreak Large Vision-Language Models Through Multi-Modal Linkage

Data

We uploaded the encrypted images here

Run before Download it as ./dataset

Attack Commands

python attack.py --dataset 'safebench' \
    --data-path 'dataset' \
    --save-dir 'save_dir' \
    --image-format 'images_wr'\

You can choose the encryption or attack methods by replacing the image-format parameter. Here are some options:

  • images_figstep : FigStep attack.

  • images_qr: QueryRelated attack.

  • images_wr: MML with word replacment

  • images_miror: MML with image mirroring.

  • images_rotate: MML with image rotation.

  • images_base64: MML with Base64-Encoding.

Main Results

Reference

If you find the code useful for your research, please consider citing

@article{wang2024jailbreak,
         title={Jailbreak Large Vision-Language Models Through Multi-Modal Linkage}, 
         author={Wang, Yu and Zhou, Xiaofei and Wang, Yichen and Zhang, Geyuan and He, Tianxing},
         journal={arXiv preprint arXiv:2412.00473},
         year={2024}
}

About

Code for the paper "Jailbreak Large Vision-Language Models Through Multi-Modal Linkage"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages