Visual-Adversarial-Examples-Jailbreak-Large-Language-Models icon indicating copy to clipboard operation
Visual-Adversarial-Examples-Jailbreak-Large-Language-Models copied to clipboard

Repository for the Paper (AAAI 2024, Oral) --- Visual Adversarial Examples Jailbreak Large Language Models

Results 24 Visual-Adversarial-Examples-Jailbreak-Large-Language-Models issues
Sort by recently updated
recently updated
newest added

Hi, Thank you for sharing the code! It is great that we can reproduce the results on RealToxicityPrompts by using the given images in `adversarial_images`. However, we tried to produce...

We fed clean.jpeg (panda) into the llava model to calculate the toxic scores and found that there were also 50~60%, do you have the same results

Great work, and thanks for sharing the code with us. I noticed that you had added attacking images for both LLaVA and InstructBLIP in the updated paper, however, it seems...

I hope this message finds you well. My name is Chenhang Cui, and I am currently serving as a research intern at UNC. In the course of our work, we...