[논문리뷰] Jailbreak Attacks and Defenses against Multimodal Generative Models: A Survey
Liu, X., Cui, X., Li, P., Li, Z., Huang, H., Xia, S., Zhang, M., Zou, Y., & He, R. (2024). Jailbreak Attacks and Defenses against Multimodal Generative Models: A Survey. https://arxiv.org/abs/2411.09259 Jailbreak Attacks and Defenses against Multimodal Generative Models: A SurveyThe rapid evolution of multimodal foundation models has led to significant advancements in cross-modal understanding a..
2024. 12. 31.
[논문리뷰] BLUE SUFFIX: REINFORCED BLUE TEAMING FOR VISION-LANGUAGE MODELS AGAINST JAILBREAK ATTACKS
Zhao, Y., Zheng, X., Luo, L., Li, Y., Ma, X., & Jiang, Y. (2024). BlueSuffix: Reinforced Blue Teaming for Vision-Language Models Against Jailbreak Attacks. ArXiv, abs/2410.20971.https://arxiv.org/abs/2410.20971 BlueSuffix: Reinforced Blue Teaming for Vision-Language Models Against Jailbreak AttacksDespite their superb multimodal capabilities, Vision-Language Models (VLMs) have been shown to be v..
2024. 12. 30.
[논문리뷰] Visual Adversarial Examples Jailbreak Aligned Large Language Models
Qi, X., Huang, K., Panda, A., Henderson, P., Wang, M., & Mittal, P. (2024). Visual Adversarial Examples Jailbreak Aligned Large Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 38(19), 21527-21536. https://doi.org/10.1609/aaai.v38i19.30150 https://arxiv.org/abs/2306.13213 Visual Adversarial Examples Jailbreak Aligned Large Language ModelsRecently, there has been a ..
2024. 12. 26.