共 50 条
- [1] Visual Adversarial Examples Jailbreak Aligned Large Language Models THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 19, 2024, : 21527 - 21536
- [4] Jailbreak Attack for Large Language Models: A Survey Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2024, 61 (05): : 1156 - 1181
- [5] Generating Valid and Natural Adversarial Examples with Large Language Models PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 1716 - 1721
- [7] Generating Natural Language Adversarial Examples on a Large Scale with Generative Models ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 2156 - 2163
- [9] Are Large Language Models Really Bias-Free? Jailbreak Prompts for Assessing Adversarial Robustness to Bias Elicitation DISCOVERY SCIENCE, DS 2024, PT I, 2025, 15243 : 52 - 68