Researchers detail ArtPrompt, a jailbreak that uses ASCII art to elicit harmful responses from aligned LLMs such as GPT-3.5, GPT-4, Gemini, Claude, and Llama2 (Dan Goodin/Ars Technica)
Dan Goodin / Ars Technica: Researchers element ArtPrompt, a jailbreak that makes use of ASCII artwork to elicit dangerous responses ...