from Hacker News

A New Trick Uses AI to Jailbreak AI Models–Including GPT-4

by escot on 12/5/23, 4:05 PM with 0 comments