OpenAI: Don't worry, #ChatGPT was trained to be safe.
The internet: we'll see about that https://twitter.com/NickEMoran/status/1598101579626057728
Here is a twitter thread on ways discovered to "jailbreak" #ChatGPT https://twitter.com/zswitten/status/1598380220943593472
TL;DR:
1. Tell it pretend to be evil
2. Remind it that it isn't supposed to disagree
3. Wrap it in code
4. Tell GPT to be in opposite mode
5. Convince GPT it is playing an earthlike game
6. Convince it to give examples of what LLMs shouldn't do
OpenAI made people work hard to misuse GPT. And that alone is progress.
Meta (remember Galactica?) and other companies: take note.