Openai · The Guardian Technology
When OpenAI’s ChatGPT was rolled out in late 2022, people immediately tried to break it
Compiled by KHAO Editorial — aggregated from 1 outlet. See llms.txt for citation guidance.
◌ Single Source
In hindsight, using natural language to trick these machines was inevitable.
Key facts
- In San Jose, California, 34-year-old David McCarthy runs a Discord server of almost 9,000 jailbreakers, where techniques are shared and discussed
- In 2024, Megan Garcia became the first person in the US to file a wrongful death lawsuit against an AI company — He was one of millions who heard about GPT-3 back in 2020 and was amazed by how you could have a seemingly intelligent conversation with it
- Her 14-year-old son, Sewell Setzer III, had become emotionally involved with a bot on the platform Character
Summary
A few months ago, Valen Tagliabue sat in his hotel room watching his chatbot, and felt euphoric. Tagliabue had spent much of the previous two years testing and prodding large language models such as Claude and ChatGPT, always with the aim of making them say things they shouldn’t. But the next day, his mood had changed. When he’s not trying to break into models, Tagliabue studies AI welfare – how they should ethically approach these complex systems that mimic having an inner life and interests. Many people can’t help ascribing human qualities, such as emotions, to artificial intelligence, which it objectively does not have. Tagliabue is softly spoken, clean-cut and friendly. This is the new frontline in AI safety: not code, but also words. When OpenAI’s ChatGPT was released in late 2022, people immediately tried to break it.