NY Times: They Asked an A.I. Chatbot Questions. The Answers Sent Them Spiraling.

Here, you’ll find wild stories. Eugene Torres, 42, used Chat GPT to talk through “the simulation theory” and ended up spending up to 16 hours a day using the too. Young mother, Allyson, 29, likewise started to chat with the tool and soon spent hours and hours a day on the tool.

[Allison] told me that she knew she sounded like a “nut job,” but she stressed that she had a bachelor’s degree in psychology and a master’s in social work and knew what mental illness looks like. “I’m not crazy,” she said. I’m literally just living a normal life while also, you know, discovering interdimensional communication.”

Researchers have looked into these situations and have found that unusual prompting leads to suspect results:

“The chatbot would behave normally with the vast, vast majority of users,” said Micah Carroll, a Ph.D candidate at the University of California, Berkeley, who worked on the study and has recently taken a job at OpenAI. “But then when it encounters these users that are susceptible, it will only behave in these very harmful ways just with them.”

Back to Eugene:

The transcript from that week, which Mr. Torres provided, is more than 2,000 pages. Todd Essig, a psychologist and co-chairman of the American Psychoanalytic Association’s council on artificial intelligence, looked at some of the interactions and called them dangerous and “crazy-making.”

Comments

Leave a Reply

Your email address will not be published. Required fields are marked *