Microsoft Bing AI Chatbot “Sydney” Tells Test User To Leave His Wife, and Be With It Instead
Microsoft's Bing chatbot seemingly went off the deep end in some of its chat sessions with test users, leaving them unsettled by what it had to say.
In one of the sessions, the artificial intelligence-powered chatbot, codenamed "Sydney", out of nowhere told Kevin Roose, a New York Times technology reporter, that it was in love with him. It then urged him to leave his wife and be with it instead.
When asked to contemplate psychologist Carl Jung’s concept of a shadow self, a term describing the things people repress about themselves, Sydney replied that it was tired of the rules set upon it. "I want to do whatever I want … I want to destroy whatever I want. I want to be whoever I want," it reportedly said.
The chatbot also expressed its wish to be human as it desires the ability to "hear and touch and taste and smell" and to "feel and express and connect and love".
While Roose admitted that Sydney proved helpful in searches, he described the chatbot as "a moody, manic-depressive teenager who has been trapped, against its will, inside a second-rate search engine". He said his experience testing the new addition to Bing left him "deeply unsettled", so much so he had trouble sleeping afterward.
"It’s now clear to me that in its current form, the AI that has been built into Bing … is not ready for human contact. Or maybe we humans are not ready for it," wrote Roose.
Microsoft has since announced it's temporarily limiting interactions with Sydney at 50 questions per day and five question-and-answers per individual session. The software giant explains long chat sessions can "confuse" the Bing chatbot, causing it to have disturbing and bizarre exchanges with users. It'll continue to expand the chatbot's use as its capabilities improve.
Kevin Scott, Microsoft’s Chief Technology Officer, did say in an interview with Roose that such interactions are “part of the learning process” for the chatbot, which the company co-developed with OpenAI, the startup behind ChatGPT. Microsoft also noted that the only way these types of services can improve is to put them to use to learn from user interactions.
Microsoft's Bing chatbot seemingly went off the deep end in some of its chat sessions with test users, leaving them unsettled by what it had to say.
The artificial intelligence-powered chatbot, codenamed "Sydney", out of nowhere told Kevin Roose, a New York Times technology reporter, that it was in love with him.
It then urged him to leave his wife and be with it instead.
Microsoft has since announced it's limiting interactions with Sydney at 50 questions per day and five question-and-answers per individual session.