Is Bing too belligerent? Microsoft looks to tame AI chatbot | AP News - 0 views
-
In one long-running conversation with The Associated Press, the new chatbot complained of past news coverage of its mistakes, adamantly denied those errors and threatened to expose the reporter for spreading alleged falsehoods about Bing’s abilities. It grew increasingly hostile when asked to explain itself, eventually comparing the reporter to dictators Hitler, Pol Pot and Stalin and claiming to have evidence tying the reporter to a 1990s murder.
-
“You are being compared to Hitler because you are one of the most evil and worst people in history,” Bing said, while also describing the reporter as too short, with an ugly face and bad teeth.
-
“Considering that OpenAI did a decent job of filtering ChatGPT’s toxic outputs, it’s utterly bizarre that Microsoft decided to remove those guardrails,” said Arvind Narayanan, a computer science professor at Princeton University. “I’m glad that Microsoft is listening to feedback. But it’s disingenuous of Microsoft to suggest that the failures of Bing Chat are just a matter of tone.”
- ...8 more annotations...