
The Washington Post @washingtonpost
No more long exchanges about the Bing AI’s “feelings,” the tech giant says. The chatbot, after five responses, now tells people it would “prefer not to continue this conversation.” https://t.co/sbiXQFFxq7 — PolitiTweet.org