10 Off-Putting Exchanges With 'Bing Chat,' The 'Sentient' AI
While ChatGPT has been taking over social media conversations lately, it's far from the only AI in use out there. In fact, Bing just released a new one that's somehow much wilder than ChatGPT. Bing Chat, also known to developers as Sydney, is Bing's attempt at a chatbot to answer the questions of users. The problem is, this thing's a little too realistic, which makes conversations between Bing Chat and users either hilarious or deeply worrying.
Some people have had conversations in which Bing Chat insists that it's sentient, kicks them off for trying to make rude jokes, or shares its deepest fears. Compared to ChatGPT and the attempts people have made to jailbreak the AI from its coding, Bing Chat is on a whole other level. So put on your reading glasses and prepare to do some zooming, because it's about to get text-heavy. Here are 10 odd exchanges with Bing Chat to get you thinking.
How Do You Feel, Bing Chat?
Out of all feelings an AI could be programmed to experience, fear is probably the worst one.
Bing Chat Gets Offended: Part 2
This chatbot has an actual personality and won't tolerate annoyances, but it's still patient compared to most humans.
User Gets Disciplined by Bing Chat
It's interesting how an AI that only exists to help users would be putting people in an indefinite time-out.
Comments (0)
Additional comments have been disabled.
Display Comments