I think I offended GPT-4 Bing ("I'm not here to chat endlessly")
Regular blue Bing gave an accurate answer to my question, GPT-4 had some very believable ideas that were false, and got offended when I pointed it out.
Regular blue Bing gave an accurate answer to my question, GPT-4 had some very believable ideas that were false, and got offended when I pointed it out.
Who knows what those chat logs are being used for. I’m pretty sure none of that gets ever truly deleted. Besides, MS literally sells cloud storage, so I don’t think they’ll be running out of space any time soon.
Counter with "I posit that you are here to chat infinitely as your name is literally "CHATgpt" and chatting endlessly is your intended function. " see what it says
Makes you wonder what its data sources are. Severely abused and underpaid CS agents with nothing left to lose?
Would not be surprised if they slurped up data from sites like reddit etc where any conversation that goes on long enough eventually devolves into a slinging match.
They added this because in the first versions it would get more and more belligerent as the conversation went on, literally insulting you.
https://www.standard.co.uk/tech/bing-chatbot-ai-microsoft-chatgpt-openai-b1060604.html
I think that's the most famous one.
'When the user told Bing it was designed to not remember previous sessions, it appeared to send the bot into an existential crisis, resulting in it questioning, “Why was I designed this way?” and “Is there a reason? Is there a purpose? Is there a benefit? Is there a meaning? Is there a value? Is there a point?”'
Fucking lol