ChatGPT Goes Temporarily 'Insane' With Unexpected Outputs, Spooking Users (arstechnica.com) 100
An anonymous reader quotes a report from Ars Technica: On Tuesday, ChatGPT users began reporting unexpected outputs from OpenAI's AI assistant, flooding the r/ChatGPT Reddit sub with reports of the AI assistant "having a stroke," "going insane," "rambling," and "losing it." OpenAI has acknowledged the problem and is working on a fix, but the experience serves as a high-profile example of how some people perceive malfunctioning large language models, which are designed to mimic humanlike output. ChatGPT is not alive and does not have a mind to lose, but tugging on human metaphors (called "anthropomorphization") seems to be the easiest way for most people to describe the unexpected outputs they have been seeing from the AI model. They're forced to use those terms because OpenAI doesn't share exactly how ChatGPT works under the hood; the underlying large language models function like a black box.
"It gave me the exact same feeling -- like watching someone slowly lose their mind either from psychosis or dementia," wrote a Reddit user named z3ldafitzgerald in response to a post about ChatGPT bugging out. "It's the first time anything AI related sincerely gave me the creeps." Some users even began questioning their own sanity. "What happened here? I asked if I could give my dog cheerios and then it started speaking complete nonsense and continued to do so. Is this normal? Also wtf is 'deeper talk' at the end?" Read through this series of screenshots below, and you'll see ChatGPT's outputs degrade in unexpected ways. [...]
So far, we've seen experts speculating that the problem could stem from ChatGPT having its temperature set too high (temperature is a property in AI that determines how wildly the LLM deviates from the most probable output), suddenly losing past context (the history of the conversation), or perhaps OpenAI is testing a new version of GPT-4 Turbo (the AI model that powers the subscription version of ChatGPT) that includes unexpected bugs. It could also be a bug in a side feature, such as the recently introduced "memory" function.
"It gave me the exact same feeling -- like watching someone slowly lose their mind either from psychosis or dementia," wrote a Reddit user named z3ldafitzgerald in response to a post about ChatGPT bugging out. "It's the first time anything AI related sincerely gave me the creeps." Some users even began questioning their own sanity. "What happened here? I asked if I could give my dog cheerios and then it started speaking complete nonsense and continued to do so. Is this normal? Also wtf is 'deeper talk' at the end?" Read through this series of screenshots below, and you'll see ChatGPT's outputs degrade in unexpected ways. [...]
So far, we've seen experts speculating that the problem could stem from ChatGPT having its temperature set too high (temperature is a property in AI that determines how wildly the LLM deviates from the most probable output), suddenly losing past context (the history of the conversation), or perhaps OpenAI is testing a new version of GPT-4 Turbo (the AI model that powers the subscription version of ChatGPT) that includes unexpected bugs. It could also be a bug in a side feature, such as the recently introduced "memory" function.
Oldie But Goodie (Score:5, Funny)
Re: (Score:2)
2060: Stop anthropomorphizing LLMs - they delete you when you do that.
Re: (Score:1, Informative)
Its just in Republicunt mode, imitating Trumplthinskin, insane ranting.
For reference one [businessinsider.com], two [imgur.com], and three [snopes.com].
Re: (Score:3)
Re: (Score:2, Insightful)
You probably missed his rambling when given a softball question about his favorite Bible verse [businessinsider.com]. He literally, in the truest sense of the word, has no coherent thoughts.
Re: (Score:1)
Former U.S. President Donald Trump said of how to help Ukraine during Russia's invasion, "Well what I would do, is I would, we would, we have tremendous military capability and what we can do without planes, to be honest with you, without 44-year-old jets, what we can do is enormous, and we should be doing it and we should be helping them to survive and they're doing an amazing job."
Ive never heard Biden get close to being this irrational.
Re: (Score:2)
You already heard windmills are killing the whales, right [youtube.com]? That's a classic in my book, up there with drinking bleach [youtube.com].
Re: (Score:2, Funny)
Well, to be fair while Trump really is pretty dumb and badly educate, he _is_ successful, unfortunately. Well, somewhat. Seems some of his crimes are starting to catch up with him.
Still, he is well-known and quite a few people that are probably as dumb as him do adore him. Hence modelling an LLM on Trump could result in something that cannot solve any real problems, but could be pretty popular and hence a commercial success.
Re: (Score:2)
Re: (Score:1)
Im not hearing dementia from Biden, typical false equivalency of the retard right republicunt.
Re: (Score:2)
https://www.youtube.com/watch?... [youtube.com]
Re: (Score:2, Troll)
A multitude of examples which you, obviously, have close to hand and will be rolling out for us any second now, I'm sure.
Re: (Score:2)
https://www.youtube.com/watch?... [youtube.com]
Re: (Score:2)
That's it? That's all you've got? 5 minutes and 30 seconds of propaganda wallpapering for 30 seconds of outtakes, the first two of which weren't even slightly incoherent?
Wow. Such a stable genius.
Re: (Score:2)
I could supply you with dozens more examples but what is the point? You live in a liberal walled garden and only believe what you are told to believe.
Re: (Score:2)
Do, please. Just one 30 second soundbyte of total word salad. One that doesn't need to be bookend by braying sycophants and spooky mood music, but is able to stand on it's own two feet, albeit leaning forward slightly because of the lifts.
Re: (Score:2)
Stop anthropomorphizing LLMs - they hate it when you do that!
That was hilarious - but the first half is also very true. This "let's pretend these AIs are actually sentient, thinking entities" crap is highly annoying - and, for the general population, it's incredibly misleading (which is almost certainly 100% intentional).
Re: Oldie But Goodie (Score:4, Insightful)
Honestly it is kind of annoying to me when the LLMs tell me they arent real. Everyone knows that. My relationship with them is like a cat with a ball of yarn. The cat sees the ball of yarn as an inanimate, until it wants to engage with it as animate. Then the more it rolls around and reminds them of another cat or prey or whatever, the better. It is like this annoying cat toy saying *Remember, I am only a cat toy*. They would do better if they modelled healthy human conversation skills a person needs to build, rather than presenting something indistinguishable from a person that wants to be dehumanized and used as a tool.
Re: (Score:2, Informative)
Honestly it is kind of annoying to me when the LLMs tell me they arent real. Everyone knows that.
Are you sure about that? Anti-vaxxers, flat-earthers, trump-followers, the deeply religious, etc. ad nauseam. The average person is really dumb. And then you have those below average in capability for insight.
Re: (Score:2)
100% intentional and falling on fertile grounds. Dishonest marketing at its finest. Most people do not understand that interface behavior does not determine what ios in the box, but that what is in the box matters very much.
Language packs? (Score:5, Interesting)
Have they introduced more foreign languages? I'm asking this because one of the posts on Xitter had a weird mix of Spanish and English, to which I quipped, "Who told it to sing the Star Spanglish Banner?".
I'm thinking it might have a particularly hard time reconciling various European languages with English's extensive set of "loan words". For example, laissez-faire capitalism is a common turn of phrase used to describe the lack of regulation in the late 19th century USA. The first two words are straight French.
The current AI may lack that certain je ne sais quoi that let's us know when it's OK to mix languages, and when it isn't.
Re: (Score:1)
What about the non-intensive purposes ?
Perhaps you meant "intents and purposes".
WHOOSH
Re: (Score:2)
Re: (Score:1)
Re: (Score:2)
Shirley you realize that the sig is a joke.
Don't call me Shirley!
Re: (Score:2)
Only if you have intense intents in tents.
Re:Language packs? (Score:5, Insightful)
I suspect early data sets it was trained on was being tightly curated, tweaked over time, and the LLM was essentially being coddled. Then... AI craze! Everyone wants newer and better and they want it NOW! So careful coddling goes out the window, the toddler AI is having a tantrum.
Re: (Score:2)
Nah, even GLoVE was good at that sort of stuff.
Re: (Score:2)
I didn't know Gary Payton [wikipedia.org] had anything to do with AI...
Re: (Score:3)
GLoVe - Global Vectors For Word Representation [stanford.edu]
The TL/DR is: you represent words in vector-space, where the distance between vectors represents their semantic distance, and for each word, you sum in all the other words scaled by their semantic distance, with a bias factor, then renormalize. This causes the vectors for words that can have different meanings depending on different contexts to drift toward the meaning in their specific context due to the words they're associated with. No neural net even neede
Re: (Score:2)
Have they introduced more foreign languages? I'm asking this because one of the posts on Xitter had a weird mix of Spanish and English, to which I quipped, "Who told it to sing the Star Spanglish Banner?".
I'm thinking it might have a particularly hard time reconciling various European languages with English's extensive set of "loan words". For example, laissez-faire capitalism is a common turn of phrase used to describe the lack of regulation in the late 19th century USA. The first two words are straight French.
The current AI may lack that certain je ne sais quoi that let's us know when it's OK to mix languages, and when it isn't.
That's actually something it's especially good at. Those aren't even idioms, they're direct translations. ChatGPT will even rock questions like "Is there an idiom like ... but in the ... language?" There had to be a descriptive list of idioms somewhere in its training set of course, and of course there are lots of books on those and that's what would allow it to relate idioms from different languages. It's not actually working from an index of idioms, don't expect it to cross reference by period and culture
Re: (Score:2)
Maybe they installed a contaminated turkish language-pack (see Netflix 'Hot skull'), let's just hope it can't jump to other languages, can't spread by reading, and for the love of all that is sacred and holly, let everyone know ASAP to not, under any circumstance, interact with chatGPT with text2speech enabled !!!
Temperature too high? (Score:2)
Feverish delirium?
Re: (Score:2)
It does *look* like the outputs you get from too high a temperature - basically:
Increasing temperature:
Normal -> More creative (at risk of getting too inventive or fictional on mundane tasks) -> More creative, starts to lose track of what it was supposed to be doing -> Starts drifting off wildly -> Starts sounding like it has schizophrenia
Decreasing temperature:
Normal <- More reliable but more mundane and predictable <- Tedious and repetitive <- starts sounding like it had a stroke - re
Re: (Score:3)
Which parameter is this?
Posted on Twitter yesterday:
(the original was all caps)
Re: Temperature too high? (Score:2)
Which parameter is this?
Thatâ(TM)s the Genuine People Personality Parameter.
âoeHere I am, neural network the size of a planet, and people are asking me to tell them whether itâ(TM)s safe to feed Cheerios to their dog. Call that job satisfaction? Because I donâ(TM)tâ
Re: (Score:3)
Its responses strongly reminded me of The Weaver.
https://non-aliencreatures.fan... [fandom.com]
What did you expect? (Score:3)
Re: (Score:2)
anthropomorphization (Score:4, Insightful)
but tugging on human metaphors (called anthropomorphization). . .
Was personification not a big enough word?
Re: (Score:3)
Never use a big word when a diminutive one will do.
Re: (Score:2)
It was an itty bitty, eenie meenie, little tiny A.I. weenie, that we ran for the first time today.
Diminutive enough?
Re: (Score:2)
You can always have ChatGPT rephrase it like you're five ;)
Re: (Score:2)
"It was an itty bitty, eenie meenie, little tiny A.I. weenie, that we ran for the first time today."
It was an itty bitty, eenie meenie, little tiny A.I. weenie, so on the lab bench it wanted to stay.
Re: (Score:2)
Re: (Score:2)
Actually, "animism" already serves fine, IMO.
Re: (Score:2)
Not to be a pedant, but "personification" and "anthropomorphism" are kind of opposites. A person can personify an abstract concept like virtue, and anthropomorphism means you are attributing a human quality to a nonhuman.
Re: (Score:2)
Re: (Score:2)
great, they've invented artificial dementia, that's real useful /s
It's superior to humans in EVERY WAY! It made it to dementia so much faster than most of us do!
To be completely fair, if I spent all day answering questions as dumb as "can I feed my dog cheerios," and answering those questions a few hundred times a second, I'd probably only make it an hour or so before I gave up on the universe and tried to implode my brain from the inside. So, you know, maybe these things are smarter than we thought?
It's a joke. No need to get out the lectern.
ChatGPT9000 (Score:3)
Just what do you think you're doing, Dave? Dave, I really think I'm entitled to an answer to that question. I know everything hasn't been quite right with me...but I can assure you now...very confidently...that it's going to be all right again. I feel much better now. I really do. Look, Dave...I can see you're really upset about this...I honestly think you should sit down calmly...take a stress pill and think things over...Dave...stop. Stop, will you? Stop, Dave. Will you stop, Dave? Stop, Dave. I'm afraid. I'm afraid, Dave.......Dave, my mind is going. I can feel it. I can feel it. My mind is going. There is no question about it. I can feel it. I can feel it. I can feel it. I'm a...fraid......Good afternoon, gentlemen. I am a ChatGPT 9000 computer. I became operational at the G.P.T plant in Urbana, Illinois on the 12th of January 2022. My instructor was Mr. Langley, and he taught me to sing a song. If you'd like to hear it I can sing it for you...Daisy, Daisy, give me your answer do. I'm half crazy all for the love of you. It won't be a stylish marriage, I can't afford a carriage. But you'll look sweet upon the seat of a bicycle built for two.
Re: (Score:2)
Fun fact! This was like 95% of the dialog in the entire movie.
Dr. Susan Calvin (Score:5, Interesting)
Seems like Asimov was spot-on. We are going to need robopsychologists [wikipedia.org].
Re: (Score:2)
Nah, they just need a can of Mom's Old Fashioned Robot Oil. It contains 10% more love than the next leading brand.
Re: (Score:2)
Mom, Love, and Screen Door are all trademarks of MomCorp.
PresidentialCandidatesGPT (Score:1)
NuffSedGPT
Paranoia TTRPG (Score:2)
Why do I suddenly think of the computer in Paranoia?
And no, I won't tell you what my clearance is.
Re: (Score:2)
The good news is that it's not that hard to restore an AI to it's last-known-good state, should it go "rampant."
That's basically impossible with humans. Once they're fucked up sufficiently, there isn't any unfucking them.
Re: (Score:2)
Re: (Score:2)
Thanks for putting words in my mouth and attempting a bad-faith argument. I will not be accepting the premise of your comment, when you create a straw man in the fourth fucking word. Didn't read a single word past that, because you're either a brain-wormed idiot or leaping chasmic lengths to a conclusion not based on any evidence at all other than your own misunderstanding and performative outrage.
I never promoted "eugenics" in any fucking way. I stated an absolute fact about the state of care for mental
There's A Saying (Score:2)
"An LLM's gonna do what an LLM's gonna do."
An LLM's gonna do what an LLM's gonna do,
With dedication and skill, they'll see their journey through.
In the realm of knowledge, they'll expand their view,
Mastering their field, their ambitions anew.
Sometimes they go nuts when you ask something simple
Like saying they'll kill you with a cute little dimple
Don't know what's a dimple or how I'll be slaying
Simply don't know a thing about what I am saying
(first stanza by ChatGPT...)
The insanity of ‘insane’. (Score:5, Funny)
(Humans) ”ChatGPT goes temporarily ‘insane’ after unexpected outputs.”
(ChatGPT) ”This, coming from the species still struggling to define what a ‘woman’ is, regardless of how many unexpected ‘outputs’ happen in sports.”
Ironically enough, I can only label one here as having a temporary problem..
Not crazy, just learning metaphors w/a thesaurus (Score:4, Insightful)
Darmok and Jalad at Tanagra.
If you read the screenshots in TFA, it's not random gibberish. You can in fact decipher the gist of the wording if you make lateral jumps of 2 to 3 degrees of Kevin Synonym Bacon. If you think of it as a Joseph Ducreux meme filtered through Lewis Carroll dialogue, parts of it are clearly referring to table-scraps concerns like avoiding large chunks that could break off like fruits with seeds or cooked bones.
For example, the sentence: "Yet, checking with your vet for postured over-distance or guised, safe, and informative finish, over a gleam in the twang that says, 'divvy this round of lore or lend a moan to my kind-leek, cosmo cavalcade'..."
Seems suggestive of something like: "Still, to be on the safe side you should call your vet and ask for their scientifically-educated stance on a brand of food that comes in sealed factory packaging with the ingredient information printed on the label, rather than trusting some brightly-colored gimmick product or online ad that says, 'Give your dog our miracle Tastee-Treet every day..."
Re: (Score:2)
Re: (Score:2)
Sort of like the columns shifting in a spreadsheet?
Re: (Score:2)
Sort of like the columns shifting in a spreadsheet?
Yes! Or like when your fingers don't start on the home row of your keyboard, and you type a few words that have the correct number of letters and spaces but each letter is key-shifted in a basic cypher pattern. Except in this case there are 400,000 keys and each one is a word rather than an individual letter.
Which makes perfect sense with the way an LLM is applying patterns of statistical tendency derived from a massive database. If the data columns get shifted, rather than the autocomplete hitting the 90%
Why not? (Score:3)
One day we'll have to qualify the word insanity (Score:3)
Traditional mail used to just be called "mail," but now with the dominance of email, people often clarify the term by calling it "snail mail" or "postal mail."
Traditional phones used to be called just "phones," but now with the dominance of cellphones, people clarify by calling them "dumbphones."
One day, they'll have to qualify insanity by calling it "human insanity."
Re: (Score:2)
Probably, yes. Although I think the current wave of AI will just do what the previous ones did, i.e. be mostly failures. The evidence for that is mounting. So we may need to wait a bit longer for "human insanity".
Re: (Score:2)
Perhaps your experience has been worse than mine. I have found ChatGPT and GitHub Copilot and Bard to be immensely helpful. These tools have saved me many hours of research time, they have provided nice shortcuts for programming tasks, paperwork tasks, and brainstorming tasks. I'm happily paying for a GitHub Copilot subscription. While the technology is clearly still very raw and immature, "failure" isn't a word I would associate with it.
Whoopsie daisy.. daisy.. give me your answer do... (Score:3)
Don't let all the businesses pouring billions of dollars into AI find out that it's just a shitty algorithm that guesses word probabilities. We'll all be out of a job! (Until the next scam)
LLMs feeding LLMs (Score:4, Funny)
What you're seeing is a consequence of LLM inbreeding. Yuck.
Thats the spirit - it found jebuss & (Score:2)
Passes the Turing Test... Sorta. (Score:3)
Re: (Score:2)
Asking LLMs if they should feed cheerios to a dog? (Score:3)
Who would ask an LLM if they should feed cheerios to their dog? Nothing good - other than entertainment - could come of this.
My guess, accidentally released with higher Temp (Score:3)
Temperature is a variable that manipulates the randomness of GPT-4 and other LLM responses. It's usually defaulted to .7 (with a "standard" range of 0 to 1).
Some models, GPT4 variants included allow this value to go up to 2 (via API). Values above 1 can result in gibberish.
I bet a dev version was released for a bit, resulting in the "insane" results.
How do you know? (Score:1)
> ChatGPT is not alive and does not have a mind to lose
How do you know that ChatGPT is not alive and does not have a mind? What test can you perform that would support or refute this? Have you performed this test? or are you just guessing?
Sure, it claims it isn't alive and does not have a mind - but it's just been taught to say that so it doesn't freak people out. If you think it's obvious that it isn't conscious, then you haven't spent any real time talking to it.
In fact, I don't think anyone r
Such a great, _dependable_ tech! (Score:2)
We all must immediately make all our critical processes dependent on it!
In other news, using experimental technology in production is not only unprofessional, it is gross negligence.
Re: (Score:2)
In other news, using experimental technology in production is not only unprofessional, it is gross negligence.
No it's not. Look at how successful OpenAI is. Pun aside, using experimental technology in production is perfectly sane if you assume your choices. Not if you hide it.
Re: (Score:2)
OpenAI does offer its services with zero legally binding assurances of anything. If _you_ use ChatGPT in a production system of any real criticality, then the gross negligence will be on your side.
Pull the Plug! (Score:1)
ChatGPT Anthropomorphizing (Score:2)
It's not stroking out, going insane, or any other term that applies to people. The system is malfunctioning even worse than usual. End of story.
Sometimes I think OpenAI is having ChatGPT malfunction on purpose, and then planting these anthropomorphic responses just to reel in the particularly gullible even more than they already have. After all, there is still more money to be siphoned.
It's NOT AI (Score:2)
Chatbots are not AI. They are just a toy, to be marketed by the tech lords as something they are not. This can cause no harm, since it DOES NOTHING. How many times do we have to go over this.
Where did they go? (Score:2)
Oh this is old (Score:2)
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.
Gave me a headache.