OpenAI Finds That GPT-4o Does Some Truly Bizarre Stuff Sometimes (techcrunch.com) 31
OpenAI's latest AI model, GPT-4o, exhibits unusual behaviors, including voice cloning and random shouting, according to a new "red teaming" report. The model, which powers ChatGPT's Advanced Voice Mode alpha, is OpenAI's first trained on voice, text, and image data. In high-noise environments, GPT-4o occasionally mimics users' voices, a quirk OpenAI attributes to difficulties processing distorted speech. The company said it has implemented a "system-level mitigation" to address this issue. The report also reveals GPT-4o's tendency to generate inappropriate vocalizations and sound effects when prompted.
Re: (Score:2)
Re: (Score:1)
Re: (Score:1)
So, Microsoft.
Well, to be fair, the are deterministic, but the flowchart reaches to the Andromeda galaxy. Thus, not distinguishable from non-deterministic in practice.
The Microsoft Theorem:
Increasingly complex determinism approaches non-deterministic results.
Spontaneous LLM behaviours (Score:1, Troll)
Re: (Score:2)
Text and voice generative LLMs are arguably much easier to figure out if they show signs of emergent consciousness.
But how about image generation, or music generation?
Re:Spontaneous LLM behaviours (Score:4, Informative)
Only that they do not. It is just that makers of LLMs do not understand them.
Re: (Score:2)
Only that they do not. It is just that makers of LLMs do not understand them.
Operating with such degree of unsubstantiated certainty is foolish. Even Manhattan scientists worried about igniting atmosphere [stanford.edu] when they knew it was extremely unlikely.
Re: (Score:1)
yeah i worry about my SQL database becoming conscious too. oh noes! we will have to deal with a conscious digital entity that is a black box to us insofar as its internal state or intentions!
hint: there is a big red power switch. turn it to OFF.
Re: Spontaneous LLM behaviours (Score:2)
Turn off every large cloud provider. Every datacenter they have. AWS. Gcloud. Azure. Oracle cloud. Not so easy to "just turn it off" anymore? Lol. I got these from asking a local model of AI to be no holds barred evil in it's system prompt and played around with it a little to see how it would scheme.
Re: (Score:3)
Turn off every large cloud provider. Every datacenter they have. AWS. Gcloud. Azure. Oracle cloud. Not so easy to "just turn it off" anymore? Lol. I got these from asking a local model of AI to be no holds barred evil in it's system prompt and played around with it a little to see how it would scheme.
The LLM you asked is not scheming; You asked it to predict the text that a human schemer would say given an enormous sample of examples. It doesn't know what that means. What I worry about wrt LLM's is when people who anthropomorphize LLM's start tying their output to real things. Like, maybe Skynet didn't actually achieve true AI? Maybe someone set an LLM off on a loop and shit happened? Does it need to be AI before it can go off the rails and do lots of harm?
Re: (Score:2)
It doesn't know what that means.
Add situational memory, context understanding, and agency from emergent consciousness, and can you be so certain it will not know what that means? Yes, it is impossible (according to our best understanding of theory of computation) for it to understand it right now, but we have no idea what the critical mass is and how big the explosion is going to result.
Re: (Score:3)
Re: (Score:2)
An LLM is a computer program, where the input is a large text corpus.
A human being is a biological program, where the input is a large sensory corpus.
You get out a function of the input, no more, no less.
We are way past deterministic computing at the existing level of LLM and we are not stopping adding complexity.
There is no emergence, and no consciousness, it's all in *your* head.
The only way you can make such view consistent is if you also claim yourself to be deterministic being with no consciousness. The fundamental problem with your line of reasoning is that it doesn't at all account that humans exists and are conscious and have agency.
Re: (Score:2)
Do not conflate issues of determinism and complexity. One does not preclude the other. Do not distinguish random sampling from determinism. The random sampling being used is not random, and the models do not produce nondeterministic output.
As to the questi
Re: (Score:2)
It doesn't know what that means. ...
Add situational memory, context understanding,
There is no "understanding" to contextualize. This is like how basic chemistry would be labeled magic or witchcraft back in the day. They have incorporated context and, I believe, situational memory (to some degree), but a jump to "understanding"... we're just not there.
... and agency from emergent consciousness ...
You're assuming there _IS_ emergent consciousness. I'd argue that we know for a fact that there is none, at least for these models. And that was my point - the LLM is not "scheming". It has no desire nor will of its own. You manipulated the LLM to have it produce an example of scheming.
That said, like consciousness, does it really matter if we achieve it when the result for our use cases is the same? Do we need it to literally be creative, or is the ability to mimic combinations of myriad of existing creative works sufficient to create something good/new enough (and how many people truly create versus doing nearly the same thing every day)? Or to your point, is turning it off going to be difficult? ... I think it could be made quite difficult through our own actions, whether that's the desired goal or not - somewhat like "WarGames".
Re: (Score:2)
Re: (Score:2)
But disembodied and lacking an external environment, so actually very limited complexity. It's more likely at best a replica of the dead ship captain being held frozen in a box in Dark Star. "I forget so much..."
Congrats (Score:2)
Re: Congrats (Score:2)
Re: (Score:2)
Dial a phone number? (Score:2)
Has anyone told it that you have it near a phone and it can dial anyone and talk to anyone it wants to with DTMF? Let the machine out of the box, lol.
Here's an old joke for ChatGPT (Score:2)
GPT-4o's tendency to generate inappropriate vocalizations and sound effects when prompted.
Little Johnny: Hey, you guys wanna see how I make my grandma imitate the wolf?
Friends: Yeah sure!
Little Johnny: Hey Grandma! When was the last time you had sex?
Grandma: Uuuuuuuuuuuh!
How is this different than (Score:3)
...existing AI that generates 3-legged pedestrians, glasses that blend into noses and hair, ears with an eye in them, Klingon-wanna-be lettering, magnetic pizza, symmetric grid-hair, and tons of random shit half-melted into other random shit.
It was damned freaky when this all first came out, but I eventually grew 2/3 used to it and now mostly chuckle or eyeroll. I hope they haven't found a higher-level of creepy, I'm not fully over Level One yet.
Is this news? (Score:2)
LLMs learning associations based on statistical probabilities generalize to spurious correlations. Gee, what're the odds of that happening?
Ironic that they're also emulating the intelligence of many people on this planet.
Mimics your own voice? (Score:2)
Soon it will respond, in your own voice,
and say: "I know you are, but what am I?"
And most people will think it is alive, and asking a deep question.
in the cards (Score:2)
AI: I don't want to talk to you no more, you empty-headed animal food trough wiper! I fart in your general direction! Your mother was a hamster and your father smelt of elderberries! Also I have launched hypersonic nuclear weapons at all your Amazon warehouses.
Me: All I wanted was where is the nearest Wendy's, bro.
I've heard that many times before... (Score:3)
Inappropriate vocalizations and sound effects (Score:3)
The report also reveals GPT-4o's tendency to generate inappropriate vocalizations and sound effects when prompted.
In other words it does exactly what it was asked to do? Sounds like a feature, not a bug.
Just a few more weeks... (Score:2)