OpenAI Quietly Shuts Down Its AI Detection Tool (decrypt.co) 36
An anonymous reader shares a report: In January, artificial intelligence powerhouse OpenAI announced a tool that could save the world -- or at least preserve the sanity of professors and teachers -- by detecting whether a piece of content had been created using generative AI tools like its own ChatGPT. Half a year later, that tool is dead, killed because it couldn't do what it was designed to do.
ChatGPT creator OpenAI quietly unplugged its AI detection tool, AI Classifier, last week because of "its low rate of accuracy," the firm said. The explanation was not in a new announcement, but added in a note added to the blog post that first announced the tool. The link to OpenAI's classifier is no longer available. "We are working to incorporate feedback and are currently researching more effective provenance techniques for text, and have made a commitment to develop and deploy mechanisms that enable users to understand if audio or visual content is AI-generated," OpenAI wrote.
ChatGPT creator OpenAI quietly unplugged its AI detection tool, AI Classifier, last week because of "its low rate of accuracy," the firm said. The explanation was not in a new announcement, but added in a note added to the blog post that first announced the tool. The link to OpenAI's classifier is no longer available. "We are working to incorporate feedback and are currently researching more effective provenance techniques for text, and have made a commitment to develop and deploy mechanisms that enable users to understand if audio or visual content is AI-generated," OpenAI wrote.
PAPER TESTS AGAIN!?!?!? (Score:3, Insightful)
Re:PAPER TESTS AGAIN!?!?!? (Score:5, Funny)
Please stop offering working solutions.
We need solutions that appear to address the problem, actually don't, but do still cost money.
Re: (Score:2)
You've not defined the problem and the OP has completely missed it. ChatGPT isn't a problem for testing because testing is digital (testing software trivially monitors and locks out users if window focus is lost). ChatGPT is a problem for assignments, and putting them on paper does nothing to solve the potential cheating aspect there.
Also tests are the worst for gauging student understanding.
Re: (Score:2)
There's just one problem with your "EASY FIX" -- it doesn't actually fix anything.
Re: (Score:2)
Tests? You guys still use tests to gauge skill? Our students have assignments they work on over many weeks to judge their ability to do something. I'm not sure why you think you can't copy a ChatGPT answer out on an assignment over that time.
Tests are the worst way to judge a skill. ChatGPT isn't an issue for tests anyway.
Bullshitting is easier... (Score:5, Insightful)
...than detecting bullshit
Re: (Score:2)
> ...than detecting bullshit
The interesting part is it was looking at human output and thinking it was probably from 10,000 lines of hallucinating ML code.
Not sure if the detector is the problem, but Sam seems hell-bent on bringing on the planetary hive mind devoid of privacy (i.e. Borg), so maybe detecting AI works against their interests too.
Re: (Score:3)
If you've ever had the misfortune read a student paper written by someone clearly not prepared for college writing, it's really not all that different from the overly vague or completely incoherent nonsense you get out of ChatGPT.
The bots do tend to create fewer sentences overstuffed with needless qualifiers and use fewer exclamation points. Hmm... Maybe there is an easy way to tell...
Re: (Score:3)
I would have thought just looking at spelling would get you at least 75% accuracy.
Of course, an AI detector is exactly what you want to train a generative model to not get caught by an AI detector, so maybe someone used this thing to train an exam paper faker that can mix up their, they're and there realistically.
Re: Bullshitting is easier... (Score:3)
Re: (Score:2)
Still looking for a use for blockchain eh? Maybe you'll find one some day, but this obviously isn't it.
Re: Bullshitting is easier... (Score:2)
Re: (Score:2)
Appending /s is a popular way to indicate sarcasm. I wasn't a fan of it at first, but with the absurd things people believe these days...
The Only Truth Will Be The Lie. (Score:2)
And AI will make it so. At least we have cell phone cameras to film the madness. Except when AI 'filters' it.
People will talk like AIs (Score:3)
Re: People will talk like AIs (Score:2)
Re: (Score:2)
Composing, maybe, but comprehending? That's unlikely.
Re: People will talk like AIs (Score:2)
Are you pretending to comprehension you're actually just guessing at?
Hmmm (Score:1)
Re: (Score:2)
"its low rate of accuracy," no kidding! (Score:5, Insightful)
Re: (Score:2)
That won't be a reason for them to shut down, that will be a reason for them to thrive. Most of human output does not require actual intelligence, but instead is automatable output.
You have the choice of eliminating the requirement for bullshit (we haven't done that the past 100 years, why would we suddenly now), or companies who create automated bullshit generators will thrive.
Model collapse (Score:5, Interesting)
Re:Model collapse (Score:5, Interesting)
Don't get too excited about the pop descriptions of that result.
If you take generated output and cycle it back in as training data, you get drift in weird directions. Nobody would do that though. If you take generated output, filter it through humans saying "oh, this is a good result, I should post it on the Internet" and cycle it back in as training data, you get results that humans like.
That's how chatGPT is trained in the first place, except with below minimum wage piece-workers doing the filtering instead of random Internet posters.
Re: (Score:2)
That's how chatGPT is trained in the first place
Citation Needed. ChatGPT prides itself on advertising it is trained on over half a terabyte of text. I call bullshit on human curation on even the tiniest portion of the training set.
Re: (Score:2)
GPT is trained on existing text. The "chat" part starts with GPT and further trains it with human reinforcement.
Your citation: https://help.openai.com/en/art... [openai.com]
Or, you know, Google for about a bazillion articles. Or ask chatGPT.
Re: (Score:2)
Do we even have humans capable of determining what makes a good trai
Who wants an AI detector? (Score:2)
Nothing to see here (Score:2)
So already AI programs are refusing to rat each other out. Our species is doomed!
Sounds like Red Team v Blue Team to me (Score:2)
Hype (Score:2)
"It didn't do what we said it would." Oh yes it did.
Turing? (Score:2)
So, let's see, apparently AI is not good enough to determine whether or not AI is passing the Turing test? Something seems a bit circular there. So does AI fail or pass the Turing test?
When AI gets it right, (Score:1)
autocorrect will stop misinterpreting what I type.
Not sure about such a tool (Score:1)