Follow Slashdot blog updates by subscribing to our blog RSS feed

 



Forgot your password?
typodupeerror
×
AI

OpenAI Threatens To Ban Users Who Probe Its 'Strawberry' AI Models (wired.com) 50

OpenAI truly does not want you to know what its latest AI model is "thinking." From a report: Since the company launched its "Strawberry" AI model family last week, touting so-called reasoning abilities with o1-preview and o1-mini, OpenAI has been sending out warning emails and threats of bans to any user who tries to probe how the model works.

Unlike previous AI models from OpenAI, such as GPT-4o, the company trained o1 specifically to work through a step-by-step problem-solving process before generating an answer. When users ask an "o1" model a question in ChatGPT, users have the option of seeing this chain-of-thought process written out in the ChatGPT interface. However, by design, OpenAI hides the raw chain of thought from users, instead presenting a filtered interpretation created by a second AI model. Nothing is more enticing to enthusiasts than information obscured, so the race has been on among hackers and red-teamers to try to uncover o1's raw chain of thought using jailbreaking or prompt injection techniques that attempt to trick the model into spilling its secrets.

This discussion has been archived. No new comments can be posted.

OpenAI Threatens To Ban Users Who Probe Its 'Strawberry' AI Models

Comments Filter:
  • We'll overlook the fact this "artificial intelligence" isn't smart enough to manage its own security.

    • They don't want you probing it, because the secret is actually horrific. All of the things you type are read to an array of heads in jars that quickly process the information and spit out the answers. Kind of like Futurama, but more slavery involved.

      • Re: (Score:2, Flamebait)

        by dgatwood ( 11270 )

        They don't want you probing it, because the secret is actually horrific. All of the things you type are read to an array of heads in jars that quickly process the information and spit out the answers. Kind of like Futurama, but more slavery involved.

        More likely a million contractors in some third-world country — kind of like slavery, but more capitalism involved.

  • by xpiotr ( 521809 ) on Wednesday September 18, 2024 @04:27PM (#64797583) Homepage
    Is the Strawberry name connected to ChatGPT:s inability to count the number of r in strawberry?
    Straight from the AI:s mouth [chatgpt.com]
  • corporate insecurity (Score:4, Interesting)

    by BishopBerkeley ( 734647 ) on Wednesday September 18, 2024 @04:39PM (#64797621) Journal
    Apple and Microsoft pay bounties for people to find bugs. Open AI, instead, goes out of its way to discourage users from finding bugs. How insecure are they about this "reasoning" system? Is strawberry too naive and fragile?
    • I wouldn't call probing AI systems as "finding bugs". The overwhelming majority of the probing of AI models isn't done by white hats for bug bounties or to make a better product. It's to get the AI bot to agree that Hitler was an all around nice guy so you can post the result on social media, or for competitors to determine what is a component of the underlying model.

      And before you said I Godwin'd this thread I invite you to scroll up where someone literally already used the holocaust to try and prove that

  • Where have I heard that before?

  • OpenAI hides the raw chain of thought from users, instead presenting a filtered interpretation created by a second AI model. ... OpenAI has been sending out warning emails and threats of bans to any user who tries to probe how the model works.

    "Pay no attention to the AI behind the curtain!"

    (Apologies to the Wizard of Oz [youtube.com].)

  • by gweihir ( 88907 ) on Wednesday September 18, 2024 @05:13PM (#64797713)

    No surprise. Their claims about that model are insane and disconnected from reality. Hence it is clear they are faking things. Obviously, they do not want people to fond out how.

    • Obviously, they do not want people to fond out how.

      Well obviously we can’t just have users deglazing the proverbial pan and tasting the special sauce inside, you could guess what went into it otherwise.

    • No surprise. Their claims about that model are insane and disconnected from reality. Hence it is clear they are faking things. Obviously, they do not want people to fond out how.

      This is gonna end up being another "there's a bunch of humans in a far away country answering" things. Isn't it?

      • by gweihir ( 88907 )

        Incidentally, that would show at least some respect for the classics in the scam area (even if no long-distance was involved): https://en.wikipedia.org/wiki/... [wikipedia.org]

        Funnily, "to tuerk" something still means "to fake it" in German: https://de.wikipedia.org/wiki/... [wikipedia.org]
        Hence this idea seems to actually have been known to a wider audience for some time now. But those that do not know history are doomed to repeat it...

        • Incidentally, that would show at least some respect for the classics in the scam area (even if no long-distance was involved): https://en.wikipedia.org/wiki/... [wikipedia.org]

          Funnily, "to tuerk" something still means "to fake it" in German: https://de.wikipedia.org/wiki/... [wikipedia.org] Hence this idea seems to actually have been known to a wider audience for some time now. But those that do not know history are doomed to repeat it...

          I've seen a lot of goofy contraptions at fairs over the years that were supposedly machines but actually had humans hidden in them. Even as a kid you could tell with most of them though.

          I do think it's funny how amazingly scammy our entire business world is right now. Add a tiny touch of tech and suddenly everybody thinks it's magic and turns off their thinking centers.

          • by gweihir ( 88907 )

            I do think it's funny how amazingly scammy our entire business world is right now. Add a tiny touch of tech and suddenly everybody thinks it's magic and turns off their thinking centers.

            Same here. To the degree this happens, it is also possible people did not actually ever start to think about things. Dark times.

  • They're giving Orwell's Ministry of Truth a run for their money with that branding.
    • by Meneth ( 872868 )
      I wonder if they could be sued for fraudulent branding.
  • how many r are in strrrrawberry?

    > There are three "r"s in "strawberry."

    that's not what I asked

    > I see what you did there! You were going for a playful twist on the word. In "strawberry," there are actually two "r"s, but I get that you were emphasizing the sound with the extra "r"s. Clever!

  • by Anonymous Coward

    Just when you thought it couldn't get any worse going from open source to closed without disclosing basic facts about your models to lobbying against open source. Now users are being threatened if they dare try and peek behind the curtain.

    OpenAI is an embarrassment.

  • "Open" AI ... laughable.
  • ... but I would think that any serious user would want to know how this thing is arriving at these types of answers, hell, for all we know it could just be another mechanical turk! Unlikely, but possible.

When your work speaks for itself, don't interrupt. -- Henry J. Kaiser

Working...