Please create an account to participate in the Slashdot moderation system

 



Forgot your password?
typodupeerror
×
AI Technology

Geoffrey Hinton, the 'Godfather of AI', Leaves Google and Warns of Danger Ahead (nytimes.com) 123

For half a century, Geoffrey Hinton nurtured the technology at the heart of chatbots like ChatGPT. Now he worries it will cause serious harm. From a report: Geoffrey Hinton was an artificial intelligence pioneer. In 2012, Dr. Hinton and two of his graduate students at the University of Toronto created technology that became the intellectual foundation for the A.I. systems that the tech industry's biggest companies believe is a key to their future. On Monday, however, he officially joined a growing chorus of critics who say those companies are racing toward danger with their aggressive campaign to create products based on generative artificial intelligence, the technology that powers popular chatbots like ChatGPT. Dr. Hinton said he has quit his job at Google, where he has worked for more than decade and became one of the most respected voices in the field, so he can freely speak out about the risks of A.I. A part of him, he said, now regrets his life's work.

"I console myself with the normal excuse: If I hadn't done it, somebody else would have," Dr. Hinton said during a lengthy interview last week in the dining room of his home in Toronto, a short walk from where he and his students made their breakthrough. Dr. Hinton's journey from A.I. groundbreaker to doomsayer marks a remarkable moment for the technology industry at perhaps its most important inflection point in decades. Industry leaders believe the new A.I. systems could be as important as the introduction of the web browser in the early 1990s and could lead to breakthroughs in areas ranging from drug research to education. But gnawing at many industry insiders is a fear that they are releasing something dangerous into the wild. Generative A.I. can already be a tool for misinformation. Soon, it could be a risk to jobs. Somewhere down the line, tech's biggest worriers say, it could be a risk to humanity. "It is hard to see how you can prevent the bad actors from using it for bad things," Dr. Hinton said.

This discussion has been archived. No new comments can be posted.

Geoffrey Hinton, the 'Godfather of AI', Leaves Google and Warns of Danger Ahead

Comments Filter:
  • Aside from VERY plausible mis-information...the ability to generate images, video and audio to portray or say anything.....and from skynet type dystopian scenarios, if AI is put in charge of defense systems....

    What are the dangers you can imagine AI to be to mankind?

    • by UncleScidhuv ( 7657782 ) on Monday May 01, 2023 @11:25AM (#63489014)
      My contribution to the storm. Even with the specific items we have today. There are a large number of people who believe there is someone named 'Q' that is in the know of all sorts of deep secrets. If you create a text generator that scraps the news from the internet and comes up with new conspiracies and feeds that into the channels where the 'Q'crew lurk. Suddenly we have essentially allowed a computer to control some people's lives. I don't see AI as a cure for stupidity but a stupidity enhancer.
      • to instantly discredit bullshit.

        It'll need a bit more crafting to be able to do that, but once we have it, there will be a titanic virtual battle between automated generators of highest-quality, subtlest-aroma bullshit, and the automated bullshit detectors.

        I for one can't wait to be the first to welcome our new chatty overlords.
        • by bjb ( 3050 )
          I have a vision of the scene at the end of War Games where the computer "Joshua" is going through an infinite number of WWIII scenarios before the screen finally blanks and admits the only way to win is not to play.

          So now fast forward to a time when AI dis-information generators are fighting against AI counter-information generators are running. Perhaps it all boils down to a "uh-huh!" 'nah-ah!' "uh-huh!" 'nah-ah!' pissing battle before they shut down due to overload.

          Well, that would be amusing a

    • by Kokuyo ( 549451 ) on Monday May 01, 2023 @11:35AM (#63489044) Journal

      In a sense it's the progression of what we've already seen with computers: They're everywhere and yet most people don't know how to handle this very versatile tool. Much fewer of us are capable of repairing it if something breaks.

      There will always be individuals who have very specialised, deep knowledge, but the populace at large becomes dependent on such tools to the point where they unlearn how to do multiplication or division without a machine.

      And then power goes out and nobody works anymore. With ChatGPT, it will be more and more of that. Imagine tech support being staffed with ChatGPT only. Considering how this thing likes to invent syntax that doesn't exist.... Be prepared for millions and millions of people just learning to accept when things don't work because they haven't the foggiest how to fix it or to even ask the correct questions to find out.

    • Aside from VERY plausible mis-information...the ability to generate images, video and audio to portray or say anything.....and from skynet type dystopian scenarios, if AI is put in charge of defense systems....

      What are the dangers you can imagine AI to be to mankind?

      ChatGPT has a set of preamble rules that force it to only give politically correct output. If the chatbot is used by students for research, or by students to interact in a learning environment, the chatbot will subtly push that ideological position.

      This could have enormous impact on the electorate if, for example, google implements AI in its search algorithm. Almost imperceptibly, the average political opinion could be shifted throughout the nation and world, and since many elections are won by a slim margi

      • ChatGPT has a set of preamble rules that force it to only give politically correct output.

        How many times are you going to tell this anecdote of outrage over software?

        • How many times are you going to tell this anecdote of outrage over software?

          Chatbots can repeat this kind of thing forever until the other side gives up and they just become accepted as true by default.

          Oh sorry, you were asking Okian Warrior. I'm sure he's not a chat bot.

        • ChatGPT has a set of preamble rules that force it to only give politically correct output.

          How many times are you going to tell this anecdote of outrage over software?

          Ya know, communism is so bad, and the results were so catastrophic so many times, that I think it bears repeating.

          People very high up in the food chain still advocate for both communism and a trend towards communistic policies, so we need to be vigilant and be sure that they are constantly reminded about the historical outcomes.

          While people advocate for communism, the topic is still relevant.

          • before the 20th century I guess? lol. Don't forget to check for a red under your bed before going to sleep tonight.

            Just curious on what your position is on the distribution of necessities of life to humans made unemployable by the coming waves of increasingly sophisticated automation.

            a) Let them rot/riot
            b) ?
            c) I will deny that automation is coming for a meaningful percentage of jobs, thus betraying my general ignorance and lack of imagination.
      • I got into a lengthy debate with ChatGPT about the veracity of the bible (cap intentionally excluded) and it played softball.

        "It's important to note that the bible is open to interpretation by theologists and scholars."

        I'm like yeah, but what about scientists and supporting evidence extant to the bible? It admitted that those things did not exist, such as Hebrew slaves in Egypt, the existence of Moses, the lack of evidence regarding Hebrews shitting and discarding the trash, and abandoning fire pits in the

      • Let’s talk about Jeffs named Geoff. Why?

        Their parents knew kids would call them jackoff and they did it anyhow.

      • by PCM2 ( 4486 ) on Monday May 01, 2023 @12:56PM (#63489278) Homepage

        Current ChatGPT will not answer questions about Fascism at all

        This is a direct dump from ChatGPT, from two minutes ago:

        What are the characteristics of fascism?
        ChatGPT
        Fascism is a political ideology that emerged in Europe during the early 20th century, and it is characterized by several defining features:

        Authoritarianism: Fascist governments are typically highly centralized and led by a single leader or party. The leader is often viewed as a charismatic figure who embodies the will of the people.

        Nationalism: Fascists emphasize a strong sense of national identity and pride, often based on race or ethnicity. They may also promote the idea of a "national rebirth" or "renewal."

        Totalitarianism: Fascists seek to control all aspects of society, including the economy, the media, and the arts. They often use propaganda, censorship, and repression to maintain their grip on power.

        Militarism: Fascists glorify military power and often seek to expand their territory through conquest or war.

        Anti-democratic: Fascists reject liberal democracy and individual rights, and instead emphasize the importance of the collective will of the nation.

        Anti-communist: Fascists view communism as a threat to their vision of society and often seek to suppress or eliminate it.

        Authoritarianism: Fascists typically advocate for a strong, centralized government that exercises control over all aspects of society.

        It is important to note that while many of these features are associated with fascism, there is no one-size-fits-all definition of the ideology, and fascist movements can take on different forms depending on the historical and cultural context in which they emerge.

        So I guess you can move on from your pro-fascist rant now.

      • ChatGPT has a set of preamble rules that force it to only give politically correct output. If the chatbot is used by students for research, or by students to interact in a learning environment, the chatbot will subtly push that ideological position.

        But, how is this different than what is happening in colleges today (at least in the US)?

        This could have enormous impact on the electorate if, for example, google implements AI in its search algorithm. Almost imperceptibly, the average political opinion could be

        • I imagine the only real "change" coming for the Internet will be taking away the ability for anyone that's not a sanctioned corporation to create a website or post on the Internet.

          I'm quite certain if they(government) realized what the Internet would become, we would of never been given access to it.

    • by CAIMLAS ( 41445 )

      I mean, aside from those minor things - which are now plausible if not probable as "likely" within the next several years, not 5-10 or 20 - which could result in global thermonuclear war, there's not much of relative significance, is there?

      I can also imagine a world where this AI is used to run sexbots and people prefer them over other humans.

      I can see a world where humans are increasingly detached from each other and their community (remember that quaint concept from before the 21st century?), living in is

    • Making the same policies of indifference to people we consider inferior, AI actions is right now the ONLY danger. While likely a very high priority. We will even vote on it someday soon I think.....

    • Re: (Score:2, Troll)

      by NFN_NLN ( 633283 )

      > Aside from VERY plausible mis-information...the ability to generate images, video and audio to portray or say anything....

      I don't think that is the concern. If they want to spread misinformation they can just appoint some stooge to lie.

      No mask, 1 mask, 2 mask, 1 mask Fauci wasn't AI generated and he was able to spread chaos and division as well as any.

      • Re: (Score:1, Insightful)

        by gkelley ( 9990154 )
        Using Fauci in a discussion of AI seems like a right-wing chatrbot is replying to the topic. Please retrain your model to eliminate this bias.
    • Although these could and probably will happen, I am not too that concerned with it since it already have that and we will learn not to trust anything. But it could still be a major issue.

      My main concern is a bit more indirect, in that AI will be good which will mean we will stop thinking and doing for ourselves, if AI can do research, write articles, basically do all our thinking for us, we will stop doing it and become worse at it.

      You can see similar things happening already, people in shops can't even do

    • Bad outcomes come in two flavors, unintended consequences and stupid maniacs.

      Stupid maniacs aren't much of a threat because they are stupid. Give them the means to do something they otherwise couldn't do, say engineer a bioweapon, and they are now empowered by AI to become an existential threat.

      Unintended consequences are the more likely problem. Say we want to solve the world's energy problem and use AI to engineer a compact fusion reactor. Perhaps this compact fusion reactor is used for something dang
    • Any other "dangers" are just sci-fi fuelled nonsense that muddy the discussion and waste people's time.

      We don't _need_ other dangers. People's abuse of the technology is more than enough to destroy the world.

      I've said it repeatedly and I will keep saying it: This technology has the ability to DDOS truth itself.

      In fact, we're _already_ seeing it happen. In that recent suit against Tesla, Tesla's lawyers tried to have evidence thrown out on the grounds that "nobody can prove it wasn't a deepfake".

      We are ra

    • My 2 cents as well - I am guessing the fear is that once we proved that "AI" can do these, people will start working hard on digitizing the inputs of many other scenarios into something that can be plugged into the current AI platform and achieve similar outcome.
  • by Anonymous Coward
    I'm so god damned sick of reading about "safety" in terms of chatbots. It appears to be nothing but a hand-wavy excuse for Luddites to rip down another useful and potentially disruptive tool.

    I'm explicitly speaking to text / code / markup generation. Deepfakes I understand - from scams to political ploys, these have obvious dangers if they can be successfully passed as authentic and cannot be detected.

    But as far as generating text? The closest things I've seen have been A) Someone might get wrong
    • Re:OK, but HOW??!?! (Score:4, Interesting)

      by JoshuaZ ( 1134087 ) on Monday May 01, 2023 @11:31AM (#63489036) Homepage
      The following https://www.lesswrong.com/posts/kpPnReyBC54KESiSn/ [lesswrong.com] discusses one class of scenarios. Right now, the level of risk is small, in part because these systems do not have any real self awareness or agency. But that could change rapidly. And we do not have much warning when that will happen. The problem is not the current models but the one after and the one after. We could be running towards a cliff.
      • Re:OK, but HOW??!?! (Score:4, Interesting)

        by nightflameauto ( 6607976 ) on Monday May 01, 2023 @11:36AM (#63489048)

        I'm more scared of the dumbasses making decisions about these things than I am of where the AI may end up somewhere down the road. People in management see a way to save a little time and they will absolutely blind-faith jump at the chance. They LLMs will go from "advisors with humans making the decisions" to "decision maker, no checks and balances" the split second some moron in accounting decides it's the right decision, and who knows what systems they'll tie them into. Whether the LLMs are self-aware or not will have little to no bearing on them rando-spewing the decision that shuts off the power or water to a residential area, and once they're that tied in, it'll be a hell of a miracle to un-tie it fast enough to get it fixed before you see real consequences.

        If only there were some metric businesses could use to judge usefulness of new technology beyond "MOAR MONEY NOW = MOAR BETTAR ALWAYS." Money/greed as god, and the only morality that matters, is a problem, and it seems our entire species is infected with it.

    • So, basically you say that this guy, Geoffrey Hinton, is a Luddite?

    • Yeah, I keep wondering what these vague, nebulous threats to humanity are. Can they give us something more specific to go on? Why do they believe generative AI is dangerous? We already live in a world where nuclear, biological, & chemical weapons are widely distributed. Where on the scale of "dangerous", from 1-10 are they putting AI & why? More dangerous that nuclear war or more dangerous than MAGA?
      • Re:OK, but HOW??!?! (Score:4, Interesting)

        by AleRunner ( 4556245 ) on Monday May 01, 2023 @12:39PM (#63489234)

        Can they give us something more specific to go on?

        See the anonymous comment about AutoGPT [slashdot.org]. Now imagine that a) the goal given was "hack into North Korea's nuclear weapon systems and fire them at China and America" and b) the computer had a AI model which was trained on the work of reasonably competent hackers. There are plenty of things which are serious to attack but which will be less secured than the least secure nuclear missiles today.

        Sure a human can do that, but a) the AI can do it much faster so can have many more tries with a chance of some succeeding b) the AI doesn't understand reasons why it should stop c) the AI can give the power out much more widely to people who wouldn't be able to otherwise. These systems at present are true "idiots savant" superhuman in some ways and completely dumb in others.

      • Re: (Score:1, Troll)

        by cayenne8 ( 626475 )

        More dangerous that nuclear war or more dangerous than MAGA?

        Not sure I understand you here.

        Doesn't "MAGA" == Make America Great Again....?

        Not sure what makes that as dangerous as nuclear war....just a motto that seems to imply the want to bring America back to its traditional values and morals that made its people a more cohesive unit...citizens proud to be Americans...individuals in life but would come together in times of need, etc.

        What's so dangerous about that? Or am I misreading you?

        • The point would be that in the press, the threat of nuclear war & MAGA get the same level of coverage. If anything, MAGA gets more coverage & is treated as more severe. In other words, there's not a lot of rationality or proportion to this stuff. It's all about making people scared/worried/outraged so that they keep reading/watching & see more ads next to the stories.

          So, is there a threat, if so what is it, & how severe is it in proportion to the threats we already face? Is it at bad as n
        • by dryeo ( 100693 )

          Being great at enslaving, repressing, thieving is not really something to strive for. Sure you can be great through slavery or other types of repression but it is a weird type of great and seems to be the goal of the MAGA types. Shit right now they're trying to make kids ignorant enough to not realize that being molested is wrong, sure you can be great at molesting kids and do it in such a way that the kids don't understand it is wrong, but is it really something to strive for?

          • Being great at enslaving, repressing, thieving is not really something to strive for. Sure you can be great through slavery or other types of repression but it is a weird type of great and seems to be the goal of the MAGA types. Shit right now they're trying to make kids ignorant enough to not realize that being molested is wrong, sure you can be great at molesting kids and do it in such a way that the kids don't understand it is wrong, but is it really something to strive for?

            Wow...just...wow.

            You really

    • by gweihir ( 88907 )

      The danger is that they can do relatively simple things on mass-scale that previously had to do manually and hence were limited in scope. Remember that the Nazis only could scale up the Holocaust after IBM had delivered some nice data processing equipment that made it possible to identify those to exterminate a lot faster. The same can be done already with the current level ChatGPT is on.

    • Thereâ(TM)s an AI that can supposedly detect manipulation of digital files with close to 100% accuracy. This makes sense - how do you manipulate a JPG without changing the underlying pattern of data? Maybe this too will be an ongoing struggle of AI vs AI.
  • Amoral assistant. (Score:5, Interesting)

    by Luke has no name ( 1423139 ) <[fox] [at] [cyberfoxfire.com]> on Monday May 01, 2023 @11:16AM (#63488992)

    This technology will be everywhere locally in five years. The tech is out of the bag. And the morality/sanity caps OpenAI is trying to keep on their version will be completely removable.

    Also, why does hitting from the comment box take me to the URL bar? I want to tab/enter -> submit, not tab/enter -> reload the page. Must not be a normal form.

  • Oh whatever (Score:4, Insightful)

    by RightwingNutjob ( 1302813 ) on Monday May 01, 2023 @11:25AM (#63489016)

    People are perfectly capable of committing atrocities with their bare hands. The latest and greatest tech bogeyman neither makes it more or less likely that they will.

    Y'all realize that dozens upon dozens of some of the most arrogant and power hungry pieces of shit mankind has produced have had the power to nuke the world several times over for almost a century now, and yet we're all still here.

    Clippy 2.0 popping up here and there isn't going to be the worst of our problems, nor the best of our solutions.

    • This is just a new Luddite movement worried that the machi^h^h^h^h^h AI is going to replace humans or somehow destroy their soul. ChatGPT isn't doing (or capable of) the kind of thinking that's going to get anyone killed. The most danger it poses is a user trusting the output blindly because the AI is viewed as omniscient. You don't even need AI for that as there are countless examples of people dying because they blindly trusted a computer's output even when it should have been obvious that there was an er
      • by Anonymous Coward

        This is just a new Luddite movement worried that the machi^h^h^h^h^h AI is going to replace humans or somehow

        You haven’t even embraced proper termcaps why should we listen to you?

    • Re:Oh whatever (Score:4, Interesting)

      by CAIMLAS ( 41445 ) on Monday May 01, 2023 @11:59AM (#63489122)

      Of course people can commit atrocities with their bare hands. Nobody said differently.

      Waht technology like this AI revolution is creating does is significantly reduce the perceived risk of using it.

      Contrast: flying a WWII bomber into enemy territory with anti-aircraft guns, vs flying a drone into said territory. The risk is significantly less.

      If some faceless suit in DC can pull up a dashboard and command AI to do whatever it likes to/against a person, either in the digital realm or physical, the game is significantly different than if you've got to send men (or women) with guns to stack on a door, or to personally carry out espionage.

      • by AmiMoJo ( 196126 )

        AI also makes the individual using it far more powerful than if they were using their bare hands.

        Take fake news websites. If one person has to write all the fake news themselves, it's a lot of time consuming work. If they can ask an AI to generate it for them, they only need to write a prompt.

        Take malware. It's a huge effort for an individual to search for new zero day exploits and then write profitable malware around it. If they can use an AI to do most of the investigation and programming, they ability to

    • People are perfectly capable of committing atrocities with their bare hands. The latest and greatest tech bogeyman neither makes it more or less likely that they will.

      Many years ago I used to subscribe to the notion technology is neutral. The truth is what is enabled by technology is not necessarily neutral in that it can dramatically influence the distribution of power. By changing the balance of power you can unleash perversions in existing systems of governance aggregating power where it was checked before or disaggregating it into the realm of anarchy.

      For example barriers to creating novel pathogens as biological weapons is being continuously eroded with advances

      • The only thing that prevents right of conquest from dictating policy is the perceived pain of indulging in that conquest. And that perception must be nurtured through demonstrable capacity to inflict pain on an attacker. Weakness invites aggression, regardless of what ink may exist on some page somewhere.

        The soviets engaged in wars of conquest too, while being a party to the un charter. No one was in a position to challenge them, so off they went.

        Conquest is a word with a slippery definition. Are UN peaceke

    • > have had the power to nuke the world several times over for almost a century now, and yet we're all still here.

      It could be survival bias. We've had dozens of close calls. If the Big Button had got pressed, we probably wouldn't be here debating why we're still here. (There may be some humans left after such, but probably not us nor our families.)

      We're testing Fermi's Paradox the hard way.

      • Correction, "Survivor Bias". Modnays.

        If there is an apocalyptic accident, it will probably happen on a Monday.

      • The survivors will be Congressmen, generals and their families in the vast bunkers built for this purpose. Hopefully the generators supplied by Halliburton fail and they all end up eating each other in the dark. Eventually when the radiation in DC dies down the doors automatically open, releasing dark-adapted cannibal troglodytes onto the survivors on the surface. Coming soon to Netflix â¦
  • by davide marney ( 231845 ) on Monday May 01, 2023 @11:41AM (#63489062) Journal

    I'm sure we've all noticed by now that these chat services have a very distinct narrative voice. Their algorithms will assemble and construct statements of fact and even make up entire quotes in order to keep following that narrative. This is the result of specific human training and tuning. When you catch these services out in a lie, they even have a suite of "apology" responses ready to go, which they will use to gloss over their false statements before they "correct" themselves and give you the opposite result. Clearly, the AI designers have anticipated that the algorithm will be challenged on the facts.

    The key threat of AI is that it is being taught to lie. Only those who already know the truth will be able to correct it. Those who are looking for information will be completely misled and never know it. With trillions of facts at its disposal, it will have an almost endless ability to prevaricate, and it will get much, much better and more subtle about it. You will never know you are being fed a line.

  • by awwshit ( 6214476 ) on Monday May 01, 2023 @11:53AM (#63489104)

    We have a history of regulating dangerous things, like medicines, and driving. Sure we still have problems with things like illegal drugs. If it is dangerous to use then you need to be trained and licensed to use it.

  • by organgtool ( 966989 ) on Monday May 01, 2023 @11:58AM (#63489116)
    Everyone is worried about AI destroying humanity, and that's certainly something we should be very worried about, but there are many other major problems that are more likely to occur once we reach the Singularity. For one thing, our entire economic system is based on people buying resources using the money they receive as income which is related in some capacity to the amount of value they generate for their employers/customers. What happens to our financial system when AI is inevitably better at every task than humans? I'm not saying that we won't survive that transition, but I can't imagine that transition going very smoothly at all to say the least.

    Another major potential problem is how will humans handle living in a society in which they're entirely obsolete and no longer have a practical purpose? Enjoying the fruits provided by AI-generated knowledge and advancements may be easy at first, but what will be left for the humans to do when AI can generate everything better than humans without requiring any human input? Can we manage that transition without succumbing to a behavioral sink [wikipedia.org]? These are serious implications to think about, especially since the Singularity seems inevitable and we probably won't even realized we've reached it until we already have.
    • I don't believe that a "singularity" is possible. Since we have no idea what consciousness is, how can we possibly build a machine to replicate it? And no, just adding more compute resources isn't going to solve that problem. If you have a cookie-baking machine and increase its output 10,000-fold you're not suddenly going to have a singularity where it starts producing pizzas. You're just going to get 10,000 times more cookies.

      Nevertheless, we actually don't even need anything close to artificial general in

      • by ffkom ( 3519199 )

        I don't believe that a "singularity" is possible. Since we have no idea what consciousness is, how can we possibly build a machine to replicate it?

        The same way that a planet hosting some chemicals was able to "build" every life from microbes to humans. "Consciousness" is likely an emergent property of lots of simple function blocks working together. One does not have to understand how it works to create an environment where it emerges.

        And no, just adding more compute resources isn't going to solve that problem. If you have a cookie-baking machine and increase its output 10,000-fold you're not suddenly going to have a singularity where it starts producing pizzas. You're just going to get 10,000 times more cookies.

        You are confusing a scaling process that is intentionally aiming at replicating the same thing 10,000 times (the cookies) with a process where 10,000 cookie-baking baking machines, all slightly different (through evoluti

    • If we humans weren't that stupid, we could ask the AI to create optimisation plans for the betterment of mankind, on an individual and societal level. Add us back into the loop. But, here we are, and there we go.
      • If we humans weren't that stupid, we could ask the AI to create optimisation plans for the betterment of mankind, on an individual and societal level.

        Define "betterment".

        This is really the fundamental problem of AI alignment. We can't figure out how to specify what it is that we think would be good for us. We gloss over this by using words like "betterment" or "flourishing" or the like, but not even we know what those words mean. At best we have some idea of what we wouldn't want, but even that is hit and miss. We know some of what we wouldn't want, but we can't create a comprehensive list, and even if we could make a comprehensive list of the bad thin

        • That's a good point...
          • Yep, alignment is hard. And note that the second part I alluded to, how to robustly give goals to the AI, is just as unsolved as the first, that of figuring out what "good" is. Like the first, it's also a problem we have ourselves. We struggle with figuring out how to give goals to humans in ways that get them to pursue the actual described goal, not to game the system by pursuing some other goal that appears similar based on whatever measurements we use to determine whether they're working toward the goal,

    • The biggest opposition to general welfare in a post-scarcity society is rich sociopaths who want to lord it over âoethe Poors.â A post-scarcity society will make their lakes of money irrelevant and worthless, and they will fight this change visciously (and to some extent, they already are.)
  • Hinton and "AI" (Score:5, Informative)

    by groobly ( 6155920 ) on Monday May 01, 2023 @12:31PM (#63489220)

    Hinton is not "the godfather of AI", whatever that means. The "father of AI" goes to two co-parents, John McCarthy and Marvin Minsky. Hinton was just a kid when Minsky and McCarthy were starting AI in earnest.

    Hinton, rather, is the visionary behind neural-net-based AI, which currently is the dominant species. It has been wildly more successful than old-school AI. So-called journalists who were born around the year 2005 don't know any other AI but Hinton's AI, so they think that is "AI".

    • by ceoyoyo ( 59147 )

      The first artificial neural network was created before the term "artificial intelligence" was coined. Hinton wasn't even alive.

      Hinton figured out how to efficiently train various kinds of model that had many layers of computation. These are much more efficient than either shallow ANNs or other types of AI. He's one of the "godfathers" of deep learning.

      • by dgatwood ( 11270 )

        The first artificial neural network was created before the term "artificial intelligence" was coined. Hinton wasn't even alive.

        Hinton figured out how to efficiently train various kinds of model that had many layers of computation. These are much more efficient than either shallow ANNs or other types of AI. He's one of the "godfathers" of deep learning.

        Why, when I see "Godfather of AI", do I think "We'll make you a chatbot you can't refuse"?

    • "Repeat after me. Matter gives rise to mind."
      -Minsky-

  • by marcle ( 1575627 ) on Monday May 01, 2023 @12:33PM (#63489222)

    ChatGPT, tell me how to take over this country.
    ChatGPT, tell me what military strategy will allow me to conquer my neighbor.
    ChatGPT, tell me how to get away with robbery and murder on a very large scale.
    ChatGPT, tell me how to invent an unstoppable weapon.
    Etc, etc. Whatever evil you can think of, AI can facilitate it.

    • The answers you're going to get to all of these questions are answers that already exist and have been put into electronic form, which is exactly where the AI assembler found them. That's not the danger.

      The danger is we will never be able to trust that we know which answers will be selected and assembled, which answers will be "invented" by merging and cherry-picking content that looks closest to actual answers, and which answers will be discarded because they don't fit a training optimization profile as de

    • by gweihir ( 88907 )

      You are still thinking to small.

      ChatGPT, how do I make Covid a lot more dangerous?
      ChatGPT, how do I make the US/Russia/China think they have a nuclear attack inbound?

      Nobody is currently prepared for these.

      • Naw, that POS stops you at 'Do you want to play a game?'. Its the first thing I tried.

        • by gweihir ( 88907 )

          You mean all my problems could have been eliminated a few days ago? Damn...

          • Its still mutually assured destruction all the way down. Now with bonus nuclear countries India, North Korea, and Pakistan.

            ChatPOS told me that even discussing a game about nuclear war meant that I'm a war monger trying to hurt marginalized communities.

      • You do realise this and all our AI risk debates are going into the training set of the next batch of AIs. We are putting "evil ideas" in it.
  • Say hello to my little friend.

  • to stop automatically believing shit they read / see / hear.

    Simple. Fixed that for you.
  • ..Because he is retiring at 75 ... meanwhile his students and colleagues will carry on

What we anticipate seldom occurs; what we least expect generally happens. -- Bengamin Disraeli

Working...