Follow Slashdot stories on Twitter

 



Forgot your password?
typodupeerror
×
AI Technology

Researchers Build AI That Builds AI (quantamagazine.org) 30

By using hypernetworks, researchers can now preemptively fine-tune artificial neural networks, saving some of the time and expense of training. From a report: Artificial intelligence is largely a numbers game. When deep neural networks, a form of AI that learns to discern patterns in data, began surpassing traditional algorithms 10 years ago, it was because we finally had enough data and processing power to make full use of them. Today's neural networks are even hungrier for data and power. Training them requires carefully tuning the values of millions or even billions of parameters that characterize these networks, representing the strengths of the connections between artificial neurons. The goal is to find nearly ideal values for them, a process known as optimization, but training the networks to reach this point isn't easy. "Training could take days, weeks or even months," said Petar Velickovic, a staff research scientist at DeepMind in London. That may soon change.

Boris Knyazev of the University of Guelph in Ontario and his colleagues have designed and trained a "hypernetwork" -- a kind of overlord of other neural networks -- that could speed up the training process. Given a new, untrained deep neural network designed for some task, the hypernetwork predicts the parameters for the new network in fractions of a second, and in theory could make training unnecessary. Because the hypernetwork learns the extremely complex patterns in the designs of deep neural networks, the work may also have deeper theoretical implications. For now, the hypernetwork performs surprisingly well in certain settings, but there's still room for it to grow -- which is only natural given the magnitude of the problem. If they can solve it, "this will be pretty impactful across the board for machine learning," said Velickovic.

This discussion has been archived. No new comments can be posted.

Researchers Build AI That Builds AI

Comments Filter:
  • Tuning? (Score:4, Interesting)

    by jythie ( 914043 ) on Wednesday January 26, 2022 @04:59PM (#62209961)
    So.. trying to make sense of the actual piece.. initially it sounded like they were talking about just architecture or hyperparameters, but later it sounds like it is also pre-populating the weights since they were comparing it to things like resnet? I am really unclear. Though I am terribly amused that they seem worried about a black box creating a black box with no way to explain the mistakes.. the horse is kinda out of the barn on that one.
    • Re:Tuning? (Score:4, Interesting)

      by Thelasko ( 1196535 ) on Wednesday January 26, 2022 @05:37PM (#62210051) Journal
      Yeah, my first thought is they were tuning hyperparameters [wikipedia.org] too. (Hyperparameters are settings that tell the training algorithm how to search for coefficients) My first time tuning an ANN I thought it was kind of obvious to do so. I wound up creating a simple search algorithm to calibrate those parameters. The whole hyperparameter tuning process is kinda meta.

      Most machine learning articles use so much jargon, it's difficult to tell what's a breakthrough and what's bullshit.
      • by narcc ( 412956 )

        Most machine learning articles use so much jargon, it's difficult to tell what's a breakthrough and what's bullshit.

        It's safe to assume that any given paper is more bullshit than breakthrough.

        On the use of jargon, this one reads a lot like a student paper. Jargon should be used only when it adds clarity or is otherwise necessary. Language overall is a bit unusual, with a lot of needless adjectives. They also hype their results in the paper, which is a red flag.

        I really don't like their use of 'experiment' and 'hypothesis', but this isn't unusual for CS papers.

        Just a first impression. The paper is here [arxiv.org], if anyone want

  • by IonOtter ( 629215 ) on Wednesday January 26, 2022 @05:05PM (#62209969) Homepage

    Hmmmm.

    "Hungry for data and power", "carefully tuning values" and "overlord".

    Yeah, see, those are three phrases and/or concepts that have absolutely no place in the context of artificial intelligence.

    It's almost as if they're taking their cues FROM Hollywood.

  • by garyisabusyguy ( 732330 ) on Wednesday January 26, 2022 @05:06PM (#62209971)

    what could go wrong with that?

    Just remember the Real Rule of Robotics, "Humans are fascinating and we want to make sure they are having a good time"

  • aka one-trick pony (impressive sure), with the potential to learn another trick or two, for the next round of press release and funding/grant applications
    • by gweihir ( 88907 )

      Pretty much yes. A lot of research has degraded from actually being useful to giving the appearance of being useful.

  • by crunchygranola ( 1954152 ) on Wednesday January 26, 2022 @05:10PM (#62209983)

    Researchers Build AI That Builds AI

    But do they have an AI to build that?

    • by narcc ( 412956 )

      They don't even have the first one. Pop sci AI article headline is bullshit. Shocking, I know.

  • Modern machine learning benefitted from two recent advances:

    * GPU-based neural network processing
    * Faster learning algorithms, so networks could converge in a reasonable time

    See here [machinelea...owledge.ai] for a historical timeline.

    GPU use started around 2008.
    Theoretical work that sped up learning in 2006 and 2011.

    • I would probably rephrase "faster learning algorithm" to "a deep network training algorithm", meaning the discovery of the algorithm (by Hinton and group) opened the floodgates.

      Prior to that there wasn't really any good algorithm for this problem, other than evolving the network with gradient descent type techniques (sure that is technically an algorithm, but more of a brute force thing of trying random variations vs an actual update formula)
  • I heard you liked AI so I added AI to your AI.

    At least we will soon be able to put the crackpot theories of paranoid futurists to the test. Of course I hope those guys were all crackpots ...

  • by UnknowingFool ( 672806 ) on Wednesday January 26, 2022 @05:17PM (#62210005)
    This is how you get Skynet.
    • I agree. Just wait for this "AI" to get into the military computers and Skynet will take over the world.
  • It oversees other AI, why don't we call it Skynet?
    • Westworld is a closer approximation. Robots designed better robots. Several generations later, humans no longer understood how they worked.

      Think how hard it is to detail how biological life works. Now imagine how intricate and specialized and interleaved sections of a giant AI brain might get.

  • by Anonymous Coward

    Use an AI to build an AI to build an AI to build an AI... it's turtles all the way down.

    What used to take humans days, weeks, months, to royally screw up, the AI can do so in milliseconds.

    Faster than you can pull the plug before the genie escapes the bottle.

    The clowns of AI are like the early scientists of nuclear energy. Only difference, the fallout from the nuclear energy scientists was somewhat contained.

  • And now he's going to build the Bitch-Brian of AI.

  • ... so I trained an AI to train AIs.

  • Because that's how you get Skynet ...

  • You will never convince me that "hypernetworks" and "hyperparameters" are real things and not something made up for the new Cyberpunk expansion.
  • "It's an anti-terminator... terminator?"

Don't panic.

Working...