Please create an account to participate in the Slashdot moderation system

 



Forgot your password?
typodupeerror
×
AI Google Technology

DeepMind Self-training Computer Creates 3D Model From 2D Snapshots (ft.com) 27

DeepMind, Google's artificial intelligence subsidiary in London, has developed a self-training vision computer that generates 'a full 3D model of a scene from just a handful of 2D snapshots," according to its chief executive. From a report: The system, called the Generative Query Network, can then imagine and render the scene from any angle [Editor's note: the link maybe paywalled; alternative source], said Demis Hassabis. GQN is a general-purpose system with a vast range of potential applications, from robotic vision to virtual reality simulation. Details were published on Thursday in the journal Science. "Remarkably, [the DeepMind scientists] developed a system that relies only on inputs from its own image sensors -- and that learns autonomously and without human supervision," said Matthias Zwicker, a computer scientist at the University of Maryland who was not involved in the research. This is the latest in a series of high-profile DeepMind projects, which are demonstrating a previously unanticipated ability by AI systems to learn by themselves, once their human programmers have set the basic parameters.
This discussion has been archived. No new comments can be posted.

DeepMind Self-training Computer Creates 3D Model From 2D Snapshots

Comments Filter:
  • I'm not sure why this is a big deal. MS had the tech for this about 5 years (Send us a buncha 2d pictures and we'll turn it into a VR set) extrapolating to models isn't that far of a reach.
    • by Anonymous Coward

      Adobe has been doing it for years as well.

    • I'm not sure why this is a big deal. MS had the tech for this about 5 years (Send us a buncha 2d pictures and we'll turn it into a VR set) extrapolating to models isn't that far of a reach.

      It sounds like the key difference here is they're predicting parts of the scene they haven't seen, such as what the other side of an object they haven't seen looks like.

      I don't know if they do that just based on clues like shadows, or the system says "that looks like the front of a sphere, therefore I can assume it's round on the other side as well."

      • by Kjella ( 173770 )

        It sounds like the key difference here is they're predicting parts of the scene they haven't seen, such as what the other side of an object they haven't seen looks like.

        Yeah, the key feature here seems to be extrapolation from a very small number of observations. Say you're in a park and you're snapping a few photos and from that you're trying to reconstruct a 3D model of the park. You don't have nearly enough data to do that properly, but you can make an educated guess. At least that seems to be the focus to me, how well can it fill in the blanks and make some copy-paste assumptions.

    • I was surprised, because I'd read that it was difficult. Specifically it was about creating a bump map from a photo of things like those decorative carved panels you sometimes see on buildings.

      It's something humans can do quite easily.

    • This is something completely different than Photosynth. Completely.
      At least just glance at TFA.

  • not unanticipated (Score:5, Insightful)

    by phantomfive ( 622387 ) on Friday June 15, 2018 @12:20PM (#56790510) Journal
    Not only was that anticipated, and not only have computers been "teaching" themselves for years in AI once the basic parameters are set, that is exactly what neural networks were DESIGNED to do nearly 50 years ago when they were invented
    • by mbkennel ( 97636 )
      No, neural networks have not been 'teaching themselves' to perform this many cognitively impressive capabilities **without significant detao;ed human-labeled data** for 50 years.

      The unsupervised or weakly supervised achivements are new.
  • How well does it work on 2D anime drawings?

  • The system, called the Generative Query Network, can then imagine and render the scene from any angle [Editor's note: the link maybe paywalled; alternative source],

    Why didn't you just use the definitely non-paywalled source?

C'est magnifique, mais ce n'est pas l'Informatique. -- Bosquet [on seeing the IBM 4341]

Working...