Become a fan of Slashdot on Facebook

 



Forgot your password?
typodupeerror
×
Supercomputing

Some Scientists Question Whether Quantum Computer Really Is Quantum 170

gbrumfiel writes "Last week, Google and NASA announced a partnership to buy a new quantum computer from Canadian firm D-Wave Systems. But NPR news reports that many scientists are still questioning whether new machine really is quantum. Long-time critic and computer scientist Scott Aaronson has a long post detailing the current state of affairs. At issue is whether the 512 quantum bits at the processor's core are 'entangled' together. Measuring that entanglement directly destroys it, so D-Wave has had a hard time convincing skeptics. As with all things quantum mechanical, the devil is in the details. Still it may not matter: D-Wave's machine appears to be far faster at solving certain kinds of problems (PDF), regardless of how it works."
This discussion has been archived. No new comments can be posted.

Some Scientists Question Whether Quantum Computer Really Is Quantum

Comments Filter:
  • Does it really matter so long as it does what it says on the tin? It works faster, surely that's all that matters?

    • by Anonymous Coward on Wednesday May 22, 2013 @08:44AM (#43793295)

      You really need to RTFA. It's slower than an optimized implementation of the same thing on a classical computer (and one that costs a lot less than $10m).

      • by Certhas ( 2310124 ) on Wednesday May 22, 2013 @09:20AM (#43793661)

        Indeed, the summary is misleading.

        Citing from Aaronsons blog:

        Among the many interesting comments below, see especially this one by Alex Selby, who says he’s written his own specialist solver for one class of the McGeoch and Wang benchmarks that significantly outperforms the software (and D-Wave machine) tested by McGeoch and Wang on those benchmarks—and who provides the Python code so you can try it yourself.

        and

        As I said above, at the time McGeoch and Wang’s paper was released to the media (though maybe not at the time it was written?), the “highly tuned implementation” of simulated annealing that they ask for had already been written and tested, and the result was that it outperformed the D-Wave machine on all instance sizes tested. In other words, their comparison to CPLEX had already been superseded by a much more informative comparison—one that gave the “opposite” result—before it ever became public. For obvious reasons, most press reports have simply ignored this fact.

        In other words, if it works, it works, except that it doesn't.

        • by unrtst ( 777550 )

          ...been superseded by a much more informative comparison—one that gave the “opposite” result—before it ever became public.

          Sounds eerily familiar to every discussion of quantum mechanics I've heard! ("opposite" result; gave it before even knew it existed; etc).

      • Here is a link to a good article.
        There were 3 tests with D-Wave going against a generic algorithm.
        It tied on 2 or the 3 tests, but beat the generic algorithm running 3,600 times faster.
        However, if it went against a specialized algorithm it was just as fast.

        http://www.economist.com/news/science-and-technology/21578027-first-real-world-contests-between-quantum-computers-and-standard-ones-faster [economist.com]

    • by RDW ( 41497 ) on Wednesday May 22, 2013 @08:45AM (#43793297)

      You may say that now, but wait until PETA find out about the number of cats and flasks of cyanide their prototype gets through every month...

    • Read the blog post (Score:5, Interesting)

      by oGMo ( 379 ) on Wednesday May 22, 2013 @08:48AM (#43793327)

      The problem is that it's not faster, and while there's a study that concludes it is, the blog post specifically invalidates this:

      Namely, the same USC paper that reported the quantum annealing behavior of the D-Wave One, also showed no speed advantage whatsoever for quantum annealing over classical simulated annealing. In more detail, Matthias Troyer’s group spent a few months carefully studying the D-Wave problem—after which, they were able to write optimized simulated annealing code that solves the D-Wave problem on a normal, off-the-shelf classical computer, about 15 times faster than the D-Wave machine itself solves the D-Wave problem! Of course, if you wanted even more classical speedup than that, then you could simply add more processors to your classical computer, for only a tiny fraction of the ~$10 million that a D-Wave One would set you back.

      About the paper claiming it's faster:

      As I said above, at the time McGeoch and Wang’s paper was released to the media (though maybe not at the time it was written?), the “highly tuned implementation” of simulated annealing that they ask for had already been written and tested, and the result was that it outperformed the D-Wave machine on all instance sizes tested. In other words, their comparison to CPLEX had already been superseded by a much more informative comparison—one that gave the “opposite” result—before it ever became public. For obvious reasons, most press reports have simply ignored this fact.

      • Sounds like it really does not matter then; who cares whether its quantum or not when it provides less value at a higher price?

        • by MozeeToby ( 1163751 ) on Wednesday May 22, 2013 @09:47AM (#43794023)

          Because if it is quantum it's a generation 0 (barely out of prototype) implementation going up against a generation... oh I don't know... 30+ classical computer. If it's not quantum, if it's basically an ASIC chip designed to solve simulated annealing problems (intentionally or not), it's worthless even as research. What they are selling is a research and training system, so that engineers can learn what kinds of problems can be solved on the hardware that will, presumably, get much more powerful going forward.

          Look at it this way, the current D-wave machine has 512 qbits and a modern PC can match it's speed. Double the qbits and you end up with a simulation space several million times larger, the 15x faster is going to seem laughable when the problem you are solving is trillions of times larger and the D-Wave solves in constant time while your PC runs an algorithm that's O(n^2). If, if, what D-wave is selling is using quantum affects.

          • by ceoyoyo ( 59147 )

            That's the problem - nobody is sure if it really IS a quantum computer. If it were just a slow quantum computer, it would be interesting. If it's a slow classical computer, it's not. Some people would like to know which it is before they buy one and their engineers invest a lot of time "learn[ing] what kinds of problems can be solved on the hardware."

            Also the "get much more powerful going forward" is in doubt. There's a lot of expert criticism that D-Wave's approach cannot be easily scaled up because th

          • It's pretty obvious who is and who is not reading the article here:

            In short, there seems to be no evidence, at present, that the D-Wave machine is going to overtake simulated annealing for any instance size.

            The author concedes that it is possible that this may happen, but:

            Well, I concede that almost anything is possible in the future—but “these experiments, while not supporting D-Wave’s claims about the usefulness of its devices, also don’t conclusively disprove those claims”

            • Aaronson's language is so political. He seems to have rabbit ears, more concerned with what people are saying than what he's doing. The real question is why can't we explore both D-Wave's approach and "academic QC programs" in parallel? Economics is not a good reason; economics should serve the advance of knowledge, not throttle it.

              • Funding has to come from somewhere, and fraud in a field tends to end future funding in that field, even if it holds promise. Case in point: organic semiconductors took a huge funding hit after the Bell labs fiasco. Another case in point: public funding of bubble fusion research was basically banned after the Oak Ridge controversy, even though the basic principle holds promise.

        • Well, that's why we have tech, obviously, so that we can make the machine that costs $10M and makes the 'BOING" noise. /s

    • by Qzukk ( 229616 )

      Does it really matter so long as it does what it says on the tin?

      But if you put in the wrong figures, do you get the correct answer?

    • by 9jack9 ( 607686 )

      Does it really matter so long as it does what it says on the tin? It works faster, surely that's all that matters?

      Nope. Need to know how it works. Need to be able to examine every aspect of it. Need to be able to model it and predict its behavior in different situations.

    • by Hentes ( 2461350 )

      It only works for certain types of problems. You won't break RSA with it.

  • by Stolpskott ( 2422670 ) on Wednesday May 22, 2013 @08:39AM (#43793229)

    So they know where the D-Wave system is, but they cannot definitively measure whether it is actually a quantum computer or not...

    • This crazy scientist kidnapped my cat and put it in a box! See if the D-Wave computer can tell me if my cat is alive or not.

      • This crazy scientist kidnapped my cat and put it in a box!

        I have a crazy cat that regularly kidnaps physicists and puts them in boxes. The cat is very skeptical of quantum mechanics though, as the physicists always wind up dead.

    • by hAckz0r ( 989977 )
      Yes. They first started with particles, then moved on to small covalently bonded molecules. Soon after they perfected a superposition of Buckyballs they eventually went for the entire machine. Its amazing to watch the confusion among scientists as the entire structure of the machine goes into a state of superposition with being real and not real all at the same time.
      • by plover ( 150551 ) on Wednesday May 22, 2013 @10:19AM (#43794335) Homepage Journal

        Its amazing to watch the confusion among scientists as the entire structure of the machine goes into a state of superposition with being real and not real all at the same time.

        All quantum computers are real, unless declared integer.

  • by dingen ( 958134 ) on Wednesday May 22, 2013 @08:40AM (#43793239)

    A real quantum computer both is and isn't at the same time.

  • by Greyfox ( 87712 ) on Wednesday May 22, 2013 @08:43AM (#43793289) Homepage Journal
    Their computer works not by quantum entanglement but by magic.
    • by RevDisk ( 740008 ) on Wednesday May 22, 2013 @08:59AM (#43793459) Journal
      Is there really any difference between quantum entanglement and magic?
      • Re: (Score:2, Funny)

        by Anonymous Coward

        Is there really any difference between quantum entanglement and magic?

        You may have to ritually sacrifice the occasional intern to keep magic working, but otherwise, no.

      • by JoshuaZ ( 1134087 ) on Wednesday May 22, 2013 @09:22AM (#43793677) Homepage

        Is there really any difference between quantum entanglement and magic?

        Yes. There's this tendency to view entanglement as spooky, magical, and hard to understand. But this really isn't the case and is more due to the confusing way that quantum mechanics if often taught, as a series of counterintuitive results tacked on to classical physics. If one adjusts one's perspective to think of quantum mechanics more as the consequences of using a 2-norm and looking then at the structure imposed on vectors by unitary transformations, things make a lot more sense. Scott Aaronson(mentioned in the summary above) has a book out recently on just this subject "Quantum Computing since Democritus" which is aimed at explaining these issues to people outside is field but with a comfortable background in other technical fields- essentially no more than some linear algebra, basic probability and complex numbers. The book is highly readable and Scott is a very funny writer, so there are a lot of amusing asides.

        • Re: (Score:3, Funny)

          by Anonymous Coward

          >using a 2-norm and looking then at the structure imposed on vectors by unitary transformations

          Like, obviously!

        • Re: (Score:2, Insightful)

          by Anonymous Coward

          " If one adjusts one's perspective to think of quantum mechanics more as the consequences of using a 2-norm and looking then at the structure imposed on vectors by unitary transformations, things make a lot more sense."

          Which means absolutely nothing to the vast majority of people; hence spooky, magical and hard to understand.

          • I think my mathematics and quantum physics are not bad (although it were genuinely strong, perhaps I would have gone further than ABD), and I have trouble making "a lot more sense" of "using a 2-norm". Yes, I understand that the wave functions are complex, but that is not enlightening without a lot of math.

            Most "spooky" results are simply the result of not accepting that everything is a wave form, and waves cannot understand that they are not supposed to be stretched across space. Which is another way of

        • You can describe magic with math too, and it starts to make a lot more sense.

  • Then we not only get a useful machine, we eventually get new science in the bargain. I *like* it!

  • Comment removed (Score:5, Informative)

    by account_deleted ( 4530225 ) on Wednesday May 22, 2013 @09:09AM (#43793559)
    Comment removed based on user account deletion
    • Re:it is and it isnt (Score:5, Informative)

      by JoshuaZ ( 1134087 ) on Wednesday May 22, 2013 @09:26AM (#43793727) Homepage
      This is the sort of thing where it helps to read Scott's post. He specifically discusses the primary claim here:

      Namely, the same USC paper that reported the quantum annealing behavior of the D-Wave One, also showed no speed advantage whatsoever for quantum annealing over classical simulated annealing. In more detail, Matthias Troyer’s group spent a few months carefully studying the D-Wave problem—after which, they were able to write optimized simulated annealing code that solves the D-Wave problem on a normal, off-the-shelf classical computer, about 15 times faster than the D-Wave machine itself solves the D-Wave problem! Of course, if you wanted even more classical speedup than that, then you could simply add more processors to your classical computer, for only a tiny fraction of the ~$10 million that a D-Wave One would set you back.

      • by Rich0 ( 548339 )

        I'm not sure that this really helps clarify whether this is actually a quantum computer.

        The question is whether the D-Wave is a quantum computer, not whether it is faster than a classical computer. A quantum computer that is slower than a desktop calculator is still useful for research purposes.

    • The problem is calling it a "Computer", which it is not really.

      A better analogy might be to call it a like on some old school computers, a math co-processor. A math co-processor wasn't a "computer" but rather a processor that offloaded certain tasks that it could handle more efficiently.

      I see this as a similar situation. It is really only good at solving very specific problems, outside of which a normal computer would be better served.

    • by ceoyoyo ( 59147 )

      Except that, as per the article, somebody has shown a purely classical algorithm that reproduces the bimodal success distribution. The blog author in the article actually accepts that distribution as evidence (finally) that the D-Wave is actually doing something quantum, but then has to back off on it when that result comes up.

  • by JoshuaZ ( 1134087 ) on Wednesday May 22, 2013 @09:15AM (#43793613) Homepage

    Scott's blog post and the comment thread there are really worth reading. Entanglement isn't the only issue. A major part of this also is whether the process that the D-Wave machine is doing is anything that is even faster (either in practice or asymptotically) than a classical computer. Right now, the answer for the first is clearly no. Scott describes mildly optimized systems which have been shown to effectively outperform D-Wave at its own problem. The second question- of asymptotic performance is a little trickier but the answer looks like "probably not". It is also worth noting that the D-Wave machines do a very specific optimization problem, of unclear usefulness, and cannot be used at all for many of things that we think of as what one wants a quantum computer for, like Shor's algorithm http://en.wikipedia.org/wiki/Shor's_algorithm [wikipedia.org] to factor integers.

    In fairness to D-Wave though if one thinks of this as essentially a research machine, then not doing as well as conventional systems isn't that much of mark against it any more than very early cars being slower than horses. However, D-Wave is trying to sell these machines commercially. And Scott expresses worry that there's going to be a serious backlash against quantum computing as a whole when the the D-Wave hype bubble bursts. Apparently, D-Wave has now gotten about 100 million in funding, so at least, this is an extremely suboptimal allocation to resources when much more promising academic quantum computer research projects are getting much less money.

    • However, D-Wave is trying to sell these machines commercially. And Scott expresses worry that there's going to be a serious backlash against quantum computing as a whole when the the D-Wave hype bubble bursts.

      No one with the money to afford one of their machines is stupid enough to buy them as anything other than a research machine. Even if someone hadn't come up with a way to match their speed with a classical computer (which has already happened) a 10,000x speed increase on a very narrow problem set at a cost of more than 100,000x just buying the additional hardware (not to mention the cost in learning how to use this new and exotic machine) makes obvious that it is not yet an economical solution for anyone.

      • by Anonymous Coward on Wednesday May 22, 2013 @10:25AM (#43794397)

        I would argue that we should be open-minded at first and see what they can actually do. Maybe analog computers are in fact not as outdated as some people claim. Maybe we could build some sort of "analog FPGA" and do massively useful things with that. I still remember an HP computer graphics subsystem using analog computers !

        Surely digital computers have the advantage of simple control of temperature, aging and general error margin issues, but it comes at a massive cost in the number of transisitors to perform a certain function. less than ten transistors can perform an analog multiplication while you need tens of thousands if not hundreds of thousands of transistors to perform a floating point multiplication. Also, the analog multiplier will be operating at much higher speed (easily 10x). Again, if we could control temperature and aging-related issues and have high integration and programmability (FPGA-style), maybe we could do massively useful things at very low power levels or with massive parallelity. I do NOT think that analog computers are dead forever. It might be more of a cultural thing we currently don't use them much ("digital is always better", "digital is modern" and similar semi-truths.

        If you put one seasoned computer scientists and one seasoned electrical engineer in one room and task them to do what I described, if you give them massive funding (say 3 million dollars), I am sure they could come up with something massively useful. For example, digital circuits could periodically calibrate the analog circuits to compensate for all kinds of drift and aging. Software could automate the drudgery of manual circuit synthesis, it could model crosstalk and similar things.

        Well, maybe Analog Devices already has this kind of thing.....

        • Damn. All those mod points last week, and now when I for once run in to an under-moded AC post, of course I have non. I would love to read a comparison between the D-Wave computer and analog computers.

          • by ceoyoyo ( 59147 )

            It would presumably get it's ass kicked. The D-Wave solves an annealing optimization problem. The analog computer called "my lunch" solves those problems on a far larger scale every time I put it in the microwave and the cheese melts.

  • Proved the Market (Score:5, Interesting)

    by bill_mcgonigle ( 4333 ) * on Wednesday May 22, 2013 @09:21AM (#43793671) Homepage Journal

    Whether this thing turns out to be the real McCoy (dammit Jim, I'm a quantum annealer) or not, one thing D-Wave has done is proven that there are customers who will pay $10M to be on the cutting edge of quantum computing for a few years. This should help boost investment and entrepreneurship in other companies. Eventually, one of them will revolutionize everything.

    • The only thing they proved is there's a sucker born every minute. It has not helped the field of quantum computing. In fact it set it back because now people are going to distrust something that's called a quantum computer.
  • by FilmedInNoir ( 1392323 ) on Wednesday May 22, 2013 @09:43AM (#43793963)
    D-Wave is probably a scam, but I work in IT, almost the entire industry is built on scams.
    Scott Aaronson comes off as an egotistical man-child that is just angry he's not directly involved in all this.
    It's still love - love in this match.
  • Now that I've RTFA and through the commentary threads, as a dumb ignorant layperson I get why Scott Aaronsen is right to call out D-Wave. I also get the counter-argument that there needs to be some sort of hype in order to sustain interest in QC. And, the damn thing's got to work eventually. What I'm wondering though is this: Are we (as a society) making an error in trying to use QC to solve problems that are particular to classical computing?

    The reason I ask is that a while back on /. I was educated about the nature of Base-10 computing. Prior to this, I'd spent my entire life thinking that Base-10 WAS mathematics, and I'd had no reason to assume or even imagine that there could be any other type of mathematics than Base-10. Base-10 was the pinnacle of mathematics to me. Then I find out that Base-10 is probably the most efficient to date for our society, but that it is not the only way to count; and that Pi is only Pi because of Base-10. Which led me to look at mathematics in a whole new light. Similar with Quantum mechanics--the more I understand about Quantum Mechanics, the more I realize that I have to set aside everything I know about Newtonian physics, because trying to understand quantum physics from a newtonian perspective will always result in failure--while there is a bridge between the two, if I don't take that "bridge" into account then I'm metaphorically trying to judge apples based on my prior experience in dog shows.

    Given this, is it fair to hold QC to the same standards as Classical Computing, or should we be looking at entirely new applications of computing? And, is there anyone out there who's staring into the vast unknown and saying "What happens if we do THIS with a QC?"

    • by JoshuaZ ( 1134087 ) on Wednesday May 22, 2013 @10:39AM (#43794551) Homepage

      he reason I ask is that a while back on /. I was educated about the nature of Base-10 computing. Prior to this, I'd spent my entire life thinking that Base-10 WAS mathematics, and I'd had no reason to assume or even imagine that there could be any other type of mathematics than Base-10. Base-10 was the pinnacle of mathematics to me. Then I find out that Base-10 is probably the most efficient to date for our society, but that it is not the only way to count; and that Pi is only Pi because of Base-10.

      No. Pi will be the same regardless of base. The digits of Pi will be different if you write it in a different base, but this is simply a representation, not a change in what the number is. If you do calculations involving Pi in one base and do the same thing with another base and then convert the answer from one to the other you will get the same thing.

      Your general question is a good one. In fact, one of the major things people want to use quantum computers for is to do simulations of quantum systems, which they can do, but which are extremely inefficient (both in terms of time and memory) on a classical computer. So people are looking at problems which are practically not doable on a classical computer. At the same time though, we know that a quantum computer can be simulated on a classical computer with massive resource overhead (essentially exponential slowdown), so we know that anything you can do on a quantum computer you can do on a classical computer if one is patient enough.

      • by arth1 ( 260657 )

        Pi will be the same regardless of base.

        Most bases. Not base 1 (unary), base 0 or base -1.

        • Those aren't bases you can write numbers in in general. That's not that Pi is different, that's that those bases can't be used to represent all real numbers. Pi would still be the same.
          • by arth1 ( 260657 )

            Those aren't bases you can write numbers in in general. That's not that Pi is different, that's that those bases can't be used to represent all real numbers. Pi would still be the same.

            Unary is still very much in use. Nearly every child starts by counting fingers, and often we still count by making a mark for each thing we count.
            Yes, unary doesn't have a way of representing reals, which means that pi can only be represented if someone manages to square the circle. Otherwise, you're stuck with fractions and other representations based on whole numbers. The concept of an exact pi only exists outside the unary system.

            As for base 0, it most certainly can represent all real numbers. In fac

            • Regarding unary, you are still confusing the issue of whether one has a representation of the number with whether it exists. These are not the same thing.

              As for base 0, it most certainly can represent all real numbers. In fact, it cannot not do so. But there's no way to convert pi from base 0 to any other base.

              I don't follow. What do you mean?

              • by arth1 ( 260657 )

                Regarding unary, you are still confusing the issue of whether one has a representation of the number with whether it exists. These are not the same thing.

                Not if you claim it can be convered from one base to another. Then the base must have a way to represent it, and unary doesn't (unless, as said, someone manages to square the circle).

                I don't follow. What do you mean?

                In base 0, all numbers are equal, real or not. 0*N = 0 for every possible value of N. This makes base 0 rather useless for mathematics, of course, but it's a valid philosophical concept, a relative to atheistic solipsism.

                • No. Changing the base doesn't make the numbers equal. It just that in base 0 you can't represent any number other than 0. Don't confuse a representation with a number.
    • by arth1 ( 260657 )

      Then I find out that Base-10 is probably the most efficient to date for our society, but that it is not the only way to count; and that Pi is only Pi because of Base-10.

      No, pi is pi in all bases equal to or higher than two, provided you assume euclidian geometry.
      If you meant that pi is 3.14159265358979323846... only because of base 10, you're correct (again, as long as you ignore relativity and curved space).

      • As I am not a mathematician, I can safely assure you I was indeed ignoring relativity and curved space :o) Even though right now I'm traveling along a curved path in spacetime around an immense gravity well, the result of which is slowly propelling me through time faster than I would be if I were seated at the center of the Earth.
    • by ceoyoyo ( 59147 )

      "I was educated about the nature of Base-10 computing. Prior to this, I'd spent my entire life thinking that Base-10 WAS mathematics."

      Really? We had to do some base-10 computing in undergrad, of the BCD nature. It was a pain in the ass and everyone was glad to get back to good old base 2. And that was base-10 coded in base-2. I haven't heard of a computer using base-10 natively since the ones in the middle of the 20th century that wore skirts.

      • You're responding to someone who is definitely NOT a mathematician here. What I was trying to convey was that prior to that revelation, I had no concept of unary, base 0, base 1, base 2, etc. All I knew was that Mathematics worked because of arabic numbers . THAT'S IT. That's all I knew--just stop for a moment to grasp the depth of this ignorance. I didn't know why it worked, and I had no context in which to even imagine being able to ask "why" it worked. It just did--and that was the limit of my experience
        • by ceoyoyo ( 59147 )

          I don't think you properly understand the implications of changing bases, possibly because you say you learned about it from Slashdot. Mathematics doesn't care what base you use. When I took honours calculus we didn't really use numbers at all - everything was symbolic. Number systems are just convenient ways we use to represent values that are larger (or smaller) than the number of unique symbols we care to make use of. As someone else pointed out, pi is pi no matter what base you use. If you calculat

          • Dude, your entire response proves my point: I'm NOT a mathematician, I don't know these things, and I'm still trying to comprehend them. That's why I'm asking the question. Please don't get pedantic on me--my ignorance is genuine and not due to stupidity.
            • by ceoyoyo ( 59147 )

              If you're trying to understand, perhaps you should read the explanation without dismissing it as pedantic.

  • Of course it is quantum. The whole universe, including the reactions in the D-Wave are based on the laws of quantum mechanics. Regular computer chips have to take quantum effects into account too, although they try to defeat quantum effects rather than utilize them. Nevertheless, at a basic level the transistors work because of the laws of quantum mechanics.

    So how do you want to describe the D-Wave? Do you want to describe it using the laws of Quantum Mechanics? Or do you want to approximate it using a

  • by jones_supa ( 887896 ) on Wednesday May 22, 2013 @11:36AM (#43795063)
    They can build these things already? I thought that quantum computers would be only a theoretical idea for many coming years.
  • There are 3 possibilities: They built a quantum computer, they are completely faking it, or they invented a whole new branch of physics. Number 1 might get them a Nobel, number 2 is unlikely, and number 3 gets them at least one Nobel with them carried to the podium on the shoulders of the other winners.

"An idealist is one who, on noticing that a rose smells better than a cabbage, concludes that it will also make better soup." - H.L. Mencken

Working...