Forgot your password?
typodupeerror
Businesses AI

Nvidia Expects To Sell 'At Least' $1 Trillion In AI Chips By 2028 (techcrunch.com) 43

An anonymous reader quotes a report from TechCrunch: Nvidia CEO Jensen Huang threw out a lot of numbers -- mostly of the technical variety -- during his keynote Monday to kick off the company's annual GTC Conference in San Jose, California. But there was one financial figure that investors surely took notice of: his projection that there will be $1 trillion worth of orders for Nvidia's Blackwell and Vera Rubin chips, a monetary reflection of a booming AI business.

About an hour into his keynote, Huang noted that last year Nvidia saw about $500 billion in demand for its Blackwell and upcoming Rubin chips through 2026. "Now, I don't know if you guys feel the same way, but $500 billion is an enormous amount of revenue," he said. "Well, I'm here to tell you that right now where I stand -- a few short months after GTC DC, one year after last GTC -- right here where I stand, I see through 2027, at least $1 trillion."

This discussion has been archived. No new comments can be posted.

Nvidia Expects To Sell 'At Least' $1 Trillion In AI Chips By 2028

Comments Filter:
  • by Anonymous Coward

    To who? Consumers can't afford them, the prices of tech is going up. Businesses can't afford them, they can't justify additional costs without proof of return.

    Other businesses will buy them? And do what? Nobody is doing anything significantly profitable and novel with AI.

    • Re: (Score:3, Insightful)

      by Anonymous Coward

      "Nobody goes there anymore. It's too crowded." -Yogi Berra

    • by nedlohs ( 1335013 ) on Tuesday March 17, 2026 @02:56PM (#66046344)

      To the hyperscalers. Who then rent them to the AI companies that are losing hundreds of billions of dollars as fast as they can, and surely won't by unable to pay their bills.

    • by ArchieBunker ( 132337 ) on Tuesday March 17, 2026 @02:58PM (#66046352)

      The AI companies will pay with the money they don't have to put in the datacenters that haven't been built.

      • The AI companies will pay with the money they don't have to put in the datacenters that haven't been built.

        Yes, the AI customers don't have the money for all the data centers. Even the money-rich hyperscalars don't have the money and have to borrow. However, these hyperscalars are self-funding for the most part and borrowing is only for around 10-20% of their spending. They are being stretched, but a case could be made that they can "handle" the financial strain. Microsoft, Alphabet, Amazon, and Meta have a combined annual profit of over $350 billion, and their actual operating cash flow is over $500 billion

      • Where are the power plants for these?

    • by Junta ( 36770 )

      If the big AI companies can't afford them, it's fine, nVidia will give them money they can use to buy the GPUs.

      • by jezwel ( 2451108 )

        If the big AI companies can't afford them, it's fine, nVidia will give them money they can use to buy the GPUs.

        Aren't we already seeing this - nVidia provide investment capital to companies ($100B to OpenAI for instance) they can buy nVidia AIPUs?

        It's like McDonalds requiring franchisees to build McDonalds restaurants on land rented to them by McDonalds - one way or (and) another they'll get their ROI.

    • Probably datacenters? Although I daresay that those too will stop getting built pretty soon. They make sense for CDNs and cloud servers, but just tossing more AI into them is something that the public just won't be willing to keep underwriting
    • That would explain why they're now considering making chips for space
  • by Dan East ( 318230 ) on Tuesday March 17, 2026 @02:54PM (#66046340) Journal

    Serious question, why haven't they architected something better than GPUs for running inference? Surely something specifically designed for the task that could do it faster using less power? Something like Groq ASIC (that's just one I've heard of). Why aren't these the future and eclipsing the stop-gap that is GPUs because they already existed and were the best fit at the time?

    • Re:Why GPUs? (Score:5, Informative)

      by Junta ( 36770 ) on Tuesday March 17, 2026 @03:26PM (#66046392)

      The datacenter "GPUs" at this point have been specifically designed for the task.

      The B300 is mostly dedicated to FP4. The only use case for 4-bit floating point is AI. If you want VDI or non-AI use, you want something other than a B300.

      • Eh, I wouldn't say it's "mostly dedicated" to FP4. It still works fine with everything up to FP32. Only FP64 gets thrown under the bus, via emulation.

        • by Junta ( 36770 )

          Sure, it has INT32 and FP32, but the lion's share is devoted to being a FP4 beast.

          I haven't personally tried to touch a B300 FP64 wise, because I was told even the 'emulation' was a no-go as the emulation still banked on *some* real FP64 units to work, and B300 has zero. So B200 with FP64 emulation leveraging some of the real FP64 was what I was told was the best to hope for in FP64 from nVidia's line.

          • True, this [glennklockwood.com] does make it seem pretty terrible at 32 bits as well. I hadn't realized that "TF32" was really a 19-bit format, and that's the biggest data type that actually runs well.

      • Good info, but parent was asking about inference.

        The beastliness of the B300 in FP4 is for compute.
        It is, like all large model inference hardware, strictly limited by its memory bandwidth for inference.
        That's why the B200 and the B300 get nearly identical t/s inference (obviously not on ttft- since the B300 does indeed have double the compute performance for FP4) - because they've got the same HBM3e @ 8TB/s, and you just can't cache data that is read sequentially.
    • by evanh ( 627108 )

      I presume Google's TPU would be classed as dedicated. There are others, like the mind bogglingly large Cerebras WSE series, but most don't get a lot of interest.

    • Serious question, why haven't they architected something better than GPUs for running inference? Surely something specifically designed for the task that could do it faster using less power? Something like Groq ASIC (that's just one I've heard of). Why aren't these the future and eclipsing the stop-gap that is GPUs because they already existed and were the best fit at the time?

      The answer is that everyone is already doing exactly what you said. Groq is now essentially Nvidia, so even Nvidia is expanding their product portfolio. They offer GPUs (Blackwells and Vera Rubins), inference systems (Groq), CPUs (Vera), and networking (including InfiniBand and ethernet, where Nvidia now has higher data center networking revenue than Cisco). Currently about 70% of Nvidia data center revenue is from GPUs, and that percentage will drop when the Groq systems ramp up.

    • For inference- the answer is actually pretty simple.
      You are using models gigabytes to terabytes in size. It's an uncacheable workload.
      This means the actual bottleneck to inference, is the speed of the RAM.

      Your question would better be posed as: Why haven't the architected better RAM than X.... and the answer is, they are.
  • by sabbede ( 2678435 ) on Tuesday March 17, 2026 @02:58PM (#66046350)
    Will there be one of those, or has Nvidia decided that the customers they built everything upon aren't worth serving anymore?
    • by Junta ( 36770 )

      No, you are now supposed to just rent remote GPU time from nVidia instead.

    • Why can't the loyal customers do something for their loyal supplier? You paid money, and got a graphics card.

      • Yeah, but that was five years ago, and my 3070ti is getting a little long in the tooth. I might be able to get a new one later this year, but won't get a 5000 series because I know they're probably about to be replaced. But, Nvidia's release cadence for consumer GPUs has slowed since the chip shortage and never seemed to recover, so maybe not. It's a pickle. (Oh, I was supposed to get pickles...)

        Or maybe it just seems that way to me because I'm an impatient human.

        Oh BALLS! I just checked. 2nd half

    • Romance and business share a lot in common and you are like Magaldi in Evita, here's the scene on youtube sorry about the low res: https://youtu.be/4sa9oMej5B8?s... [youtu.be]
      • Let me stop you right there. I hate musicals. Yes, I am in love with Jensen Huang and wish to marry him, but I don't think either of us are actually gay (ref. my opinion on musicals), and I think we are both married to other people. But, what really breaks my heart is that store shelves aren't filled with reasonably priced RTX 6000's, won't be for over a year, and they will probably be even more expensive than the already overpriced ones.

        But if Jensen is reading this, hey, I'm open to it if you are!

      • Look, I tried. I did follow the link and saw things but was immediately hampered by my disinterest in Madonna and distaste for the Peronist dictatorship. Might as well be making musicals about Rachele Guidi or Carmen Polo.

        But that's my opinion, I won't tell you not to enjoy something you like. Have at it. Or don't. I don't get to tell you what to do.

    • Now would be the time for Intel to step into the market. They already have the Arc line.

  • How are they planning on selling these? Will it be to consumers or will they be investing 1 trillion into other companies on the basis said companies buy 1 trillion nvidia chips?
    • by zlives ( 2009072 )

      nah they are going to sell 3 for the trillion dollars
      1st is to fuckerburg for the new AR glasshole anal plug
      2nd is to bezos to create minime ai bots for his own self pleasure
      3rd is for elmo so he can say he is a leet gamer

      profit

  • They are just assuming that the data center frenzy will continue and that they will be able to push up prices because no one will be able to make enough chips to meet the inflated demand. At some point the boom must bust, but they are probably hoping it is far enough out to be the next guy's problem.
    • by jezwel ( 2451108 )
      Depends on whether consumers - including businesses - want locally running agentic AI on local hardware or centrally running in datacentres.

      Owned hardware means potential for locally run AI, which can be open source and therefore no subscription costs for use. I'm not just talking PCs creating stuff, or a calendar wall display summarising my schedule, but robots performing household tasks.

  • " last year Nvidia saw about $500 billion in demand for its Blackwell and upcoming Rubin chips through 2026"

    Very good revenue, and then this next generation already available that is "3.5x faster than the Blackwell architecture on model-training tasks and 5x faster on inference tasks". If you are building AI data centers this translates to very serious money. The Rubin chip architecture could easily be worth twice the price of Blackwell.

  • Did he do the pinky-to-corner-of-mouth thing when he said it?

    "One TRILLION dollars" /pinky

Everybody likes a kidder, but nobody lends him money. -- Arthur Miller

Working...