Forgot your password?
typodupeerror
AI Hardware

Furiosa's Energy-Efficient 'NPU' AI Chips Start Mass Production This Month, Challenging Nvidia (msn.com) 25

The Wall Street Journal profiles "the startup that is now one of a handful of chip makers nipping at the heels of Nvidia." Furiosa's AI chip is dubbed "RNGD" — short for renegade — and slated to start mass production this month. Valued at nearly $700 million based on its most recent fundraising, Furiosa has attracted interest from big tech firms. Last year, Meta Platforms attempted to acquire it, though the startup declined the offer. OpenAI used a Furiosa chip for a recent demonstration in Seoul. LG's AI research unit is testing the chip and said it offered "excellent real-world performance." Furiosa said it is engaged in talks with potential customers.

Nvidia's graphic processing units, or GPUs, dominated the initial push to train AI models. But companies like Furiosa are betting that for the next stage — referred to as "inference," or using AI models after they're trained — their specialty chips can be competitive. Furiosa makes chips called neural processing units, or NPUs, which are a rising class of chips designed specifically to handle the type of computing calculations underpinning AI and use less energy than GPUs. [Founder/CEO June] Paik said Furiosa's chips can provide similar performance as Nvidia's advanced GPUs with less electricity usage. That would drive down the total costs of deploying AI. The tech world, Paik says, shouldn't be so reliant on one chip maker for AI computing. "A market dominated by a single player — that's not a healthy ecosystem, is it?" Paik said...

In 2024, at Stanford's prestigious Hot Chips conference, Paik debuted Furiosa's RNGD chip as a solution for what he called "sustainable AI computing" in a keynote speech. Paik presented data showing how the chip could run the then-latest version of Meta's Llama large language model with more than twice the power efficiency of Nvidia's high-end chips. Furiosa's booth was swarmed with engineers from big tech firms, including Google, Meta and Amazon.com, wanting to see a live demo of the chip. "It was a moment where we felt we could really move forward with our chip with confidence," Paik said.

This discussion has been archived. No new comments can be posted.

Furiosa's Energy-Efficient 'NPU' AI Chips Start Mass Production This Month, Challenging Nvidia

Comments Filter:
  • acronym (Score:3, Funny)

    by lastman71 ( 1314797 ) on Sunday January 04, 2026 @11:45AM (#65901113)

    RNGD is for Random Number Generator Daemon?

  • by will4 ( 7250692 ) on Sunday January 04, 2026 @12:05PM (#65901139)

    Read the article and did not find any mention other than a "uses less electricity" mention.

    The news should include a comparison including power usage per chip and server rack, AC cooling requirement power usage for a workload for this chip and a NVIDIA system.

    The news editor could help with AI quality control by: "In this news article, remove all mentions of childhood, outdoor scenery, clothes, fashion trends, physical descriptions of people, emotional appeals, softening language, and vignettes."

    Removing all of those would help to show that the article is 98% filler material (CEO personal profile + office decor + vignette) with little information.

  • Human beings are more ecologically sound than AI, and unlike AI, they have a right to exist.
    • Hopefully, someone will collect metrics on Youtube videos produce by AI, categorize them into "entertainment including brainrot" and everything else and then estimate the electricity used for the ai produced entertainment and scale it in terms of how many life saving medical operations could be performed in poor countries

  • by Tomahawk ( 1343 ) on Sunday January 04, 2026 @01:19PM (#65901243) Homepage

    I might finally be able to build a decent gaming PC?

    • RAM and storage are still going to be an extremely short supply for the next 4 or 5 years minimum.
    • Only marginable decent.
      But for old versions of Pac Man, it might be okay.

    • by Hentes ( 2461350 )

      This is actually bad for gamers. The bottleneck is how many chips TSMC can make, new designs aren't going to change that. But at least with GPUs you will be able to buy them on the cheap once the bubble bursts, while inference chips will end up in a landfill. Now some of these chips are just generic systolic arrays doing matrix multiplication which could have uses in a desktop, but you would need to get drivers for it somehow and also no software is going to support them.

    • by tlhIngan ( 30335 )

      I might finally be able to build a decent gaming PC?

      You can if you disregard future proofing for now. Older platforms capable of running DDR4 memory are coming back into vogue because DDR4 memory prices, while still inflated, are still roughly half the price of an equivalent amount of DDR5.

      This has meant things like AMD AM4, and Intel 13th and 14th gen CPUs are back which can be paired with a decent midrange GPU.

      The other meta if that PCs with 16GB of RAM would become common again, with 32GB being "high end

  • I checked at the web site, and like all the other AI chip designers, actual fabrication is by TSMC at 5nm and 4nm. My understanding is that TSMC's fab complex in Arizona can do 5nm parts, and will be able to do 4nm relatively soon. Also that they are at Arizona's capacity for those nodes doing work for Apple and AMD. So FuriosaAI's parts will be made in Taiwan.

    The US AI stock bubble is hideously dependent on a few fabs in Taiwan.
  • CUDA locks people into using NVIDIA hardware, and I doubt that'll change. Yes, the big guys can spend the engineering effort to rewrite software, but that'll give NVIDIA time to release their own inference optimized chips. I'd bet they're already working on some.

    • But do NPUs lock you into doing AI? After the AI bubble bursts, your CUDA/Nvidia hardware can be used for other things. While I prefer more open solutions such as OpenCL on AMD, the same OpenCL code will also run on Nvidia hardware.

      I don't know about the details of NPUs, but I've got the impression that efficient AI uses very low precision arithmetic, such as 8-bit and even lower. So while it could be possible to use NPUs for general vector/matrix math, the low precision would be a deal breaker for a lot

Feel disillusioned? I've got some great new illusions, right here!

Working...