Connect with us

Published

on

Breaking down AI chips, from Nvidia GPUs to ASICs by Google and Amazon

Nvidia outperformed all expectations, reporting soaring profits Wednesday thanks to its graphics processing units that excel at AI workloads. But more categories of AI chips are gaining ground.

Custom ASICs, or application-specific integrated circuits, are now being designed by all the major hyperscalers, from Google‘s TPU to Amazon‘s Trainium and OpenAI’s plans with Broadcom. These chips are smaller, cheaper, accessible and could reduce these companies’ reliance on Nvidia GPUs. Daniel Newman of the Futurum Group told CNBC that he sees custom ASICs “growing even faster than the GPU market over the next few years.”

Besides GPUs and ASICs, there are also field-programmable gate arrays, which can be reconfigured with software after they’re made for use in all sorts of applications, like signal processing, networking and AI. There’s also an entire group of AI chips that power AI on devices rather than in the cloud. Qualcomm, Apple and others have championed those on-device AI chips.

CNBC talked to experts and insiders at the Big Tech companies to break down the crowded space and the various kinds of AI chips out there.

GPUs for general compute

Once used primarily for gaming, GPUs made Nvidia the world’s most valuable public company after their use shifted toward AI workloads. Nvidia shipped some 6 million current-generation Blackwell GPUs over the past year.

Nvidia senior director of AI infrastructure Dion Harris shows CNBC’s Katie Tarasov how 72 Blackwell GPUs work together as one in a GB200 NVL72 rack-scale server system for AI at Nvidia headquarters in Santa Clara, California, on November 12, 2025.

Marc Ganley

The shift from gaming to AI started around 2012, when Nvidia’s GPUs were used by researchers to build AlexNet, what many consider to be modern AI’s big bang moment. AlexNet was a tool that was entered into a prominent image recognition contest. Whereas others in the contest used central processing units for their applications, AlexNet reliance on GPUs provided incredible accuracy and obliterated its competition.

AlexNet’s creators discovered that the same parallel processing that helps GPUs render lifelike graphics was also great for training neural networks, in which a computer learns from data rather than relying on a programmer’s code. AlexNet showcased the potential of GPUs.

Today, GPUs are often paired with CPUs and sold in server rack systems to be placed in data centers, where they run AI workloads in the cloud. CPUs have a small number of powerful cores running sequential general-purpose tasks, while GPUs have thousands of smaller cores more narrowly focused on parallel math like matrix multiplication.

Because GPUs can perform many operations simultaneously, they’re ideal for the two main phases of AI computation: training and inference. Training teaches the AI model to learn from patterns in large amounts of data, while inference uses the AI to make decisions based on new information.

GPUs are the general-purpose workhorses of Nvidia and its top competitor, Advanced Micro Devices. Software is a major differentiator between the two GPU leaders. While Nvidia GPUs are tightly optimized around CUDA, Nvidia’s proprietary software platform, AMD GPUs use a largely open-source software ecosystem.

AMD and Nvidia sell their GPUs to cloud providers like Amazon, Microsoft, Google, Oracle and CoreWeave. Those companies then rent the GPUs to AI companies by the hour or minute. Anthropic’s $30 billion deal with Nvidia and Microsoft, for example, includes 1 gigawatt of compute capacity on Nvidia GPUs. AMD has also recently landed big commitments from OpenAI and Oracle.

Nvidia also sells directly to AI companies, like a recent deal to sell at least 4 million GPUs to OpenAI, and to foreign governments, including South Korea, Saudi Arabia and the U.K.

The chipmaker told CNBC that it charges around $3 million for one of its server racks with 72 Blackwell GPUs acting as one, and ships about 1,000 each week. 

Dion Harris, Nvidia’s senior director of AI infrastructure, told CNBC he couldn’t have imagined this much demand when he joined Nvidia over eight years ago.

“When we were talking to people about building a system that had eight GPUs, they thought that was overkill,” he said.

ASICs for custom cloud AI

Training on GPUs has been key in the early boom days of large language models, but inference is becoming more crucial as the models mature. Inference can happen on less powerful chips that are programmed for more specific tasks. That’s where ASICs come in.

While a GPU is like a Swiss Army Knife able to do many kinds of parallel math for different AI workloads, an ASIC is like a single-purpose tool. It’s very efficient and fast, but hard-wired to do the exact math for one type of job.

Google released its 7th generation TPU, Ironwood, in November 2025, a decade after making its first custom ASIC for AI in 2015.

Google

“You can’t change them once they’re already carved into silicon, and so there’s a trade off in terms of flexibility,” said Chris Miller, author of “Chip War.”

Nvidia’s GPUs are flexible enough for adoption by many AI companies, but they cost up to $40,000 and can be hard to get. Still, startups rely on GPUs because designing a custom ASIC has an even higher up-front cost, starting at tens of millions of dollars, according to Miller.

For the biggest cloud providers who can afford them, analysts say custom ASICs pay off in the long-run.

“They want to have a little bit more control over the workloads that they build,” Newsom said. “At the same time, they’re going to continue to work very closely with Nvidia, with AMD, because they also need the capacity. The demand is so insatiable.”

Google was the first Big Tech company to make a custom ASIC for AI acceleration, coining the term Tensor Processing Unit when its first ASIC came out in 2015. Google said it considered making a TPU as far back as 2006, but the situation became “urgent” in 2013 as it realized AI was going to double its number of data centers. In 2017, the TPU also contributed to Google’s invention of the Transformer, the architecture powering almost all modern AI.

A decade after its first TPU, Google released its seventh generation TPU in November. Anthropic announced it will train its LLM Claude on up to 1 million TPUs. Some people think TPUs are technically on par or superior to Nvidia’s GPUs, Miller said.

“Traditionally, Google has only used them for in-house purposes,” Miller said. “There’s a lot of speculation that in the longer run, Google might open up access to TPUs more broadly.”

Amazon Web Services was the next cloud provider to design its own AI chips, after acquiring Israeli chip startup Annapurna labs in 2015. AWS announced Inferentia in 2018, and it launched Trainium in 2022. AWS is expected to announce Trainium’s third generation as soon December.

Ron Diamant, Trainium’s head architect, told CNBC that Amazon’s ASIC has 30% to 40% better price performance compared to other hardware vendors in AWS.

“Over time, we’ve seen that Trainium chips can serve both inference and training workloads quite well,” Diamant said.

CNBC’s Katie Tarasov holds Amazon Web Services’ Trainium 2 AI chip that fill its new AI data center in New Carlisle, Indiana, on October 8, 2025.

Erin Black

In October, CNBC went to Indiana for the first on-camera tour of Amazon’s biggest AI data center, where Anthropic is training its models on half a million Trainium2 chips. AWS fills its other data centers with Nvidia GPUs to meet the demand from AI customers like OpenAI.

Building ASICs isn’t easy. This is why companies turn to chip designers Broadcom and Marvell. They “provide the IP and the know-how and the networking” to help their clients build their ASICs, Miller said.

“So you’ve seen Broadcom in particular be one of the biggest beneficiaries of the AI boom,” Miller said.

Broadcom helped build Google’s TPUs and Meta‘s Training and Inference Accelerator launched in 2023, and has a new deal to help OpenAI build its own custom ASICs starting in 2026.

Microsoft is also getting into the ASIC game, telling CNBC that its in-house Maia 100 chips are currently deployed in its data centers in the eastern U.S. Others include Qualcomm with the A1200, Intel with its Gaudi AI accelerators and Tesla with its AI5 chip. There’s also a slew of start-ups going all in on custom AI chips, including Cerebras, which makes huge full-wafer AI chips, and Groq, with inference-focused language processing units.

In China, Huawei, ByteDance, and Alibaba are making custom ASICs, although export controls on the most advanced equipment and AI chips pose a challenge.

Edge AI with NPUs and FPGAs

The final big category of AI chips are those made to run on devices, rather than in the cloud. These chips are typically built into a device’s main System on a Chip, SoC. Edge AI chips, as they’re called, enable devices to have AI capabilities while helping them save battery life and space for other components.

“You’ll be able to do that right on your phone with very low latency, so you don’t have to have communication all the way back to a data center,” said Saif Khan, former White House AI and semiconductor policy advisor. “And you can preserve privacy of your data on your phone.”

Neural processing units are a major type of edge AI chip. Qualcomm, Intel and AMD are making NPUs that enable AI capabilities in personal computers.

Although Apple doesn’t use the term NPU, the in-house M-series chips inside its MacBooks include a dedicated neural engine. Apple also built neural accelerators into the latest iPhone A-series chips.

“It is efficient for us. It is responsive. We know that we are much more in control over the experience,” Tim Millet, Apple platform architecture vice president, told CNBC in an exclusive September interview

The latest Android phones also have NPUs built into their primary Qualcomm Snapdragon chips, and Samsung has its own NPU on its Galaxy phones, too. NPUs by companies like NXP and Nvidia power AI embedded in cars, robots, cameras, smart home devices and more.

“Most of the dollars are going towards the data center, but over time that’s going to change because we’ll have AI deployed in our phones and our cars and wearables, all sorts of other applications to a much greater degree than today,” Miller said.

Then there’s field-programmable gate arrays, or FPGAs, which can be reconfigured with software after they’re made. Although far more flexible than NPUs or ASICs, FPGAs have lower raw performance and lower energy efficiency for AI workloads.

AMD became the largest FPGA maker after acquiring Xilinx for $49 billion in 2022, with Intel in second thanks to its $16.7 billion purchase of Altera in 2015.

These players designing AI chips rely on a single company to manufacture them all: Taiwan Semiconductor Manufacturing Company.

TSMC has a giant new chip fabrication plant in Arizona, where Apple has committed to moving some chip production. In October, Nvidia CEO Jensen Huang said Blackwell GPUs were in “full production” in Arizona, too. 

Although the AI chip space is crowded, dethroning Nvidia won’t come easily.

“They have that position because they’ve earned it and they’ve spent the years building it,” Newman said. “They’ve won that developer ecosystem.”

Watch the video to see a breakdown of how all the AI chips work: https://www.cnbc.com/video/2025/11/21/nvidia-gpus-google-tpus-aws-trainium-comparing-the-top-ai-chips.html

Continue Reading

Technology

CNBC Daily Open: Debt worries continue to weigh on AI-related stocks

Published

on

By

CNBC Daily Open: Debt worries continue to weigh on AI-related stocks

Traders work on the floor at the New York Stock Exchange in New York City, U.S., Dec. 15, 2025.

Brendan McDermid | Reuters

U.S. stocks of late have been shaky as investors turn away from artificial intelligence shares, especially those related to AI infrastructure, such as Oracle, Broadcom and CoreWeave.

The worry is that those companies are running into high levels of debt to finance their multibillion-dollar deals.

Oracle, for instance, said Wednesday it would need to raise capital expenditure by an additional $15 billion for its current fiscal year and increase its lease commitments for data centers. The company is turning to debt to finance all that.

The stock lost 2.7% on Monday, while shares of CoreWeave, its fellow player in the AI data center trade dropped around 8%. Broadcom also retreated over concerns over margin compression, sliding about 5.6%.

That said, the broader market was not affected too adversely as investors continued rotating into sectors such as consumer discretionary and industrials. The S&P 500 slipped 0.16%, the Dow Jones Industrial Average ticked down just 0.09% and the Nasdaq Composite, comprising more tech firms, fell 0.59%.

The broader market performance suggests that the fears are mostly contained within the AI infrastructure space.

“It definitely requires the ROI [return on investment] to be there to keep funding this AI investment,” Matt Witheiler, head of late-stage growth at Wellington Management, told CNBC’s “Money Movers” on Monday. “From what we’ve seen so far that ROI is there.”

Witheiler said the bullish side of the story is that, “every single AI company on the planet is saying if you give me more compute I can make more revenue.”

The ready availability of clients, according to that argument, means those companies that provide the compute — Oracle and CoreWeave — just need to make sure their finances are in order.

— CNBC’s Ari Levy contributed to this report.

What you need to know today

U.S. stocks edged down Monday. All major indexes slid as AI-related stocks continued to weigh down markets. Europe’s regional Stoxx 600 climbed 0.74%. The continent’s defense stocks fell as Ukraine offered to give up on joining NATO.

Tesla testing driverless Robotaxis in Austin, Texas. “Testing is underway with no occupants in the car,” CEO Elon Musk wrote in a post on his social network X over the weekend. Shares of Tesla rose 3.6% on Monday to close at their highest this year.

U.S. collects $200 billion in tariffs. The country’s Customs and Border Protection agency said Monday that the tally comprises only new tariffs, including “reciprocal” and “fentanyl” levies, imposed by U.S. President Trump in his second term.

Ukraine-Russia peace deal is nearly complete. That’s according to U.S. officials, who held talks with Ukraine President Volodymyr Zelenskyy beginning Sunday. Ukraine has offered to give up its NATO bid, while Russia is open to Ukraine joining the EU, officials said.

[PRO] Wall Street’s favorite stocks for 2026. These S&P 500 stocks have a consensus buy rating and an upside to average price target of at least 35%, based on CNBC Pro’s screening of data from LSEG.

And finally…

Customers walk in the parking lot outside a Costco store on December 02, 2025 in Chicago, Illinois.

Scott Olson | Getty Images

Continue Reading

Technology

Merriam-Webster declares ‘slop’ its word of the year in nod to growth of AI

Published

on

By

Merriam-Webster declares 'slop' its word of the year in nod to growth of AI

The logos of Google Gemini, ChatGPT, Microsoft Copilot, Claude by Anthropic, Perplexity, and Bing apps are displayed on the screen of a smartphone in Reno, United States, on November 21, 2024.

Jaque Silva | Nurphoto | Getty Images

Merriam-Webster declared “slop” its 2025 word of the year on Monday, a sign of growing wariness around artificial intelligence.

Slop is now defined as “digital content of low quality that is produced usually in quantity by means of artificial intelligence,” according to Merriam-Webster’s dictionary. The word has previously been used primarily to connote a “product of little value” or “food waste fed to animals”

Mainstream social networks saw a flood of AI-generated content, including what 404 Media described as a “video of a bizarre creature turning into a spider, turning into a nightmare giraffe inside of a busy mall,” that the publication reported had been viewed more than 362 million times on Meta apps. 

In September, Meta launched Vibes, a separate feed for AI-generated videos. Days later, OpenAI released its Sora app. Those services, along with TikTok, YouTube and others, are increasingly rife with AI slop, which can often generate revenue with enough engagement.

Spotify said in September that it had to remove over 75 million AI-generated, “spammy tracks” from its service, and roll out formal policies to protect artists from AI impersonation and deception. The streaming company faced widespread criticism after The Velvet Sundown racked up 1 million monthly listeners on without initially making it clear they produced their songs with generative AI. The artist later clarified on its bio page that it’s a “synthetic music project.”

According to CNBC’s latest All-America Economic Survey, published Dec. 15, fewer respondents have been using AI platforms, such as ChatGPT, Microsoft Copilot and Google Gemini, in the last two to three months compared to the summer months.

Just 48% of those surveyed said they had used AI platforms recently, down from 53% in August.

WATCH: OpenAI’s Sora 2 sparks AI ‘slop’ backlash

OpenAI's Sora 2 sparks AI 'slop' backlash

Continue Reading

Technology

PayPal applies to form bank that can offer small business loans and savings accounts

Published

on

By

PayPal applies to form bank that can offer small business loans and savings accounts

PayPal CEO Alex Chriss speaks at the Global Fintech Fest in Mumbai, India, on Oct. 7, 2025.

Indranil Aditya | Nurphoto | Getty Images

PayPal said Monday that it has applied for approval to form PayPal Bank, which would be able to offer loans to small businesses.

“Establishing PayPal Bank will strengthen our business and improve our efficiency, enabling us to better support small business growth and economic opportunities across the U.S.,” PayPal CEO Alex Chriss said in a statement.

The U.S. Federal Deposit Insurance Corporation will review an application proposing the establishment of PayPal Bank, along with Utah’s Department of Financial Institutions, PayPal said.

The company, which owns popular payment app Venmo, hopes to also offer interest-bearing savings accounts to its customers, the statement said. PayPal already makes credit lines available to consumers and has been trying to expand its roster of banking-like services as it competes with a growing number of fintech companies that are aiming to take business from traditional brick-and-mortar banks.

Shares of PayPal rose 1.5% in extended trading following the announcement.

In October, PayPal said quarterly revenue increased 7% year over year to $8.42 billion, more than analysts had expected. But in 2025 the stock has slumped about 29%, while the S&P 500 index has gained almost 16% in the same period.

WATCH: E-commerce consumption could bump 20% because of agentic AI, says Mizuho’s Dan Dolev

E-commerce consumption could bump 20% because of agentic AI, says Mizuho's Dan Dolev

Continue Reading

Trending