Connect with us

Published

on

Larry Ellison and Monica Seles and Bill Gates (back row) watch Carlos Alcaraz of Spain play against Alexander Zverev of Germany in their Quarterfinal match during the BNP Paribas Open in Indian Wells, California, on March 14, 2024.

Clive Brunskill | Getty Images

It’s been a good year for Larry Ellison.

Oracle’s co-founder has gained roughly $75 billion in paper wealth as the software company he started in 1979 enjoyed its biggest stock rally since 1999 and the dot-com boom.

While the S&P 500 index has gained 27% in 2024, Oracle shares have shot up 63%, lifting Ellison’s net worth to more than $217 billion, according to Forbes, behind only Tesla CEO Elon Musk and Amazon founder Jeff Bezos among the world’s richest people.

At 80, Ellison is a senior citizen in the tech industry, where his fellow billionaire founders are generally decades younger. Meta CEO Mark Zuckerberg, whose net worth has also ballooned past $200 billion, is half his age.

But Ellison has found the fountain of youth both personally and professionally. After being divorced several times, Ellison was reported this month to be involved with a 33-year-old woman. And at a meeting with analysts in Las Vegas in September, Ellison was as engaged as ever, mentioning offhand that the night before, he and his son were having dinner with his good friend Musk, who’s advising President-elect Donald Trump (then the Republican nominee) while running Tesla and his other ventures.

His big financial boon has come from Oracle, which has maneuvered its way into the artificial intelligence craze with its cloud infrastructure technology and has made its databases more accessible.

ChatGPT creator OpenAI said in June that it will use Oracle’s cloud infrastructure. Earlier this month, Oracle said it had also picked up business from Meta.

Startups, which often opt for market leader Amazon Web Services when picking a cloud, have been engaging Oracle as well. Last year, video generation startup Genmo set up a system to train an AI model with Nvidia graphics processing units, or GPUs, in Oracle’s cloud, CEO Paras Jain said. Genmo now relies on the Oracle cloud to produce videos based on the prompts that users type in on its website.

“Oracle produced a different product than what you can get elsewhere with GPU computing,” Jain said. The company offers “bare metal” computers that can sometimes yield better performance than architectures that employ server virtualization, he said.

In its latest earnings report earlier this month, Oracle came up short of analysts’ estimates and issued a forecast that was also weaker than Wall Street was expecting. The stock had its worst day of 2024, falling almost 7% and eating into the year’s gains.

Oracle has the best infrastructure for hosting GPUs anywhere, says Citizens JMP's Patrick Walravens

Still, Ellison was bullish for the future.

“Oracle Cloud Infrastructure trains several of the world’s most important generative AI models because we are faster and less expensive than other clouds,” Ellison said in the earnings release.

For the current fiscal year, which ends in May, Oracle is expected to record revenue growth of about 10%, which would mark its second-strongest year of expansion since 2011.

Jain said that when Genmo has challenges, he communicates with Oracle sales executives and engineers through a Slack channel. The collaboration has resulted in better reliability and performance, he said. Jain said Oracle worked with Genmo to ensure that developers could launch the startup’s Mochi open-source video generator on Oracle’s cloud hardware with a single click.

“Oracle was also more price-competitive than these large hyperscalers,” Jain said.

‘That’s going to be so easy’

Three months before its December earnings report, at the analyst event in Las Vegas, Oracle had given a rosy outlook for the next three years. Executive Vice President Doug Kehring declared that the company would produce more than $66 billion in revenue in the 2026 fiscal year, and over $104 billion in fiscal 2029. The numbers suggested acceleration, with a compound annual growth rate of over 16%, compared with 9% in the latest quarter.

After Kehring and CEO Safra Catz spoke, it was Ellison’s turn. The company’s chairman, technology chief and top shareholder strutted onto the stage in a black sweater and jeans, waved to the analysts, licked his lips and sat down. For the next 74 minutes, he answered questions from seven analysts.

“Did — did he say $104 billion?” Ellison said, referring to Kehring’s projection. Some in the crowd giggled. “That’s going to be so easy. It is kind of crazy.”

Oracle’s revenue in fiscal 2023 was just shy of $50 billion.

The new target impressed Eric Lynch, managing director of Scharf Investments, which held $167 million in Oracle shares at the end of September.

“For a company doing single digits for a decade or so, that’s unbelievable,” Lynch told CNBC in an interview.

Oracle co-founder and Chairman Larry Ellison delivers a keynote address during the Oracle OpenWorld on October 22, 2018 in San Francisco, California. 

Justin Sullivan | Getty Images

Oracle is still far behind in cloud infrastructure. In 2023, Amazon controlled 39% share of market, followed by Microsoft at 23% and Google at 8.2%, according to industry researcher Gartner. That left Oracle with 1.4%.

But in database software, Oracle remains a stalwart. Gartner estimated that the company had 17% market share in database management systems in 2023.

Ellison’s challenge is to find opportunities for expansion.

Last year, he visited Microsoft headquarters in Redmond, Washington, for the first time to announce a partnership that would enable organizations to use Oracle’s database through Microsoft’s Azure cloud. Microsoft even installed Oracle hardware in its data centers.

In June, Oracle rolled out a similar announcement with Google. Then, in September, Oracle finally partnered with Amazon, introducing its database on AWS.

Oracle and Amazon had exchanged barbs for years. AWS introduced a database called Aurora in 2014, and Amazon worked hard to move itself off Oracle. Following a CNBC report on the effort, Ellison expressed doubt about Amazon’s ability to reach its goal. But the project succeeded.

In 2019, Amazon published a blog post titled, “Migration Complete – Amazon’s Consumer Business Just Turned off its Final Oracle Database.”

Friendlier vibe

Ellison looked back on the history between the two companies at the analyst meeting in September.

“I got kind of got cute commenting about Amazon uses Oracle, doesn’t use AWS, blah, blah,” he said. “And that hurt some people’s feelings. I probably shouldn’t have said it.”

He said a friend at a major New York bank had asked him to make sure the Oracle database works on AWS.

“I said, ‘Great. It makes sense to me,'” Ellison said.

The multi-cloud strategy should deliver gains in database market share, said analyst Siti Panigrahi of Mizuho, which has the equivalent of a buy rating on Oracle shares. Cloud deals related to AI will also help Oracle deliver on its promise for faster revenue growth, he said.

“Oracle right now has an end-to-end stack for enterprises to build their AI strategy,” said Panigrahi, who worked on applications at Oracle in the 2000s.

So far, Oracle has been mainly cutting high-value AI deals with the likes of OpenAI and Musk’s X.ai. Of Oracle’s $97 billion in remaining performance obligations, or revenue that hasn’t yet been recognized, 40% or 50% of it is tied to renting out GPUs, Panigrahi said.

Oracle didn’t respond to a request for comment.

Panigrahi predicts that a wider swath of enterprises will begin adopting AI, which will be a boon to Oracle given its hundreds of thousands of big customers.

There’s also promise in Oracle Health, the segment that came out of the company’s $28.2 billion acquisition of electronic health record software vendor Cerner in 2022.

Yoshiki Hayashi, Marc Benioff and Larry Ellison attend the Transformative Medicine of USC: Rebels with a Cause Gala in Santa Monica, California, on Oct. 24, 2019.

Joshua Blanchard | Getty Images

Unlike rival Epic, Oracle Health lost U.S. market share in 2023, according to estimates from KLAS Research. But Ellison’s connection to Musk, who is set to co-lead Trump’s Department of Government Efficiency, might benefit Oracle Health “if there is a bigger push towards modernizing existing healthcare systems,” analysts at Evercore said in a note last week. They recommend buying the stock.

For now, Oracle is busy using AI to rewrite Cerner’s entire code base, Ellison said at the analyst event.

“This is another pillar for growth,” he said. “I think you haven’t quite seen it yet.”

Hours earlier, Ellison had put in a call to Marc Benioff, co-founder and CEO of Salesforce. Benioff knows Ellison as well as anyone, having worked for him for 13 years before starting the cloud software company that’s now a big competitor.

“It was awesome,” Benioff said in a wide-ranging interview the next day, regarding his chat with Ellison.

Benioff spoke about his former boss’s latest run of fortune.

“Larry really deeply wants this,” Benioff said. “This is very important to him, that he is building a great company, what he believes is one of the most important companies in the world, and also, wealth is very important to him.”

Don’t miss these insights from CNBC PRO

Oracle and C3.ai both came into their earnings reports hot, says Jim Cramer

Continue Reading

Technology

Amazon cut thousands of engineers in its record layoffs, despite saying it needs to innovate faster

Published

on

By

Amazon cut thousands of engineers in its record layoffs, despite saying it needs to innovate faster

The Amazon Puget Sound Headquarters is pictured on Oct. 28, 2025 in Seattle, Washington.

Stephen Brashear | Getty Images

Amazon‘s 14,000-plus layoffs announced last month touched almost every piece of the company’s sprawling business, from cloud computing and devices to advertising, retail and grocery stores. But one job category bore the brunt of cuts more than others: engineers.

Documents filed in New York, California, New Jersey and Amazon’s home state of Washington showed that nearly 40% of the more than 4,700 job cuts in those states were engineering roles. The data was reported by Amazon in Worker Adjustment and Retraining Notification, or WARN, filings to state agencies.

The figures represent a segment of the total layoffs announced in October. Not all data was immediately available because of differences in state WARN reporting requirements.

In announcing the steepest round of cuts in its 31-year history, Amazon joined a growing roster of tech companies that have slashed jobs this year even as cash piles have mounted and profits soared. In total, there have been almost 113,000 job cuts at 231 tech companies, according to Layoffs.fyi, continuing a trend that began in 2022 as businesses readjusted to life after the Covid pandemic.

Amazon CEO Andy Jassy has been on a multiyear mission to transform the company’s corporate culture into one that operates like what he calls “the world’s largest startup.” He’s looked to make Amazon leaner and less bureaucratic by urging staffers to do more with less and cutting organizational bloat.

Amazon is expected to carry out further job reductions in January, CNBC previously reported.

Andy Jassy, chief executive officer of Amazon.com Inc., speaks during an unveiling event in New York, US, on Wednesday, Feb. 26, 2025.

Michael Nagle | Bloomberg | Getty Images

The company said it’s also shifting resources to invest more in artificial intelligence. The technology is already poised to reshape Amazon’s white-collar workforce, with Jassy predicting in June that its corporate head count will shrink in the coming years alongside efficiency gains from AI.

Human resources chief Beth Galetti, in her memo announcing the layoffs, focused on the importance of innovating, which the company will now have to do with fewer people, specifically engineers.

“This generation of AI is the most transformative technology we’ve seen since the Internet, and it’s enabling companies to innovate much faster than ever before,” Galetti wrote. “We’re convinced that we need to be organized more leanly, with fewer layers and more ownership, to move as quickly as possible for our customers and business.”

Amazon said in a statement that AI is not the driver behind the vast majority of the job cuts, and that the bigger goal was to reduce bureaucracy and emphasize speed.

Jassy said on Amazon’s earnings call last month that the cuts were in response to a “culture” issue inside the company, spurred in part by an extended hiring spree that left it with “a lot more layers” and slower decision-making.

The layoffs impacted a mix of software engineer levels, but SDE II roles, or mid-level employees, were disproportionately affected, the WARN filings show.

The AI boom is making software development jobs harder to come by as companies adopt coding assistants or so-called vibe coding platforms from vendors like Cursor, OpenAI and Cognition. Amazon has released its own competitor called Kiro.

Read more CNBC tech news

‘Significant role reductions’

Amazon spends billions on AI arms race as it guts corporate ranks

Game designers, artists and producers made up more than a quarter of the total cuts in Irvine, and they were roughly 11% of staffers laid off at Amazon’s San Diego offices, according to filings.

The company also told staffers it’s halting much of its work on big-budget, or triple A, game development, specifically around massively multiplayer online, or MMO, games, Boom wrote. Amazon has released MMOs including Crucible and New World. It was also developing an MMO based on “Lord of the Rings.”

Beyond its gaming division, Amazon also significantly cut back its visual search and shopping teams, according to multiple employee posts on LinkedIn. The unit is responsible for products like Amazon Lens and Lens Live, AI shopping tools that enable users to find products via their camera in real time or images saved to their device. The company rolled out Lens Live in September.

The team was primarily based in Palo Alto, California, and Amazon’s WARN filings indicate that software engineers, applied scientists and quality assurance engineers were heavily impacted across its offices there.

Amazon’s online ad business, one of its biggest profit centers, was downsized as well. More than 140 ad sales and marketing roles were eliminated across Amazon’s New York offices, accounting for about 20% of the roughly 760 positions cut, according to state documents viewed by CNBC.

WATCH: Box joining AWS marketplace in new partnership

AI's impact on reshaping the workforce

Continue Reading

Technology

The market’s surprising reversal, Gap’s viral ad, AI regulation and more in Morning Squawk

Published

on

By

The market's surprising reversal, Gap's viral ad, AI regulation and more in Morning Squawk

Dado Ruvic | Reuters

This is CNBC’s Morning Squawk newsletter. Subscribe here to receive future editions in your inbox.

Here are five key things investors need to know to start the trading day:

1. Hero to zero

Stock investors didn’t end up getting the post-Nvidia earnings market bounce they hoped for. After opening yesterday’s trading session higher, stocks took a dramatic midday tumble, once again casting doubt on the artificial intelligence trade.

Here’s what to know:

  • Nvidia shares gave up their 5% post-earnings gain, ending the session down more than 3% despite the chipmaker’s blockbuster quarterly results and guidance. The AI darling’s stock is on track to finish the week down 5%.
  • The Dow Jones Industrial Average swung more than 1,100 between its session highs and lows. All three major averages closed solidly in the red, with the tech-heavy Nasdaq Composite ending the day down 2.15%.
  • Meanwhile, the CBOE Volatility Index — better known as Wall Street’s fear gauge — ended the session at a level not seen since April.
  • Bitcoin fell to lows going back to April, further illustrating the shift away from risk assets.
  • Before stocks’ midday reversal, Bridgewater founder Ray Dalio told CNBC that “we are in that territory of a bubble,” but that you don’t need to sell stocks because of it.
  • The three major indexes are all on track to end the week in the red.
  • Follow live markets updates here.

2. Prediction market

A ‘Now Hiring’ sign is posted outside of a business on Oct. 3, 2025 in Miami, Florida.

Joe Raedle | Getty Images

The belated September jobs report was finally released yesterday, and the headline number was much hotter than economists expected with an increase of 119,000 jobs. On the other hand, the unemployment rate ticked up to 4.4%, its highest level since 2021.

The chance of a rate cut at the Federal Reserve’s next meeting remained low after the report, according to the CME FedWatch Tool. But the odds flipped this morning after New York Fed President John Williams said he sees “room for a further adjustment” in interest rates, reviving hopes of a December cut.

Get Morning Squawk directly in your inbox

3. Better than yours

Merchandise on display in a Gap store on November 21, 2024 in Miami Beach, Florida. 

Joe Raedle | Getty Images

Gap‘s “Milkshake” ad brought all the shoppers to the store. The retailer’s viral “Better in Denim” campaign with girl group Katseye helped drive comparable sales up 5% in its third quarter, beating analyst expectations.

The Old Navy and Banana Republic parent also surpassed Wall Street’s estimates on both the top and bottom lines, sending shares rising 4.5% in overnight trading. Athleta was the notable outlier, with the athleisure brand’s sales falling 11%.

Gap’s report comes at the end of a busy week for retail earnings. As CNBC’s Melissa Repko reports, one key theme of this quarter’s results has been that value-oriented retailers are winning favor with shoppers across income brackets.

4. AI in D.C.

U.S. President Donald Trump speaks in the Oval Office at the White House on Oct. 6, 2025 in Washington, DC.

Anna Moneymaker | Getty Images

The White House is putting together an executive order that would thwart states’ individual AI laws. A draft obtained by CNBC shows the order would focus on staging legal challenges and blocking federal funding for states to ensure their compliance.

The draft would work to the advantage of many AI industry leaders who have pushed back on a state-by-state approach to the technology’s regulation. A White House official told CNBC that any discussion around the draft is speculation until an official announcement.

Click here to read the full draft.

5. Flight fight

Courtesy: Archer Aviation

Joby Aviation is taking air taxi competitor Archer Aviation to court. In a lawsuit filed Wednesday, Joby accused Archer of using information stolen by a former employee to “one-up” a deal with a real estate developer.

Joby alleges that George Kivork, its former U.S. state and local policy lead, took files and information before jumping to the competitor in an act of “corporate espionage.” Archer called the case “baseless litigation” and said it’s “entirely without merit.”

The Daily Dividend

Here are our recommendations for stories to circle back to this weekend:

CNBC’s Liz Napolitano, Tasmin Lockwood, Melissa Repko, Jeff Cox, Sarah Min, Emily Wilkins, Mary Catherine Wellons and Samantha Subin contributed to this report. Josephine Rozzelle edited this edition.

Continue Reading

Technology

Nvidia is king in AI chips, but Google and Amazon want to catch up by making their own

Published

on

By

Nvidia is king in AI chips, but Google and Amazon want to catch up by making their own

Breaking down AI chips, from Nvidia GPUs to ASICs by Google and Amazon

Nvidia outperformed all expectations, reporting soaring profits Wednesday thanks to its graphics processing units that excel at AI workloads. But more categories of AI chips are gaining ground.

Custom ASICs, or application-specific integrated circuits, are now being designed by all the major hyperscalers, from Google‘s TPU to Amazon‘s Trainium and OpenAI’s plans with Broadcom. These chips are smaller, cheaper, accessible and could reduce these companies’ reliance on Nvidia GPUs. Daniel Newman of the Futurum Group told CNBC that he sees custom ASICs “growing even faster than the GPU market over the next few years.”

Besides GPUs and ASICs, there are also field-programmable gate arrays, which can be reconfigured with software after they’re made for use in all sorts of applications, like signal processing, networking and AI. There’s also an entire group of AI chips that power AI on devices rather than in the cloud. Qualcomm, Apple and others have championed those on-device AI chips.

CNBC talked to experts and insiders at the Big Tech companies to break down the crowded space and the various kinds of AI chips out there.

GPUs for general compute

Once used primarily for gaming, GPUs made Nvidia the world’s most valuable public company after their use shifted toward AI workloads. Nvidia shipped some 6 million current-generation Blackwell GPUs over the past year.

Nvidia senior director of AI infrastructure Dion Harris shows CNBC’s Katie Tarasov how 72 Blackwell GPUs work together as one in a GB200 NVL72 rack-scale server system for AI at Nvidia headquarters in Santa Clara, California, on November 12, 2025.

Marc Ganley

The shift from gaming to AI started around 2012, when Nvidia’s GPUs were used by researchers to build AlexNet, what many consider to be modern AI’s big bang moment. AlexNet was a tool that was entered into a prominent image recognition contest. Whereas others in the contest used central processing units for their applications, AlexNet reliance on GPUs provided incredible accuracy and obliterated its competition.

AlexNet’s creators discovered that the same parallel processing that helps GPUs render lifelike graphics was also great for training neural networks, in which a computer learns from data rather than relying on a programmer’s code. AlexNet showcased the potential of GPUs.

Today, GPUs are often paired with CPUs and sold in server rack systems to be placed in data centers, where they run AI workloads in the cloud. CPUs have a small number of powerful cores running sequential general-purpose tasks, while GPUs have thousands of smaller cores more narrowly focused on parallel math like matrix multiplication.

Because GPUs can perform many operations simultaneously, they’re ideal for the two main phases of AI computation: training and inference. Training teaches the AI model to learn from patterns in large amounts of data, while inference uses the AI to make decisions based on new information.

GPUs are the general-purpose workhorses of Nvidia and its top competitor, Advanced Micro Devices. Software is a major differentiator between the two GPU leaders. While Nvidia GPUs are tightly optimized around CUDA, Nvidia’s proprietary software platform, AMD GPUs use a largely open-source software ecosystem.

AMD and Nvidia sell their GPUs to cloud providers like Amazon, Microsoft, Google, Oracle and CoreWeave. Those companies then rent the GPUs to AI companies by the hour or minute. Anthropic’s $30 billion deal with Nvidia and Microsoft, for example, includes 1 gigawatt of compute capacity on Nvidia GPUs. AMD has also recently landed big commitments from OpenAI and Oracle.

Nvidia also sells directly to AI companies, like a recent deal to sell at least 4 million GPUs to OpenAI, and to foreign governments, including South Korea, Saudi Arabia and the U.K.

The chipmaker told CNBC that it charges around $3 million for one of its server racks with 72 Blackwell GPUs acting as one, and ships about 1,000 each week. 

Dion Harris, Nvidia’s senior director of AI infrastructure, told CNBC he couldn’t have imagined this much demand when he joined Nvidia over eight years ago.

“When we were talking to people about building a system that had eight GPUs, they thought that was overkill,” he said.

ASICs for custom cloud AI

Training on GPUs has been key in the early boom days of large language models, but inference is becoming more crucial as the models mature. Inference can happen on less powerful chips that are programmed for more specific tasks. That’s where ASICs come in.

While a GPU is like a Swiss Army Knife able to do many kinds of parallel math for different AI workloads, an ASIC is like a single-purpose tool. It’s very efficient and fast, but hard-wired to do the exact math for one type of job.

Google released its 7th generation TPU, Ironwood, in November 2025, a decade after making its first custom ASIC for AI in 2015.

Google

“You can’t change them once they’re already carved into silicon, and so there’s a trade off in terms of flexibility,” said Chris Miller, author of “Chip War.”

Nvidia’s GPUs are flexible enough for adoption by many AI companies, but they cost up to $40,000 and can be hard to get. Still, startups rely on GPUs because designing a custom ASIC has an even higher up-front cost, starting at tens of millions of dollars, according to Miller.

For the biggest cloud providers who can afford them, analysts say custom ASICs pay off in the long-run.

“They want to have a little bit more control over the workloads that they build,” Newsom said. “At the same time, they’re going to continue to work very closely with Nvidia, with AMD, because they also need the capacity. The demand is so insatiable.”

Google was the first Big Tech company to make a custom ASIC for AI acceleration, coining the term Tensor Processing Unit when its first ASIC came out in 2015. Google said it considered making a TPU as far back as 2006, but the situation became “urgent” in 2013 as it realized AI was going to double its number of data centers. In 2017, the TPU also contributed to Google’s invention of the Transformer, the architecture powering almost all modern AI.

A decade after its first TPU, Google released its seventh generation TPU in November. Anthropic announced it will train its LLM Claude on up to 1 million TPUs. Some people think TPUs are technically on par or superior to Nvidia’s GPUs, Miller said.

“Traditionally, Google has only used them for in-house purposes,” Miller said. “There’s a lot of speculation that in the longer run, Google might open up access to TPUs more broadly.”

Amazon Web Services was the next cloud provider to design its own AI chips, after acquiring Israeli chip startup Annapurna labs in 2015. AWS announced Inferentia in 2018, and it launched Trainium in 2022. AWS is expected to announce Trainium’s third generation as soon December.

Ron Diamant, Trainium’s head architect, told CNBC that Amazon’s ASIC has 30% to 40% better price performance compared to other hardware vendors in AWS.

“Over time, we’ve seen that Trainium chips can serve both inference and training workloads quite well,” Diamant said.

CNBC’s Katie Tarasov holds Amazon Web Services’ Trainium 2 AI chip that fill its new AI data center in New Carlisle, Indiana, on October 8, 2025.

Erin Black

In October, CNBC went to Indiana for the first on-camera tour of Amazon’s biggest AI data center, where Anthropic is training its models on half a million Trainium2 chips. AWS fills its other data centers with Nvidia GPUs to meet the demand from AI customers like OpenAI.

Building ASICs isn’t easy. This is why companies turn to chip designers Broadcom and Marvell. They “provide the IP and the know-how and the networking” to help their clients build their ASICs, Miller said.

“So you’ve seen Broadcom in particular be one of the biggest beneficiaries of the AI boom,” Miller said.

Broadcom helped build Google’s TPUs and Meta‘s Training and Inference Accelerator launched in 2023, and has a new deal to help OpenAI build its own custom ASICs starting in 2026.

Microsoft is also getting into the ASIC game, telling CNBC that its in-house Maia 100 chips are currently deployed in its data centers in the eastern U.S. Others include Qualcomm with the A1200, Intel with its Gaudi AI accelerators and Tesla with its AI5 chip. There’s also a slew of start-ups going all in on custom AI chips, including Cerebras, which makes huge full-wafer AI chips, and Groq, with inference-focused language processing units.

In China, Huawei, ByteDance, and Alibaba are making custom ASICs, although export controls on the most advanced equipment and AI chips pose a challenge.

Edge AI with NPUs and FPGAs

The final big category of AI chips are those made to run on devices, rather than in the cloud. These chips are typically built into a device’s main System on a Chip, SoC. Edge AI chips, as they’re called, enable devices to have AI capabilities while helping them save battery life and space for other components.

“You’ll be able to do that right on your phone with very low latency, so you don’t have to have communication all the way back to a data center,” said Saif Khan, former White House AI and semiconductor policy advisor. “And you can preserve privacy of your data on your phone.”

Neural processing units are a major type of edge AI chip. Qualcomm, Intel and AMD are making NPUs that enable AI capabilities in personal computers.

Although Apple doesn’t use the term NPU, the in-house M-series chips inside its MacBooks include a dedicated neural engine. Apple also built neural accelerators into the latest iPhone A-series chips.

“It is efficient for us. It is responsive. We know that we are much more in control over the experience,” Tim Millet, Apple platform architecture vice president, told CNBC in an exclusive September interview

The latest Android phones also have NPUs built into their primary Qualcomm Snapdragon chips, and Samsung has its own NPU on its Galaxy phones, too. NPUs by companies like NXP and Nvidia power AI embedded in cars, robots, cameras, smart home devices and more.

“Most of the dollars are going towards the data center, but over time that’s going to change because we’ll have AI deployed in our phones and our cars and wearables, all sorts of other applications to a much greater degree than today,” Miller said.

Then there’s field-programmable gate arrays, or FPGAs, which can be reconfigured with software after they’re made. Although far more flexible than NPUs or ASICs, FPGAs have lower raw performance and lower energy efficiency for AI workloads.

AMD became the largest FPGA maker after acquiring Xilinx for $49 billion in 2022, with Intel in second thanks to its $16.7 billion purchase of Altera in 2015.

These players designing AI chips rely on a single company to manufacture them all: Taiwan Semiconductor Manufacturing Company.

TSMC has a giant new chip fabrication plant in Arizona, where Apple has committed to moving some chip production. In October, Nvidia CEO Jensen Huang said Blackwell GPUs were in “full production” in Arizona, too. 

Although the AI chip space is crowded, dethroning Nvidia won’t come easily.

“They have that position because they’ve earned it and they’ve spent the years building it,” Newman said. “They’ve won that developer ecosystem.”

Watch the video to see a breakdown of how all the AI chips work: https://www.cnbc.com/video/2025/11/21/nvidia-gpus-google-tpus-aws-trainium-comparing-the-top-ai-chips.html

Continue Reading

Trending