Jensen Huang, co-founder and chief executive officer of Nvidia Corp., during the Nvidia GPU Technology Conference (GTC) in San Jose, California, US, on Tuesday, March 19, 2024.
David Paul Morris | Bloomberg | Getty Images
Nvidia’s 27% rally in May pushed its market cap to $2.7 trillion, behind only Microsoft and Apple among the most-valuable public companies in the world. The chipmaker reported a tripling in year-over-year sales for the third straight quarter driven by soaring demand for its artificial intelligence processors.
Mizuho Securities estimates that Nvidia controls between 70% and 95% of the market for AI chips used for training and deploying models like OpenAI’s GPT. Underscoring Nvidia’s pricing power is a 78% gross margin, a stunningly high number for a hardware company that has to manufacture and ship physical products.
Rival chipmakers Intel and Advanced Micro Devices reported gross margins in the latest quarter of 41% and 47%, respectively.
Nvidia’s position in the AI chip market has been described as a moat by some experts. Its flagship AI graphics processing units (GPUs), such as the H100, coupled with the company’s CUDA software led to such a head start on the competition that switching to an alternative can seem almost unthinkable.
Still, Nvidia CEO Jensen Huang, whose net worth has swelled from $3 billion to about $90 billion in the past five years, has said he’s “worried and concerned” about his 31-year-old company losing its edge. He acknowledged at a conference late last year that there are many powerful competitors on the rise.
“I don’t think people are trying to put me out of business,” Huang said in November. “I probably know they’re trying to, so that’s different.”
Nvidia has committed to releasing a new AI chip architecture every year, rather than every other year as was the case historically, and to putting out new software that could more deeply entrench its chips in AI software.
But Nvidia’s GPU isn’t alone in being able to run the complex math that underpins generative AI. If less powerful chips can do the same work, Huang might be justifiably paranoid.
The transition from training AI models to what’s called inference — or deploying the models — could also give companies an opportunity to replace Nvidia’s GPUs, especially if they’re less expensive to buy and run. Nvidia’s flagship chip costs roughly $30,000 or more, giving customers plenty of incentive to seek alternatives.
“Nvidia would love to have 100% of it, but customers would not love for Nvidia to have 100% of it,” said Sid Sheth, co-founder of aspiring rival D-Matrix. “It’s just too big of an opportunity. It would be too unhealthy if any one company took all of it.”
Founded in 2019, D-Matrix plans to release a semiconductor card for servers later this year that aims to reduce the cost and latency of running AI models. The company raised $110 million in September.
In addition to D-Matrix, companies ranging from multinational corporations to nascent startups are fighting for a slice of the AI chip market that could reach $400 billion in annual sales in the next five years, according to market analysts and AMD. Nvidia has generated about $80 billion in revenue over the past four quarters, and Bank of America estimates the company sold $34.5 billion in AI chips last year.
Many companies taking on Nvidia’s GPUs are betting that a different architecture or certain trade-offs could produce a better chip for particular tasks. Device makers are also developing technology that could end up doing a lot of the computing for AI that’s currently taking place in large GPU-based clusters in the cloud.
“Nobody can deny that today Nvidia is the hardware you want to train and run AI models,” Fernando Vidal, co-founder of 3Fourteen Research, told CNBC. “But there’s been incremental progress in leveling the playing field, from hyperscalers working on their own chips, to even little startups, designing their own silicon.”
AMD CEO Lisa Su wants investors to believe there’s plenty of room for many successful companies in the space.
“The key is that there are a lot of options there,” Su told reporters in December, when her company launched its most recent AI chip. “I think we’re going to see a situation where there’s not only one solution, there will be multiple solutions.”
Other big chipmakers
Lisa Su displays an AMD Instinct MI300 chip as she delivers a keynote address at CES 2023 in Las Vegas, Nevada, on Jan. 4, 2023.
David Becker | Getty Images
AMD makes GPUs for gaming and, like Nvidia, is adapting them for AI inside of data centers. Its flagship chip is the Instinct MI300X. Microsoft has already bought AMD processors, offering access to them through its Azure cloud.
At launch, Su highlighted the chip’s excellence at inference, as opposed to competing with Nvidia for training. Last week, Microsoft said it was using AMD Instinct GPUs to serve its Copilot models. Morgan Stanley analysts took the news as a sign that AMD’s AI chip sales could surpass $4 billion this year, the company’s public target.
Intel, which was surpassed by Nvidia last year in terms of revenue, is also trying to establish a presence in AI. The company recently announced the third version of its AI accelerator, Gaudi 3. This time Intel compared it directly to the competition, describing it as a more cost-effective alternative and better than Nvidia’s H100 in terms of running inference, while faster at training models.
Bank of America analysts estimated recently that Intel will have less than 1% of the AI chip market this year. Intel says it has a $2 billion order of backlogs for the chip.
The main roadblock to broader adoption may be software. AMD and Intel are both participating in a big industry group called the UXL foundation, which includes Google, that’s working to create free alternatives to Nvidia’s CUDA for controlling hardware for AI applications.
Nvidia’s top customers
One potential challenge for Nvidia is that it’s competing against some of its biggest customers. Cloud providers including Google, Microsoft and Amazon are all building processors for internal use. The Big Tech three, plus Oracle, make up over 40% of Nvidia’s revenue.
Amazon introduced its own AI-oriented chips in 2018, under the Inferentia brand name. Inferentia is now on its second version. In 2021, Amazon Web Services debuted Tranium targeted to training. Customers can’t buy the chips but they can rent systems through AWS, which markets the chips as more cost efficient than Nvidia’s.
Google is perhaps the cloud provider most committed to its own silicon. The company has been using what it calls Tensor Processing Units (TPUs) since 2015 to train and deploy AI models. In May, Google announced the sixth version of its chip, Trillium, which the company said was used to develop its models, including Gemini and Imagen.
Google also uses Nvidia chips and offers them through its cloud.
Microsoft isn’t as far along. The company said last year that it was building its own AI accelerator and processor, called Maia and Cobalt.
Meta isn’t a cloud provider, but the company needs massive amounts of computing power to run its software and website and to serve ads. While the Facebook parent company is buying billions of dollars worth of Nvidia processors, it said in April that some of its homegrown chips were already in data centers and enabled “greater efficiency” compared to GPUs.
JPMorgan analysts estimated in May that the market for building custom chips for big cloud providers could be worth as much as $30 billion, with potential growth of 20% per year.
Startups
Cerebras’ WSE-3 chip is one example of new silicon from upstarts designed to run and train artificial intelligence.
Cerebras Systems
Venture capitalists see opportunities for emerging companies to jump into the game. They invested $6 billion in AI semiconductor companies in 2023, up slightly from $5.7 billion a year earlier, according to data from PitchBook.
It’s a tough area for startups as semiconductors are expensive to design, develop and manufacture. But there are opportunities for differentiation.
For Cerebras Systems, an AI chipmaker in Silicon Valley, the focus is on basic operations and bottlenecks for AI, versus the more general purpose nature of a GPU. The company was founded in 2015 and was valued at $4 billion during its most recent fundraising, according to Bloomberg.
The Cerebras chip, WSE-2, puts GPU capabilities as well as central processing and additional memory into a single device, which is better for training large models, said CEO Andrew Feldman.
“We use a giant chip, they use a lot of little chips,” Feldman said. “They’ve got challenges of moving data around, we don’t.”
Feldman said his company, which counts Mayo Clinic, GlaxoSmithKline, and the U.S. Military as clients, is winning business for its supercomputing systems even going up against Nvidia.
“There’s ample competition and I think that’s healthy for the ecosystem,” Feldman said.
Sheth from D-Matrix said his company plans to release a card with its chiplet later this year that will allow for more computation in memory, as opposed to on a chip like a GPU. D-Matrix’s product can be slotted into an AI server along existing GPUs, but it takes work off of Nvidia chips, and helps to lower the cost of generative AI.
Customers “are very receptive and very incentivized to enable a new solution to come to market,” Sheth said.
Apple and Qualcomm
Apple iPhone 15 series devices are displayed for sale at The Grove Apple retail store on release day in Los Angeles, California, on September 22, 2023.
Patrick T. Fallon | Afp | Getty Images
The biggest threat to Nvidia’s data center business may be a change in where processing happens.
Developers are increasingly betting that AI work will move from server farms to the laptops, PCs and phones we own.
Big models like the ones developed by OpenAI require massive clusters of powerful GPUs for inference, but companies like Apple and Microsoft are developing “small models” that require less power and data and can run on a battery-powered device. They may not be as skilled as the latest version of ChatGPT, but there are other applications they perform, such as summarizing text or visual search.
Apple and Qualcomm are updating their chips to run AI more efficiently, adding specialized sections for AI models called neural processors, which can have privacy and speed advantages.
Qualcomm recently announced a PC chip that will allow laptops to run Microsoft AI services on the device. The company has also invested in a number of chipmakers making lower-power processors to run AI algorithms outside of a smartphone or laptop.
Apple has been marketing its latest laptops and tablets as optimized for AI because of the neural engine on its chips. At its upcoming developer conference, Apple is planning to show off a slew of new AI features, likely running on the company’s iPhone-powering silicon.
Almost 600 people have signed an open letter to leaders at venture firm Sequoia Capital after one of its partners, Shaun Maguire, posted what the group described as a “deliberate, inflammatory attack” against the Muslim Democratic mayoral candidate in New York City.
Maguire, a vocal supporter of President Donald Trump, posted on X over the weekend that Zohran Mamdani, who won the Democratic primary last month, “comes from a culture that lies about everything” and is out to advance “his Islamist agenda.”
The post had 5.3 million views as of Monday afternoon. Maguire, whose investments include Elon Musk’s SpaceX and X as well as artificial intelligence startup Safe Superintelligence, also published a video on X explaining the remark.
Those signing the letter are asking Sequoia to condemn Maguire’s comments and apologize to Mamdani and Muslim founders. They also want the firm to authorize an independent investigation of Maguire’s behavior in the past two years and post “a zero-tolerance policy on hate speech and religious bigotry.”
They are asking the firm for a public response by July 14, or “we will proceed with broader public disclosure, media outreach and mobilizing our networks to ensure accountability,” the letter says.
Sequoia declined to comment. Maguire didn’t respond to a request for comment, but wrote in a post about the letter on Wednesday that, “You can try everything you want to silence me, but it will just embolden me.”
Among the signees are Mudassir Sheikha, CEO of ride-hailing service Careem, and Amr Awadallah, CEO of AI startup Vectara. Also on the list is Abubakar Abid, who works in machine learning Hugging Face, which is backed by Sequoia, and Ahmed Sabbah, CEO of Telda, a financial technology startup that Sequoia first invested in four years ago.
At least three founders of startups that have gone through startup accelerator program Y Combinator added their names to the letter.
Sequoia as a firm is no stranger to politics. Doug Leone, who led the firm until 2022 and remains a partner, is a longtime Republican donor, who supported Trump in the 2024 election. Following Trump’s victory in November, Leone posted on X, “To all Trump voters: you no longer have to hide in the shadows…..you’re the majority!!”
By contrast, Leone’s predecessor, Mike Moritz, is a Democratic megadonor, who criticized Trump and, in August, slammed his colleagues in the tech industry for lining up behind the Republican nominee. In a Financial Times opinion piece, Moritz wrote Trump’s tech supporters were “making a big mistake.”
“I doubt whether any of them would want him as part of an investment syndicate that they organised,” wrote Moritz, who stepped down from Sequoia in 2023, over a decade after giving up a management role at the firm. “Why then do they dismiss his recent criminal conviction as nothing more than a politically inspired witch-hunt over a simple book-keeping error?”
Neither Leone nor Moritz returned messages seeking comment.
Roelof Botha, Sequoia’s current lead partner, has taken a more neutral stance. Botha said at an event last July that Sequoia as a partnership doesn’t “take a political point of view,” adding that he’s “not a registered member of either party.” Boelof said he’s “proud of the fact that we’ve enabled many of our partners to express their respected individual views along the way, and given them that freedom.”
Maguire has long been open with his political views. He said on X last year that he had “just donated $300k to President Trump.”
Mamdani, a self-described democratic socialist, has gained the ire of many people in tech and in the business community more broadly since defeating former New York Gov. Andrew Cuomo in the June primary.
Samsung signage during the Nvidia GPU Technology Conference (GTC) in San Jose, California, US, on Thursday, March 20, 2025.
David Paul Morris | Bloomberg | Getty Images
South Korea’s Samsung Electronics on Tuesday forecast a 56% fall in profits for the second as the company struggles to capture demand from artificial intelligence chip leader Nvidia.
The memory chip and smartphone maker said in its guidance that operating profit for the quarter ending June was projected to be around 4.6 trillion won, down from 10.44 trillion Korean won year over year.
The figure is a deeper plunge compared to smart estimates from LSEG, which are weighted toward forecasts from analysts who are more consistently accurate.
According to the smart estimates, Samsung was expected to post an operating profit of 6.26 trillion won ($4.57 billion)for the quarter. Meanwhile, Samsung projected its revenue to hit 74 trillion won, falling short of LSEG smart estimates of 75.55 trillion won.
Samsung is a leading player in the global smartphone market and is also one of the world’s largest makers of memory chips, which are utilized in devices such as laptops and servers.
However, the company has been falling behind competitors like SK Hynix and Micron in high-bandwidth memory chips — an advanced type of memory that is being deployed in AI chips.
“The disappointing earnings are due to ongoing operating losses in the foundry business, while the upside in high-margin HBM business remains muted this quarter,” MS Hwang, Research Director at Counterpoint Research, said about the earnings guidance.
SK Hynix, the leader in HBM, has secured a position as Nvidia’s key supplier. While Samsung has reportedly been working to get the latest version of its HBM chips certified by Nvidia, a report from a local outlet suggests these plans have been pushed back to at least September.
The company did not respond to a request for comment on the status of its deals with Nvidia.
Ray Wang, Research Director of Semiconductors, Supply Chain and Emerging Technology at Futurum Group told CNBC that it is clear that Samsung has yet to pass Nvidia’s qualification for its most advanced HBM.
“Given that Nvidia accounts for roughly 70% of global HBM demand, the delay meaningfully caps near-term upside,” Wang said. He noted that while Samsung has secured some HBM supply for AI processors from AMD, this win is unlikely to contribute to second-quarter results due to the timing of production ramps.
Reuters reported in September that Samsung had instructed its subsidiaries worldwide to cut 30% of staff in some divisions, citing sources familiar with the matter.
A Waymo autonomous self-driving Jaguar electric vehicle sits parked at an EVgo charging station in Los Angeles, California, on May 15, 2024.
Patrick T. Fallon | AFP | Getty Images
Waymo said it will begin testing in Philadelphia, with a limited fleet of vehicles and human safety drivers behind the wheel.
“This city is a National Treasure,” Waymo wrote in a post on X on Monday. “It’s a city of love, where eagles fly with a gritty spirit and cheese that spreads and cheese that steaks. Our road trip continues to Philly next.”
The Alphabet-owned company confirmed to CNBC that it will be testing in Pennsylvania’s largest city through the fall, adding that the initial fleet of cars will be manually driven through the more complex parts of Philadelphia, including downtown and on freeways.
“Folks will see our vehicles driving at all hours throughout various neighborhoods, from North Central to Eastwick, and from University City to as far east as the Delaware River,” a Waymo spokesperson said.
With its so-called road trips, Waymo seeks to collect mapping data and evaluate how its autonomous technology, Waymo Driver, performs in new environments, handling traffic patterns and local infrastructure. Road trips are often used a way for the company to gauge whether it can potentially offer a paid ride share service in a particular location.
The expanded testing, which will go through the fall, comes as Waymo aims for a broader rollout. Last month, the company announced plans to drive vehicles manually in New York for testing, marking the first step toward potentially cracking the largest U.S. city. Waymo applied for a permit with the New York City Department of Transportation to operate autonomously with a trained specialist behind the wheel in Manhattan. State law currently doesn’t allow for such driverless operations.
Waymo One provides more than 250,000 paid trips each week across Phoenix, San Francisco, Los Angeles, and Austin, Texas, and is preparing to bring fully autonomous rides to Atlanta, Miami, and Washington, D.C., in 2026.
Alphabet has been under pressure to monetize artificial intelligence products as it bolsters spending on infrastructure. Alphabet’s “Other Bets” segment, which includes Waymo, brought in revenue of $1.65 billion in 2024, up from $1.53 billion in 2023. However, the segment lost $4.44 billion last year, compared to a loss of $4.09 billion the previous year.