Nvidia CEO Jensen Huang speaks during a press conference at The MGM during CES 2018 in Las Vegas on January 7, 2018.
Mandel Ngan | AFP | Getty Images
Software that can write passages of text or draw pictures that look like a human created them has kicked off a gold rush in the technology industry.
Companies like Microsoft and Google are fighting to integrate cutting-edge AI into their search engines, as billion-dollar competitors such as OpenAI and Stable Diffusion race ahead and release their software to the public.
Powering many of these applications is a roughly $10,000 chip that’s become one of the most critical tools in the artificial intelligence industry: The Nvidia A100.
The A100 has become the “workhorse” for artificial intelligence professionals at the moment, said Nathan Benaich, an investor who publishes a newsletter and report covering the AI industry, including a partial list of supercomputers using A100s. Nvidia takes 95% of the market for graphics processors that can be used for machine learning, according to New Street Research.
The A100 is ideally suited for the kind of machine learning models that power tools like ChatGPT, Bing AI, or Stable Diffusion. It’s able to perform many simple calculations simultaneously, which is important for training and using neural network models.
The technology behind the A100 was initially used to render sophisticated 3D graphics in games. It’s often called a graphics processor, or GPU, but these days Nvidia’s A100 is configured and targeted at machine learning tasks and runs in data centers, not inside glowing gaming PCs.
Big companies or startups working on software like chatbots and image generators require hundreds or thousands of Nvidia’s chips, and either purchase them on their own or secure access to the computers from a cloud provider.
Hundreds of GPUsare required to train artificial intelligence models, like large language models. The chips need to be powerful enough to crunch terabytes of data quickly to recognize patterns. After that, GPUs like the A100 are also needed for “inference,” or using the model to generate text, make predictions, or identify objects inside photos.
This means that AI companies need access to a lot of A100s. Some entrepreneurs in the space even see the number of A100s they have access to as a sign of progress.
“A year ago we had 32 A100s,” Stability AI CEO Emad Mostaque wrote on Twitter in January. “Dream big and stack moar GPUs kids. Brrr.” Stability AI is the company that helped develop Stable Diffusion, an image generator that drew attention last fall, and reportedly has a valuation of over $1 billion.
Now, Stability AI has access to over 5,400 A100 GPUs, according to one estimate from the State of AI report, which charts and tracks which companies and universities have the largest collection of A100 GPUs — although it doesn’t include cloud providers, which don’t publish their numbers publicly.
Nvidia’s riding the A.I. train
Nvidia stands to benefit from the AI hype cycle. During Wednesday’s fiscal fourth-quarter earnings report, although overall sales declined 21%, investors pushed the stock up about 14% on Thursday, mainly because the company’s AI chip business — reported as data centers — rose by 11% to more than $3.6 billion in sales during the quarter, showing continued growth.
Nvidia shares are up 65% so far in 2023, outpacing the S&P 500 and other semiconductor stocks alike.
Nvidia CEO Jensen Huang couldn’t stop talking about AI on a call with analysts on Wednesday, suggesting that the recent boom in artificial intelligence is at the center of the company’s strategy.
“The activity around the AI infrastructure that we built, and the activity around inferencing using Hopper and Ampere to influence large language models has just gone through the roof in the last 60 days,” Huang said. “There’s no question that whatever our views are of this year as we enter the year has been fairly dramatically changed as a result of the last 60, 90 days.”
Ampere is Nvidia’s code name for the A100 generation of chips. Hopper is the code name for the new generation, including H100, which recently started shipping.
More computers needed
Nvidia A100 processor
Nvidia
Compared to other kinds of software, like serving a webpage, which uses processing power occasionally in bursts for microseconds, machine learning tasks can take up the whole computer’s processing power, sometimes for hours or days.
This means companies that find themselves with a hit AI product often need to acquire more GPUs to handle peak periods or improve their models.
These GPUs aren’t cheap. In addition to a single A100 on a card that can be slotted into an existing server, many data centers use a system that includes eight A100 GPUs working together.
This system, Nvidia’s DGX A100, has a suggested price of nearly $200,000, although it comes with the chips needed. On Wednesday, Nvidia said it would sell cloud access to DGX systems directly, which will likely reduce the entry cost for tinkerers and researchers.
It’s easy to see how the cost of A100s can add up.
For example, an estimate from New Street Research found that the OpenAI-based ChatGPT model inside Bing’s search could require 8 GPUs to deliver a response to a question in less than one second.
At that rate, Microsoft would need over 20,000 8-GPU servers just to deploy the model in Bing to everyone, suggesting Microsoft’s feature could cost $4 billion in infrastructure spending.
“If you’re from Microsoft, and you want to scale that, at the scale of Bing, that’s maybe $4 billion. If you want to scale at the scale of Google, which serves 8 or 9 billion queries every day, you actually need to spend $80 billion on DGXs.” said Antoine Chkaiban, a technology analyst at New Street Research. “The numbers we came up with are huge. But they’re simply the reflection of the fact that every single user taking to such a large language model requires a massive supercomputer while they’re using it.”
The latest version of Stable Diffusion, an image generator, was trained on 256 A100 GPUs, or 32 machines with 8 A100s each, according to information online posted by Stability AI, totaling 200,000 compute hours.
At the market price, training the model alone cost $600,000, Stability AI CEO Mostaque said on Twitter, suggesting in a tweet exchange the price was unusually inexpensive compared to rivals. That doesn’t count the cost of “inference,” or deploying the model.
Huang, Nvidia’s CEO, said in an interview with CNBC’s Katie Tarasov that the company’s products are actually inexpensive for the amount of computation that these kinds of models need.
“We took what otherwise would be a $1 billion data center running CPUs, and we shrunk it down into a data center of $100 million,” Huang said. “Now, $100 million, when you put that in the cloud and shared by 100 companies, is almost nothing.”
Huang said that Nvidia’s GPUs allow startups to train models for a much lower cost than if they used a traditional computer processor.
“Now you could build something like a large language model, like a GPT, for something like $10, $20 million,” Huang said. “That’s really, really affordable.”
New competition
Nvidia isn’t the only company making GPUs for artificial intelligence uses. AMD and Intel have competing graphics processors, and big cloud companies like Google and Amazon are developing and deploying their own chips specially designed for AI workloads.
Still, “AI hardware remains strongly consolidated to NVIDIA,” according to the State of AI compute report. As of December, more than 21,000 open-source AI papers said they used Nvidia chips.
Most researchersincluded in the State of AI Compute Index used the V100, Nvidia’s chip that came out in 2017, but A100 grew fast in 2022 to be the third-most used Nvidia chip, just behind a $1500-or-less consumer graphics chip originally intended for gaming.
The A100 also has the distinction of being one of only a few chips to have export controls placed on it because of national defense reasons. Last fall, Nvidia said in an SEC filing that the U.S. government imposed a license requirement barring the export of the A100 and the H100 to China, Hong Kong, and Russia.
“The USG indicated that the new license requirement will address the risk that the covered products may be used in, or diverted to, a ‘military end use’ or ‘military end user’ in China and Russia,” Nvidia said in its filing. Nvidia previously said it adapted some of its chips for the Chinese market to comply with U.S. export restrictions.
The fiercest competition for the A100 may be its successor. The A100 was first introduced in 2020, an eternity ago in chip cycles. The H100, introduced in 2022, is starting to be produced in volume — in fact, Nvidia recorded more revenue from H100 chips in the quarter ending in January than the A100, it said on Wednesday, although the H100 is more expensive per unit.
The H100, Nvidia says, is the first one of its data center GPUs to be optimized for transformers, an increasingly important technique that many of the latest and top AI applications use. Nvidia said on Wednesday that it wants to make AI training over 1 million percent faster. That could mean that, eventually, AI companies wouldn’t need so many Nvidia chips.
Baidu has launched a slew of AI applications after its Ernie chatbot received public approval.
Sopa Images | Lightrocket | Getty Images
Chinese tech giant Baidu saw its shares in Hong Kong soar nearly 16% on Wednesday as the company ramps up its artificial intelligence plans and partnerships.
Shares in the Beijing-based firm, which holds a dominant position in China’s search engine market, had gained nearly 8% overnight in U.S. trading.
The strong stock performance comes after Baidu earlier this week secured an AI-related deal with China Merchants Group, a major state-owned enterprise, focused on transportation, finance, and property development.
“Both sides plan to focus on applications of large language models, AI agents and ‘digital employees,’ vowing to make scalable and sustainable progress in industrial intelligence based on real-life business scenarios,” according to Baidu’s statement translated by CNBC.
Baidu has been aggressively pursuing its AI business, which includes its popular large language model and AI chatbot Ernie Bot.
As it seeks to gain an edge in China’s competitive AI space, the company on Tuesday disclosed a 4.4 billion yuan ($56.2 million) offshore bond offering. This follows a $2 billion bond issuance back in March.
Other Chinese AI players, such as Tencent, have also been raising funds, including via debt sales this year, to support the billions being poured into their AI capabilities.
Signs of AI strength
At a developer conference last week, Baidu unveiled a series of AI advancements, including the company’s latest reasoning model, Ernie X 1.1.
According to the company, multiple benchmark results showed that its model’s overall performance surpassed that of Chinese AI start-up DeepSeek’s latest reasoning model. CNBC could not independently verify that claim.
To train its AI models, the company has also started using internally designed chips, The Information reported last week, citing people with direct knowledge of the matter.
In addition to providing a new potential business venture, Baidu’s chip drive could help it reduce reliance on AI chips from Nvidia, which has been subject to shifting export controls from Washington.
Gimme Credit Senior Bond Analyst, Saurav Sen, said in a report last week that Baidu’s recent capital allocation revealed that the company is making an “all-in AI pivot.”
Baidu, whose Hong Kong shares have gained nearly 59% this year, reported a drop in second-quarter revenue last month as its core advertising business struggled and returns from AI investments remained limited.
Andy Jassy, CEO of Amazon, speaks during an unveiling event in New York on Feb. 26, 2025.
Michael Nagle | Bloomberg | Getty Images
Amazon CEO Andy Jassy said Tuesday that he’s working to root out bureaucracy from within the company’s ranks as part of an effort to reset its culture.
Speaking at Amazon’s annual conference for third-party sellers in Seattle, Jassy said the changes are necessary for the company to be able to innovate faster.
“I would say bureaucracy is really anathema to startups and to entrepreneurial organizations,” Jassy said. “As you get larger, it’s really easy to accumulate bureaucracy, a lot of bureaucracy that you may not see.”
A year ago, as part of a mandate requiring corporate employees to work in the office five days a week, Jassy set a goal to flatten organizations across Amazon. He called for the company to increase worker-to-manager ratios by at least 15% by the end of the first quarter of this year.
Jassy also announced the creation of a “no bureaucracy email alias” so that employees can flag unnecessary processes or excessive rules within the company.
Amazon has received about 1,500 emails in the past year, and the company has changed about 455 processes based on that feedback, Jassy said.
The changes are linked to Jassy’s broad strategy to overhaul Amazon’s corporate culture and operate like the “world’s largest startup” as it looks to stay competitive.
Jassy, who took the helm from founder Jeff Bezos in 2021, has been on a campaign to slash costs across the company in recent years. Amazon has laid off more than 27,000 employees since 2022, and axed some of its more unprofitable initiatives. Jassy has also urged employees to do more with less at the same time that the company invests heavily in artificial intelligence.
Transforming Amazon into a startup-like environment isn’t an easy task. The company operates sprawling businesses across retail, cloud computing, advertising, and other areas. It’s the U.S. second-largest private employer, with more than 1.5 million employees globally.
“You have to keep remembering your roots and how useful it is to be scrappy,” Jassy said.
The StubHub logo is seen at its headquarters in San Francisco.
Andrej Sokolow | Picture Alliance | Getty Images
Online ticket platform StubHub is pricing its IPO at $23.50, CNBC’s Leslie Picker confirmed on Tuesday.
The pricing comes at the midpoint of the expected range that the company gave last week. At $23.50, the pricing gives StubHub a valuation of $8.6 billion. StubHub will trade on the New York Stock Exchange under the symbol “STUB.”
The San Francisco-based company was co-founded by Eric Baker in 2000, and was acquired by eBay for $310 million seven years later. Baker reacquired StubHub in 2020 for roughly $4 billion through his new company Viagogo, which operates a ticket marketplace in Europe.
StubHub has been trying to go public for the past several years, but delayed its public debut twice. The most recent stall came in April after President Donald Trump‘s “Liberation Day” tariffs roiled markets.
The company filed an updated prospectus in August, effectively restarting the process to go public.
The IPO market has bounced back in recent months after an extended dry spell due to high inflation and rising interest rates. Klarna made its debut on the NYSE last week after the online lender also delayed its IPO in April. Tyler and Cameron Winklevoss’ Gemini, stablecoin issuer Circle, Peter Thiel-backed cryptocurrency exchangeBullish and design software company Figma have all soared in their respective debuts.
At the top of the pricing range StubHub offered last week, the company would have been valued at $9.2 billion. StubHub had sought a $16.5 billion valuation before it began the IPO process, CNBC previously reported.
StubHub said in its updated prospectus that first-quarter revenue increased 10% from a year earlier to $397.6 million. Operating income came in at $26.8 million for the period.
The company’s net loss widened to $35.9 million from $29.7 million a year ago.