Nvidia CEO Jensen Huang speaks during a press conference at The MGM during CES 2018 in Las Vegas on January 7, 2018.
Mandel Ngan | AFP | Getty Images
Software that can write passages of text or draw pictures that look like a human created them has kicked off a gold rush in the technology industry.
Companies like Microsoft and Google are fighting to integrate cutting-edge AI into their search engines, as billion-dollar competitors such as OpenAI and Stable Diffusion race ahead and release their software to the public.
Powering many of these applications is a roughly $10,000 chip that’s become one of the most critical tools in the artificial intelligence industry: The Nvidia A100.
The A100 has become the “workhorse” for artificial intelligence professionals at the moment, said Nathan Benaich, an investor who publishes a newsletter and report covering the AI industry, including a partial list of supercomputers using A100s. Nvidia takes 95% of the market for graphics processors that can be used for machine learning, according to New Street Research.
The A100 is ideally suited for the kind of machine learning models that power tools like ChatGPT, Bing AI, or Stable Diffusion. It’s able to perform many simple calculations simultaneously, which is important for training and using neural network models.
The technology behind the A100 was initially used to render sophisticated 3D graphics in games. It’s often called a graphics processor, or GPU, but these days Nvidia’s A100 is configured and targeted at machine learning tasks and runs in data centers, not inside glowing gaming PCs.
Big companies or startups working on software like chatbots and image generators require hundreds or thousands of Nvidia’s chips, and either purchase them on their own or secure access to the computers from a cloud provider.
Hundreds of GPUsare required to train artificial intelligence models, like large language models. The chips need to be powerful enough to crunch terabytes of data quickly to recognize patterns. After that, GPUs like the A100 are also needed for “inference,” or using the model to generate text, make predictions, or identify objects inside photos.
This means that AI companies need access to a lot of A100s. Some entrepreneurs in the space even see the number of A100s they have access to as a sign of progress.
“A year ago we had 32 A100s,” Stability AI CEO Emad Mostaque wrote on Twitter in January. “Dream big and stack moar GPUs kids. Brrr.” Stability AI is the company that helped develop Stable Diffusion, an image generator that drew attention last fall, and reportedly has a valuation of over $1 billion.
Now, Stability AI has access to over 5,400 A100 GPUs, according to one estimate from the State of AI report, which charts and tracks which companies and universities have the largest collection of A100 GPUs — although it doesn’t include cloud providers, which don’t publish their numbers publicly.
Nvidia’s riding the A.I. train
Nvidia stands to benefit from the AI hype cycle. During Wednesday’s fiscal fourth-quarter earnings report, although overall sales declined 21%, investors pushed the stock up about 14% on Thursday, mainly because the company’s AI chip business — reported as data centers — rose by 11% to more than $3.6 billion in sales during the quarter, showing continued growth.
Nvidia shares are up 65% so far in 2023, outpacing the S&P 500 and other semiconductor stocks alike.
Nvidia CEO Jensen Huang couldn’t stop talking about AI on a call with analysts on Wednesday, suggesting that the recent boom in artificial intelligence is at the center of the company’s strategy.
“The activity around the AI infrastructure that we built, and the activity around inferencing using Hopper and Ampere to influence large language models has just gone through the roof in the last 60 days,” Huang said. “There’s no question that whatever our views are of this year as we enter the year has been fairly dramatically changed as a result of the last 60, 90 days.”
Ampere is Nvidia’s code name for the A100 generation of chips. Hopper is the code name for the new generation, including H100, which recently started shipping.
More computers needed
Nvidia A100 processor
Nvidia
Compared to other kinds of software, like serving a webpage, which uses processing power occasionally in bursts for microseconds, machine learning tasks can take up the whole computer’s processing power, sometimes for hours or days.
This means companies that find themselves with a hit AI product often need to acquire more GPUs to handle peak periods or improve their models.
These GPUs aren’t cheap. In addition to a single A100 on a card that can be slotted into an existing server, many data centers use a system that includes eight A100 GPUs working together.
This system, Nvidia’s DGX A100, has a suggested price of nearly $200,000, although it comes with the chips needed. On Wednesday, Nvidia said it would sell cloud access to DGX systems directly, which will likely reduce the entry cost for tinkerers and researchers.
It’s easy to see how the cost of A100s can add up.
For example, an estimate from New Street Research found that the OpenAI-based ChatGPT model inside Bing’s search could require 8 GPUs to deliver a response to a question in less than one second.
At that rate, Microsoft would need over 20,000 8-GPU servers just to deploy the model in Bing to everyone, suggesting Microsoft’s feature could cost $4 billion in infrastructure spending.
“If you’re from Microsoft, and you want to scale that, at the scale of Bing, that’s maybe $4 billion. If you want to scale at the scale of Google, which serves 8 or 9 billion queries every day, you actually need to spend $80 billion on DGXs.” said Antoine Chkaiban, a technology analyst at New Street Research. “The numbers we came up with are huge. But they’re simply the reflection of the fact that every single user taking to such a large language model requires a massive supercomputer while they’re using it.”
The latest version of Stable Diffusion, an image generator, was trained on 256 A100 GPUs, or 32 machines with 8 A100s each, according to information online posted by Stability AI, totaling 200,000 compute hours.
At the market price, training the model alone cost $600,000, Stability AI CEO Mostaque said on Twitter, suggesting in a tweet exchange the price was unusually inexpensive compared to rivals. That doesn’t count the cost of “inference,” or deploying the model.
Huang, Nvidia’s CEO, said in an interview with CNBC’s Katie Tarasov that the company’s products are actually inexpensive for the amount of computation that these kinds of models need.
“We took what otherwise would be a $1 billion data center running CPUs, and we shrunk it down into a data center of $100 million,” Huang said. “Now, $100 million, when you put that in the cloud and shared by 100 companies, is almost nothing.”
Huang said that Nvidia’s GPUs allow startups to train models for a much lower cost than if they used a traditional computer processor.
“Now you could build something like a large language model, like a GPT, for something like $10, $20 million,” Huang said. “That’s really, really affordable.”
New competition
Nvidia isn’t the only company making GPUs for artificial intelligence uses. AMD and Intel have competing graphics processors, and big cloud companies like Google and Amazon are developing and deploying their own chips specially designed for AI workloads.
Still, “AI hardware remains strongly consolidated to NVIDIA,” according to the State of AI compute report. As of December, more than 21,000 open-source AI papers said they used Nvidia chips.
Most researchersincluded in the State of AI Compute Index used the V100, Nvidia’s chip that came out in 2017, but A100 grew fast in 2022 to be the third-most used Nvidia chip, just behind a $1500-or-less consumer graphics chip originally intended for gaming.
The A100 also has the distinction of being one of only a few chips to have export controls placed on it because of national defense reasons. Last fall, Nvidia said in an SEC filing that the U.S. government imposed a license requirement barring the export of the A100 and the H100 to China, Hong Kong, and Russia.
“The USG indicated that the new license requirement will address the risk that the covered products may be used in, or diverted to, a ‘military end use’ or ‘military end user’ in China and Russia,” Nvidia said in its filing. Nvidia previously said it adapted some of its chips for the Chinese market to comply with U.S. export restrictions.
The fiercest competition for the A100 may be its successor. The A100 was first introduced in 2020, an eternity ago in chip cycles. The H100, introduced in 2022, is starting to be produced in volume — in fact, Nvidia recorded more revenue from H100 chips in the quarter ending in January than the A100, it said on Wednesday, although the H100 is more expensive per unit.
The H100, Nvidia says, is the first one of its data center GPUs to be optimized for transformers, an increasingly important technique that many of the latest and top AI applications use. Nvidia said on Wednesday that it wants to make AI training over 1 million percent faster. That could mean that, eventually, AI companies wouldn’t need so many Nvidia chips.
Every weekday, the CNBC Investing Club with Jim Cramer holds a “Morning Meeting” livestream at 10:20 a.m. ET. Here’s a recap of Thursday’s key moments. 1. Stocks were little changed Thursday as Wall Street overlooked mixed labor market data. U.S. layoff announcements in November pushed the year’s total above 1.1 million, the highest level since 2020, according to job placement firm Challenger, Gray & Christmas. Initial jobless claims, however, came in lower than expected for the week ending Nov. 29. Despite the muted session, Jim Cramer says the market’s still overbought. That means we’re not looking to put new money to work right now. Meta Platforms was an outperformer in the portfolio. Shares jumped 4% after Bloomberg reported that the Facebook parent plans to make deep cuts to its metaverse unit. 2. Costco reported U.S. sales for November that were slightly weaker than the month before, sending shares down 3%. Company-wide same-store sales, however, accelerated last month, up 6.9% from October’s 6.6% gain. The stock’s weakness Thursday doesn’t present a buying opportunity just yet, according to Jim, because its multiple is still too high. “There are periods of underperformance in Costco, but if you look at the longer term, it’s one of the greatest performers of all time,” he added. 3. Salesforce stock was up after management posted a huge quarterly earnings beat and raised guidance Wednesday evening. The company missed slightly on revenue. We liked all the paid deals Agentforce, Salesforce’s AI-powered platform, pulled in this quarter. Still, generative AI adoption continues to pose a risk to Salesforce’s seat-based business model. CEO Marc Benioff will be on “Mad Money” on Thursday. 4. Stocks covered in Thursday’s rapid fire at the end of the video were: Snowflake , Five Below , Hormel Foods , PayPal , and Kroger. (Jim Cramer’s Charitable Trust is long META, CRM, COST. See here for a full list of the stocks.) As a subscriber to the CNBC Investing Club with Jim Cramer, you will receive a trade alert before Jim makes a trade. Jim waits 45 minutes after sending a trade alert before buying or selling a stock in his charitable trust’s portfolio. If Jim has talked about a stock on CNBC TV, he waits 72 hours after issuing the trade alert before executing the trade. THE ABOVE INVESTING CLUB INFORMATION IS SUBJECT TO OUR TERMS AND CONDITIONS AND PRIVACY POLICY , TOGETHER WITH OUR DISCLAIMER . NO FIDUCIARY OBLIGATION OR DUTY EXISTS, OR IS CREATED, BY VIRTUE OF YOUR RECEIPT OF ANY INFORMATION PROVIDED IN CONNECTION WITH THE INVESTING CLUB. NO SPECIFIC OUTCOME OR PROFIT IS GUARANTEED.
People walk next to the Google Cloud logo, during the 2025 Mobile World Congress (MWC) in Barcelona, Spain, March 4, 2025.
Albert Gea | Reuters
Google Cloud announced Thursday a multi-year partnership with artificial intelligence coding startup Replit, giving the search giant fresh firepower against the coding products of rivals, including Anthropic and Cursor.
Under the partnership, Replit will expand usage of Google Cloud services, add more of Google’s models onto its platform, and support AI coding use cases for enterprise customers.
Google will continue to be Replit’s primary cloud provider.
Replit, founded nearly a decade ago, is a leader in the fast-growing AI vibe-coding space.
In September, the startup closed a $250 million funding round that almost tripled its valuation to $3 billion, and said it grew annualized revenue from $2.8 million to $150 million in less than a year.
And new data from Ramp, a fintech company that also tracks enterprise spending on its platform, found that Replit had the fastest new customer growth among software vendors. Google, meanwhile, is adding new customers and spending faster than any other company on Ramp’s platform.
Put those together, and you get a clearer picture of why both companies see opportunity.
Read more CNBC tech news
Vibe-coding emerged as a phenomenon earlier this year after AI models became more adept at generating code using only natural language prompts, allowing users with little experience in programming to use AI to create functioning code and potentially full applications.
Anthropic announced on Tuesday that its product Claude Code hit $1 billion in run-rate revenue. The coding startup Cursor, in November, closed a funding round that valued it at $29.3 billion, while also announcing it reached $1 billion in annualized revenue.
Replit, which bills itself as an easy-to-use product for non-developers, could help drive Google Cloud adoption among enterprises, and expand the reach of its AI efforts beyond traditional engineers.
Google is riding on the momentum of its new top-scoring model, Gemini 3. Shares of Alphabet have risen more than 12% since its debut.
Is the “year of efficiency” Mark Zuckerberg back at Meta Platforms ? Shares of the social media giant rallied more than 5% to $676 each at Thursday’s highs after Bloomberg reported that Zuckerberg is set to reduce metaverse spending up to 30%. The metaverse group, which works on the company’s virtual “Horizon World” environment and Quest line of virtual reality headsets. It’s been a long time coming. Meta stock took a beating back in 2022 when, in addition to aggressive interest rate hikes from the Federal Reserve to combat sky-high inflation, investors grew concerned that Zuckerberg was going to spend countless sums of money building out a virtual world with little idea as to when, or even if, the investment would see any return. Since then, Zuckerberg has smartly avoided much talk about the metaverse. Wall Street is wondering whether cutting the metaverse budget is a true turning point for Zuckerberg, who has been on an artificial intelligence spending spree, both on the capital expenditures side and in poaching AI talent for top dollar, or whether it’s more about facing the reality that he’s been throwing good money after a vanity project that no one cares about. That spending question has been top of mind as Meta shares have dropped more than 20% since reporting earnings in late October , on fears that Zuckerberg was losing his way on efficiency and preparing to continue to ramp up investments without a clear view on returns. Judging by Thursday’s rally in Meta shares, the Bloomberg report has eased some of those concerns. We remember the power of spending discipline: Zuckerberg dubbed 2023 the “year of efficiency,” embarking on massive layoffs and cost-cutting. Shares surged nearly 195% in 2023, followed by a 65% gain last year. The metaverse and other VR-related investments are housed within the company’s Reality Labs operating segment, alongside the company’s smart glasses. Reality Labs lost just over $4.4 billion in the last quarter alone, with Bloomberg highlighting more than $70 billion in losses since its launch in 2021. The article does not appear to indicate a pullback in smart glasses investments, which, by all indications, have been better received by the mass market than the company’s virtual reality offerings. Bloomberg does report that Zuckerberg still believes in the metaverse and thinks that people will one day work and play in virtual worlds. META 5Y mountain Meta Platforms 5 years Our view? While focusing on a future metaverse isn’t wrong, it’s just a difficult narrative for investors to digest. When folks hear the term metaverse, they think about a digital playground that is nothing more than a highly immersive video game or entertainment experience. From that point of view, it’s easy to understand the skepticism about a return on investment on such a big swing. As Meta shareholders for the Club, we applaud any decision to cut spending on the more ambitious aspects of the metaverse vision, but take a different view of Zuckerberg’s north star. The technology needed to achieve his vision will still be invested in, just in a more methodical manner. Zuckerberg is choosing to focus on the technology that can be monetized more quickly, such as smart glasses and AI, while leaving open the idea of a metaverse-like world in the future. You aren’t going to be able to run a fully immersive digital world, in which players/users can interact, without AI. So, rather than talk about the grand vision of a metaverse, Zuckerberg can simply talk about AI and how it’s helping in the here and now, by reducing costs and boosting engagement, while aiding topline growth. That’s a narrative investors are all too happy to talk about. The recently released display glasses — which take the idea of smart glasses to the next level without the bulk of VR goggles — would certainly lend themselves to user interactions in a virtual environment. By tying that effort to the screenless Ray-Ban and Oakley smart glasses, Zuckerberg has a better chance to start monetizing the research R & D investments that went into the metaverse in the first place. We think that Zuckerberg’s long-term view hasn’t changed so much as he has learned to be more methodical in his long-term investing roadmap, while at the same time becoming a better storyteller as it relates to the narrative of these investments. We think this bodes well for 2026 earnings – perhaps a cost guidance cut with the next earnings release – and perhaps, even more important given Meta’s already attractive valuation, a reversal of the negative sentiment since the company last reported what were nothing short of fantastic quarterly results. Analysts at Mizuho are out with a note following the Bloomberg report, saying such metaverse cuts could add as much as $2 to 2026 earnings per share. Assuming a valuation multiple of 20 to 25 times earnings, that would be expected to add anywhere from $40 to $50 to the share price. “The stock is up, but it’s not up nearly as much as I think it could be given the fact that it’s only up 13% for the year, and is not expensive on a P/E multiple,” Jim Cramer said Thursday during the Club’s Morning Meeting . On a forward basis, the stock trades at 22.3 times full-year 2026 earnings estimates. That’s right in line with the S & P 500 ‘s valuation, despite expectations that Meta can grow earnings twice as fast in the coming year as the overall market. (Jim Cramer’s Charitable Trust is long META. See here for a full list of the stocks.) As a subscriber to the CNBC Investing Club with Jim Cramer, you will receive a trade alert before Jim makes a trade. Jim waits 45 minutes after sending a trade alert before buying or selling a stock in his charitable trust’s portfolio. If Jim has talked about a stock on CNBC TV, he waits 72 hours after issuing the trade alert before executing the trade. THE ABOVE INVESTING CLUB INFORMATION IS SUBJECT TO OUR TERMS AND CONDITIONS AND PRIVACY POLICY , TOGETHER WITH OUR DISCLAIMER . NO FIDUCIARY OBLIGATION OR DUTY EXISTS, OR IS CREATED, BY VIRTUE OF YOUR RECEIPT OF ANY INFORMATION PROVIDED IN CONNECTION WITH THE INVESTING CLUB. NO SPECIFIC OUTCOME OR PROFIT IS GUARANTEED.