Nvidia CEO Jensen Huang speaks during a press conference at The MGM during CES 2018 in Las Vegas on January 7, 2018.
Mandel Ngan | AFP | Getty Images
Software that can write passages of text or draw pictures that look like a human created them has kicked off a gold rush in the technology industry.
Companies like Microsoft and Google are fighting to integrate cutting-edge AI into their search engines, as billion-dollar competitors such as OpenAI and Stable Diffusion race ahead and release their software to the public.
Powering many of these applications is a roughly $10,000 chip that’s become one of the most critical tools in the artificial intelligence industry: The Nvidia A100.
The A100 has become the “workhorse” for artificial intelligence professionals at the moment, said Nathan Benaich, an investor who publishes a newsletter and report covering the AI industry, including a partial list of supercomputers using A100s. Nvidia takes 95% of the market for graphics processors that can be used for machine learning, according to New Street Research.
The A100 is ideally suited for the kind of machine learning models that power tools like ChatGPT, Bing AI, or Stable Diffusion. It’s able to perform many simple calculations simultaneously, which is important for training and using neural network models.
The technology behind the A100 was initially used to render sophisticated 3D graphics in games. It’s often called a graphics processor, or GPU, but these days Nvidia’s A100 is configured and targeted at machine learning tasks and runs in data centers, not inside glowing gaming PCs.
Big companies or startups working on software like chatbots and image generators require hundreds or thousands of Nvidia’s chips, and either purchase them on their own or secure access to the computers from a cloud provider.
Hundreds of GPUsare required to train artificial intelligence models, like large language models. The chips need to be powerful enough to crunch terabytes of data quickly to recognize patterns. After that, GPUs like the A100 are also needed for “inference,” or using the model to generate text, make predictions, or identify objects inside photos.
This means that AI companies need access to a lot of A100s. Some entrepreneurs in the space even see the number of A100s they have access to as a sign of progress.
“A year ago we had 32 A100s,” Stability AI CEO Emad Mostaque wrote on Twitter in January. “Dream big and stack moar GPUs kids. Brrr.” Stability AI is the company that helped develop Stable Diffusion, an image generator that drew attention last fall, and reportedly has a valuation of over $1 billion.
Now, Stability AI has access to over 5,400 A100 GPUs, according to one estimate from the State of AI report, which charts and tracks which companies and universities have the largest collection of A100 GPUs — although it doesn’t include cloud providers, which don’t publish their numbers publicly.
Nvidia’s riding the A.I. train
Nvidia stands to benefit from the AI hype cycle. During Wednesday’s fiscal fourth-quarter earnings report, although overall sales declined 21%, investors pushed the stock up about 14% on Thursday, mainly because the company’s AI chip business — reported as data centers — rose by 11% to more than $3.6 billion in sales during the quarter, showing continued growth.
Nvidia shares are up 65% so far in 2023, outpacing the S&P 500 and other semiconductor stocks alike.
Nvidia CEO Jensen Huang couldn’t stop talking about AI on a call with analysts on Wednesday, suggesting that the recent boom in artificial intelligence is at the center of the company’s strategy.
“The activity around the AI infrastructure that we built, and the activity around inferencing using Hopper and Ampere to influence large language models has just gone through the roof in the last 60 days,” Huang said. “There’s no question that whatever our views are of this year as we enter the year has been fairly dramatically changed as a result of the last 60, 90 days.”
Ampere is Nvidia’s code name for the A100 generation of chips. Hopper is the code name for the new generation, including H100, which recently started shipping.
More computers needed
Nvidia A100 processor
Nvidia
Compared to other kinds of software, like serving a webpage, which uses processing power occasionally in bursts for microseconds, machine learning tasks can take up the whole computer’s processing power, sometimes for hours or days.
This means companies that find themselves with a hit AI product often need to acquire more GPUs to handle peak periods or improve their models.
These GPUs aren’t cheap. In addition to a single A100 on a card that can be slotted into an existing server, many data centers use a system that includes eight A100 GPUs working together.
This system, Nvidia’s DGX A100, has a suggested price of nearly $200,000, although it comes with the chips needed. On Wednesday, Nvidia said it would sell cloud access to DGX systems directly, which will likely reduce the entry cost for tinkerers and researchers.
It’s easy to see how the cost of A100s can add up.
For example, an estimate from New Street Research found that the OpenAI-based ChatGPT model inside Bing’s search could require 8 GPUs to deliver a response to a question in less than one second.
At that rate, Microsoft would need over 20,000 8-GPU servers just to deploy the model in Bing to everyone, suggesting Microsoft’s feature could cost $4 billion in infrastructure spending.
“If you’re from Microsoft, and you want to scale that, at the scale of Bing, that’s maybe $4 billion. If you want to scale at the scale of Google, which serves 8 or 9 billion queries every day, you actually need to spend $80 billion on DGXs.” said Antoine Chkaiban, a technology analyst at New Street Research. “The numbers we came up with are huge. But they’re simply the reflection of the fact that every single user taking to such a large language model requires a massive supercomputer while they’re using it.”
The latest version of Stable Diffusion, an image generator, was trained on 256 A100 GPUs, or 32 machines with 8 A100s each, according to information online posted by Stability AI, totaling 200,000 compute hours.
At the market price, training the model alone cost $600,000, Stability AI CEO Mostaque said on Twitter, suggesting in a tweet exchange the price was unusually inexpensive compared to rivals. That doesn’t count the cost of “inference,” or deploying the model.
Huang, Nvidia’s CEO, said in an interview with CNBC’s Katie Tarasov that the company’s products are actually inexpensive for the amount of computation that these kinds of models need.
“We took what otherwise would be a $1 billion data center running CPUs, and we shrunk it down into a data center of $100 million,” Huang said. “Now, $100 million, when you put that in the cloud and shared by 100 companies, is almost nothing.”
Huang said that Nvidia’s GPUs allow startups to train models for a much lower cost than if they used a traditional computer processor.
“Now you could build something like a large language model, like a GPT, for something like $10, $20 million,” Huang said. “That’s really, really affordable.”
New competition
Nvidia isn’t the only company making GPUs for artificial intelligence uses. AMD and Intel have competing graphics processors, and big cloud companies like Google and Amazon are developing and deploying their own chips specially designed for AI workloads.
Still, “AI hardware remains strongly consolidated to NVIDIA,” according to the State of AI compute report. As of December, more than 21,000 open-source AI papers said they used Nvidia chips.
Most researchersincluded in the State of AI Compute Index used the V100, Nvidia’s chip that came out in 2017, but A100 grew fast in 2022 to be the third-most used Nvidia chip, just behind a $1500-or-less consumer graphics chip originally intended for gaming.
The A100 also has the distinction of being one of only a few chips to have export controls placed on it because of national defense reasons. Last fall, Nvidia said in an SEC filing that the U.S. government imposed a license requirement barring the export of the A100 and the H100 to China, Hong Kong, and Russia.
“The USG indicated that the new license requirement will address the risk that the covered products may be used in, or diverted to, a ‘military end use’ or ‘military end user’ in China and Russia,” Nvidia said in its filing. Nvidia previously said it adapted some of its chips for the Chinese market to comply with U.S. export restrictions.
The fiercest competition for the A100 may be its successor. The A100 was first introduced in 2020, an eternity ago in chip cycles. The H100, introduced in 2022, is starting to be produced in volume — in fact, Nvidia recorded more revenue from H100 chips in the quarter ending in January than the A100, it said on Wednesday, although the H100 is more expensive per unit.
The H100, Nvidia says, is the first one of its data center GPUs to be optimized for transformers, an increasingly important technique that many of the latest and top AI applications use. Nvidia said on Wednesday that it wants to make AI training over 1 million percent faster. That could mean that, eventually, AI companies wouldn’t need so many Nvidia chips.
Startup Figure AI is developing general-purpose humanoid robots.
Figure AI
Figure AI, an Nvidia-backed developer of humanoid robots, was sued by the startup’s former head of product safety who alleged that he was wrongfully terminated after warning top executives that the company’s robots “were powerful enough to fracture a human skull.”
Robert Gruendel, a principal robotic safety engineer, is the plaintiff in the suit filed Friday in a federal court in the Northern District of California. Gruendel’s attorneys describe their client as a whistleblower who was fired in September, days after lodging his “most direct and documented safety complaints.”
The suit lands two months after Figure was valued at $39 billion in a funding round led by Parkway Venture Capital. That’s a 15-fold increase in valuation from early 2024, when the company raised a round from investors including Jeff Bezos, Nvidia, and Microsoft.
In the complaint, Gruendel’s lawyers say the plaintiff warned Figure CEO Brett Adcock and Kyle Edelberg, chief engineer, about the robot’s lethal capabilities, and said one “had already carved a ¼-inch gash into a steel refrigerator door during a malfunction.”
The complaint also says Gruendel warned company leaders not to “downgrade” a “safety road map” that he had been asked to present to two prospective investors who ended up funding the company.
Gruendel worried that a “product safety plan which contributed to their decision to invest” had been “gutted” the same month Figure closed the investment round, a move that “could be interpreted as fraudulent,” the suit says.
The plaintiff’s concerns were “treated as obstacles, not obligations,” and the company cited a “vague ‘change in business direction’ as the pretext” for his termination, according to the suit.
Gruendel is seeking economic, compensatory and punitive damages and demanding a jury trial.
Figure didn’t immediately respond to a request for comment. Nor did attorneys for Gruendel.
The humanoid robot market remains nascent today, with companies like Tesla and Boston Dynamics pursuing futuristic offerings, alongside Figure, while China’s Unitree Robotics is preparing for an IPO. Morgan Stanley said in a report in May that adoption is “likely to accelerate in the 2030s” and could top $5 trillion by 2050.
Concerns about stock valuations in companies tied to artificial intelligence knocked the market around this week. Whether these worries will recede, as they did Friday, or flare up again will certainly be something to watch in the days and weeks ahead. We understand the concerns about valuations in the speculative aspects of the AI trade, such as nuclear stocks and neoclouds. Jim Cramer has repeatedly warned about them. But, in the past week, the broader AI cohort — including real companies that make money and are driving what many are calling the fourth industrial revolution — has been getting hit. We own many of them: Nvidia and Broadcom on the chip side, and GE Vernova and Eaton on the derivative trade of powering these energy-gobbling AI data centers. That’s not what should be happening based on their fundamentals. Outside of valuations, worries also center on capital expenditures and the depreciation that results from massive investments in AI infrastructure. On this point, investors face a choice. You can go with the bears who are glued to their spreadsheets and extrapolating the usable life of tech assets based on history, a seemingly understandable approach, and applying those depreciation rates to their financial models, arguing the chips should be near worthless after three years. Or, you can go with the commentary from management teams running the largest companies driving the AI trade, and what Jim has gleaned from talking with the smartest CEOs in the world. When it comes to the real players driving this AI investment cycle, like the ones we’re invested in, we don’t think valuations are all that high or unreasonable when you consider their growth rates and importance to the U.S., and by extension, the global economy. We’re talking about Nvidia CEO Jensen Huang, who would tell you that advancements in his company’s CUDA software have extended the life of GPU chip platforms to roughly five to six years. Don’t forget, CoreWeave recently re-contracted for H100s from Nvidia, which were released in late 2022. The bears with their spreadsheets would tell you those chips are worthless. However, we know that H100s have held most of their value. Or listen to Lisa Su, CEO of Advanced Micro Devices , who said last week that her customers are at the point now where “they can see the return on the other side” of these massive investments. For our part, we understand the spending concerns and the depreciation issues that will arise if these companies are indeed overstating the useful lives of these assets. However, those who have bet against the likes of Jensen Huang and Lisa Su, or Meta Platforms CEO Mark Zuckerberg, Microsoft CEO Satya Nadella, and others who have driven innovation in the tech world for over a decade, have been burned time and again. While the bears’ concerns aren’t invalid, long-term investors are better off taking their cues from technology experts. AI is real, and it will increasingly lead to productivity gains as adoption ramps up and the technology becomes ingrained in our everyday lives, just as the internet has. We have faith in the management teams of the AI stocks in which we are invested, and while faith is not an investment strategy, that faith is based on a historical track record of strong execution, the knowledge that offerings from these companies are best in class, and scrutiny of their underlying business fundamentals and financial profiles. Siding with these technology expert management teams, over the loud financial expert bears, has kept us on the right side of the trade for years, and we don’t see that changing in the future. (See here for a full list of the stocks in Jim Cramer’s Charitable Trust, including NVDA, AVGO, GEV, ETN, META, MSFT.) As a subscriber to the CNBC Investing Club with Jim Cramer, you will receive a trade alert before Jim makes a trade. Jim waits 45 minutes after sending a trade alert before buying or selling a stock in his charitable trust’s portfolio. If Jim has talked about a stock on CNBC TV, he waits 72 hours after issuing the trade alert before executing the trade. THE ABOVE INVESTING CLUB INFORMATION IS SUBJECT TO OUR TERMS AND CONDITIONS AND PRIVACY POLICY , TOGETHER WITH OUR DISCLAIMER . NO FIDUCIARY OBLIGATION OR DUTY EXISTS, OR IS CREATED, BY VIRTUE OF YOUR RECEIPT OF ANY INFORMATION PROVIDED IN CONNECTION WITH THE INVESTING CLUB. NO SPECIFIC OUTCOME OR PROFIT IS GUARANTEED.
Every weekday, the CNBC Investing Club with Jim Cramer releases the Homestretch — an actionable afternoon update, just in time for the last hour of trading on Wall Street. Markets: The S & P 500 bounced back Friday, recovering from the prior session’s sharp losses. The broad-based index, which was still tracking for a nearly 1.5% weekly decline, started off the session a little shaky as Club stock Nvidia drifted lower after the open. It was looking like concerns about the artificial intelligence trade, which have been dogging the market, were going to dominate back-to-back sessions. But when New York Federal Reserve President John Williams suggested that central bankers could cut interest rates for a third time this year, the market jumped higher. Rate-sensitive stocks saw big gains Friday. Home Depot rose more than 3.5% on the day, mitigating a tough week following Tuesday’s lackluster quarterly release. Eli Lilly hit an all-time high, becoming the first drugmaker to reach a $1 trillion market cap. TJX also topped its all-time high after the off-price retailer behind T.J. Maxx, Marshalls, and HomeGoods, delivered strong quarterly results Wednesday. Carry trade: We’re also monitoring developments in Japan, which is dealing with its own inflation problem and questions about whether to resume interest rate hikes. That brings us to the popular Japanese yen carry trade, which is getting squeezed as borrowing costs there are rising. The yen carry trade involves borrowing yen at a low rate, then converting them into, say, dollars, and investing in higher-yielding foreign assets. That’s all well and good when the cost to borrow yen is low. It’s a different story now that borrowing costs in Japan are hitting 30-year highs. When rates rise, the profit margin on the carry trade gets crunched, or vanishes completely. As a result, investors need to get out, which means forced selling and price action that becomes divorced from fundamentals. It’s unclear if any of this is adding pressure to U.S. markets. We didn’t see anything in the recent quarterly earnings reports from U.S. companies to suggest corporate fundamentals are deteriorating in any meaningful way. That’s why we’re looking for other potential external factors, alongside the well-known concerns about artificial intelligence spending, the depreciation resulting from those capital expenditures, and general worries about consumer sentiment and inflation here in America. Wall Street call: HSBC downgraded Palo Alto Networks to a sell-equivalent rating from a hold following the company’s quarterly earnings report Wednesday. Analysts, who left their $157 price target unchanged, cited decelerating sales growth as the driver of the rerating, describing the quarter as “sufficient, not transformational.” Still, the Club name delivered a beat-and-raise quarter, which topped estimates across every key metric. None of this stopped Palo Alto shares from falling on the release. We chalked the post-earnings decline up to high expectations heading into the quarter, coupled with investor concerns over a new acquisition of cloud management and monitoring company Chronosphere. Palo Alto is still working to close its multi-billion-dollar acquisition of identity security company CyberArk , announced in July. HSBC now argues the stock’s risk-versus-reward is turning negative, with limited potential for upward estimate revisions for fiscal years 2026 and 2027. We disagree with HSBC’s call, given the momentum we’re seeing across Palo Alto’s businesses. The cybersecurity leader is dominating through its “platformization” strategy, which bundles its products and services. Plus, Palo Alto keeps adding net new platformizations each quarter, converting customers to use its security platform, and is on track to reach its fiscal 2030 target. We also like management’s playbook for acquiring businesses just before they see an industry inflection point. With Chronosphere, Palo Alto believes the entire observability industry needs to change due to the growing presence of AI. We’re reiterating our buy-equivalent 1 rating and $225 price target on the stock. Up next: There are no Club earnings reports next week. Outside of the portfolio, Symbotic, Zoom Communications , Semtech , and Fluence Energy will report after Monday’s close. Wall Street will also get a slew of delayed economic data during the shortened holiday trading week. U.S. retail sales and September’s consumer price index are scheduled for release early Tuesday. Durable goods orders and the Conference Board consumer sentiment are released on Wednesday morning. (See here for a full list of the stocks in Jim Cramer’s Charitable Trust.) As a subscriber to the CNBC Investing Club with Jim Cramer, you will receive a trade alert before Jim makes a trade. Jim waits 45 minutes after sending a trade alert before buying or selling a stock in his charitable trust’s portfolio. If Jim has talked about a stock on CNBC TV, he waits 72 hours after issuing the trade alert before executing the trade. THE ABOVE INVESTING CLUB INFORMATION IS SUBJECT TO OUR TERMS AND CONDITIONS AND PRIVACY POLICY , TOGETHER WITH OUR DISCLAIMER . NO FIDUCIARY OBLIGATION OR DUTY EXISTS, OR IS CREATED, BY VIRTUE OF YOUR RECEIPT OF ANY INFORMATION PROVIDED IN CONNECTION WITH THE INVESTING CLUB. NO SPECIFIC OUTCOME OR PROFIT IS GUARANTEED.