Connect with us

Published

on

Nvidia CEO Jensen Huang speaks during a press conference at The MGM during CES 2018 in Las Vegas on January 7, 2018.

Mandel Ngan | AFP | Getty Images

Software that can write passages of text or draw pictures that look like a human created them has kicked off a gold rush in the technology industry.

Companies like Microsoft and Google are fighting to integrate cutting-edge AI into their search engines, as billion-dollar competitors such as OpenAI and Stable Diffusion race ahead and release their software to the public.

Powering many of these applications is a roughly $10,000 chip that’s become one of the most critical tools in the artificial intelligence industry: The Nvidia A100.

The A100 has become the “workhorse” for artificial intelligence professionals at the moment, said Nathan Benaich, an investor who publishes a newsletter and report covering the AI industry, including a partial list of supercomputers using A100s. Nvidia takes 95% of the market for graphics processors that can be used for machine learning, according to New Street Research.

A.I. is the catalyst behind Nividia's earnings beat, says Susquehanna's Christopher Rolland

The A100 is ideally suited for the kind of machine learning models that power tools like ChatGPT, Bing AI, or Stable Diffusion. It’s able to perform many simple calculations simultaneously, which is important for training and using neural network models.

The technology behind the A100 was initially used to render sophisticated 3D graphics in games. It’s often called a graphics processor, or GPU, but these days Nvidia’s A100 is configured and targeted at machine learning tasks and runs in data centers, not inside glowing gaming PCs.

Big companies or startups working on software like chatbots and image generators require hundreds or thousands of Nvidia’s chips, and either purchase them on their own or secure access to the computers from a cloud provider.

Hundreds of GPUs are required to train artificial intelligence models, like large language models. The chips need to be powerful enough to crunch terabytes of data quickly to recognize patterns. After that, GPUs like the A100 are also needed for “inference,” or using the model to generate text, make predictions, or identify objects inside photos.

This means that AI companies need access to a lot of A100s. Some entrepreneurs in the space even see the number of A100s they have access to as a sign of progress.

“A year ago we had 32 A100s,” Stability AI CEO Emad Mostaque wrote on Twitter in January. “Dream big and stack moar GPUs kids. Brrr.” Stability AI is the company that helped develop Stable Diffusion, an image generator that drew attention last fall, and reportedly has a valuation of over $1 billion.

Now, Stability AI has access to over 5,400 A100 GPUs, according to one estimate from the State of AI report, which charts and tracks which companies and universities have the largest collection of A100 GPUs — although it doesn’t include cloud providers, which don’t publish their numbers publicly.

Nvidia’s riding the A.I. train

Nvidia stands to benefit from the AI hype cycle. During Wednesday’s fiscal fourth-quarter earnings report, although overall sales declined 21%, investors pushed the stock up about 14% on Thursday, mainly because the company’s AI chip business — reported as data centers — rose by 11% to more than $3.6 billion in sales during the quarter, showing continued growth.

Nvidia shares are up 65% so far in 2023, outpacing the S&P 500 and other semiconductor stocks alike.

Nvidia CEO Jensen Huang couldn’t stop talking about AI on a call with analysts on Wednesday, suggesting that the recent boom in artificial intelligence is at the center of the company’s strategy.

“The activity around the AI infrastructure that we built, and the activity around inferencing using Hopper and Ampere to influence large language models has just gone through the roof in the last 60 days,” Huang said. “There’s no question that whatever our views are of this year as we enter the year has been fairly dramatically changed as a result of the last 60, 90 days.”

Ampere is Nvidia’s code name for the A100 generation of chips. Hopper is the code name for the new generation, including H100, which recently started shipping.

More computers needed

Nvidia A100 processor

Nvidia

Compared to other kinds of software, like serving a webpage, which uses processing power occasionally in bursts for microseconds, machine learning tasks can take up the whole computer’s processing power, sometimes for hours or days.

This means companies that find themselves with a hit AI product often need to acquire more GPUs to handle peak periods or improve their models.

These GPUs aren’t cheap. In addition to a single A100 on a card that can be slotted into an existing server, many data centers use a system that includes eight A100 GPUs working together.

This system, Nvidia’s DGX A100, has a suggested price of nearly $200,000, although it comes with the chips needed. On Wednesday, Nvidia said it would sell cloud access to DGX systems directly, which will likely reduce the entry cost for tinkerers and researchers.

It’s easy to see how the cost of A100s can add up.

For example, an estimate from New Street Research found that the OpenAI-based ChatGPT model inside Bing’s search could require 8 GPUs to deliver a response to a question in less than one second.

At that rate, Microsoft would need over 20,000 8-GPU servers just to deploy the model in Bing to everyone, suggesting Microsoft’s feature could cost $4 billion in infrastructure spending.

“If you’re from Microsoft, and you want to scale that, at the scale of Bing, that’s maybe $4 billion. If you want to scale at the scale of Google, which serves 8 or 9 billion queries every day, you actually need to spend $80 billion on DGXs.” said Antoine Chkaiban, a technology analyst at New Street Research. “The numbers we came up with are huge. But they’re simply the reflection of the fact that every single user taking to such a large language model requires a massive supercomputer while they’re using it.”

The latest version of Stable Diffusion, an image generator, was trained on 256 A100 GPUs, or 32 machines with 8 A100s each, according to information online posted by Stability AI, totaling 200,000 compute hours.

At the market price, training the model alone cost $600,000, Stability AI CEO Mostaque said on Twitter, suggesting in a tweet exchange the price was unusually inexpensive compared to rivals. That doesn’t count the cost of “inference,” or deploying the model.

Huang, Nvidia’s CEO, said in an interview with CNBC’s Katie Tarasov that the company’s products are actually inexpensive for the amount of computation that these kinds of models need.

“We took what otherwise would be a $1 billion data center running CPUs, and we shrunk it down into a data center of $100 million,” Huang said. “Now, $100 million, when you put that in the cloud and shared by 100 companies, is almost nothing.”

Huang said that Nvidia’s GPUs allow startups to train models for a much lower cost than if they used a traditional computer processor.

“Now you could build something like a large language model, like a GPT, for something like $10, $20 million,” Huang said. “That’s really, really affordable.”

New competition

Nvidia isn’t the only company making GPUs for artificial intelligence uses. AMD and Intel have competing graphics processors, and big cloud companies like Google and Amazon are developing and deploying their own chips specially designed for AI workloads.

Still, “AI hardware remains strongly consolidated to NVIDIA,” according to the State of AI compute report. As of December, more than 21,000 open-source AI papers said they used Nvidia chips.

Most researchers included in the State of AI Compute Index used the V100, Nvidia’s chip that came out in 2017, but A100 grew fast in 2022 to be the third-most used Nvidia chip, just behind a $1500-or-less consumer graphics chip originally intended for gaming.

The A100 also has the distinction of being one of only a few chips to have export controls placed on it because of national defense reasons. Last fall, Nvidia said in an SEC filing that the U.S. government imposed a license requirement barring the export of the A100 and the H100 to China, Hong Kong, and Russia.

“The USG indicated that the new license requirement will address the risk that the covered products may be used in, or diverted to, a ‘military end use’ or ‘military end user’ in China and Russia,” Nvidia said in its filing. Nvidia previously said it adapted some of its chips for the Chinese market to comply with U.S. export restrictions.

The fiercest competition for the A100 may be its successor. The A100 was first introduced in 2020, an eternity ago in chip cycles. The H100, introduced in 2022, is starting to be produced in volume — in fact, Nvidia recorded more revenue from H100 chips in the quarter ending in January than the A100, it said on Wednesday, although the H100 is more expensive per unit.

The H100, Nvidia says, is the first one of its data center GPUs to be optimized for transformers, an increasingly important technique that many of the latest and top AI applications use. Nvidia said on Wednesday that it wants to make AI training over 1 million percent faster. That could mean that, eventually, AI companies wouldn’t need so many Nvidia chips.

Continue Reading

Technology

Musk’s xAI sues Apple, OpenAI alleging anticompetitive scheme harmed X, Grok

Published

on

By

Musk’s xAI sues Apple, OpenAI alleging anticompetitive scheme harmed X, Grok

Elon Musk, CEO of SpaceX and Tesla, attends the Viva Technology conference at the Porte de Versailles exhibition center in Paris on June 16, 2023.

Gonzalo Fuentes | Reuters

Elon Musk‘s xAI sued Apple and OpenAI on Monday, accusing the pair of an “anticompetitive scheme” to thwart artificial intelligence rivals.

The lawsuit, filed by Musk’s AI startup xAI and its social network business X, alleges Apple and OpenAI have “colluded” to maintain monopolies in the smartphone and generative AI markets.

Musk’s xAI acquired X in March in an all-stock transaction.

It accuses Apple of deprioritizing so-called “super apps” and generative AI chatbot competitors, such as xAI’s Grok, in its App Store rankings, while favoring OpenAI by integrating its ChatGPT chatbot into Apple products.

“In a desperate bid to protect its smartphone monopoly, Apple has joined forces with the company that most benefits from inhibiting competition and innovation in AI: OpenAI, a monopolist in the market for generative AI chatbots,” according to the complaint, which was filed in U.S. District Court for the Northern District of Texas.

An OpenAI spokesperson said in a statement: “This latest filing is consistent with Mr. Musk’s ongoing pattern of harassment.”

Representatives from Apple didn’t immediately respond to a request for comment.

The Tesla CEO launched xAI in 2023 in a bid to compete with OpenAI and other leading chatbot makers.

Read more CNBC tech news

Musk earlier this month threatened to sue Apple for “an unequivocal antitrust violation,” saying in a post on X that the company “is behaving in a manner that makes it impossible for any AI company besides OpenAI to reach #1 in the App Store.”

After Musk threatened to sue Apple, OpenAI CEO Sam Altman responded: “This is a remarkable claim given what I have heard alleged that Elon does to manipulate X to benefit himself and his own companies and harm his competitors and people he doesn’t like.”

An Apple spokesperson previously said its App Store was designed to be “fair and free of bias,” and that the company features “thousands of apps” using a variety of signals.

Apple last year partnered with OpenAI to integrate ChatGPT into iPhone, iPad, Mac laptop and desktop products.

Several users replied to Musk’s post on X via its Community Notes feature saying that rival chatbot apps such as DeepSeek and Perplexity were ranked No. 1 on the App Store after Apple and OpenAI announced their partnership.

The lawsuit is the latest twist in an ongoing clash between Musk and Altman. Musk co-founded OpenAI alongside Altman in 2015, before leaving the startup in 2018 due to disagreements over OpenAI’s direction.

Musk sued OpenAI and Altman last year, accusing them of breach of contract by putting commercial interests ahead of its original mission to develop AI “for the benefit of humanity broadly.”

In a counter claim, OpenAI has alleged that Musk and xAI engaged in “harassment” through litigation, attacks on social media and in the press, and through a “sham bid” to buy the ChatGPT-maker for $97.4 billion designed to harm the company’s business relationships.

OpenAI says Musk's filing is 'consistent with his ongoing pattern of harassment

Continue Reading

Technology

Nvidia’s new ‘robot brain’ goes on sale for $3,499 as company targets robotics for growth

Published

on

By

Nvidia's new 'robot brain' goes on sale for ,499 as company targets robotics for growth

Jensen Huang, CEO of Nvidia, is seen on stage next to a small robot during the Viva Technology conference dedicated to innovation and startups at Porte de Versailles exhibition center in Paris, France, on June 11, 2025.

Gonzalo Fuentes | Reuters

Nvidia announced Monday that its latest robotics chip module, the Jetson AGX Thor, is now on sale for $3,499 as a developer kit.

The company calls the chip a “robot brain.” The first kits ship next month, Nvidia said last week, and the chips will allow customers to create robots.

After a company uses the developer kit to prototype their robot, Nvidia will sell Thor T5000 modules that can be installed in production-ready robots. If a company needs more than 1,000 Thor chips, Nvidia will charge $2,999 per module.

CEO Jensen Huang has said robotics is the company’s largest growth opportunity outside of artificial intelligence, which has led to the Nvidia’s overall sales more than tripling in the past two years.

“We do not build robots, we do not build cars, but we enable the whole industry with our infrastructure computers and the associated software,” said Deepu Talla, Nvidia’s vice president of robotics and edge AI, on a call with reporters Friday.

The Jetson Thor chips are based on a Blackwell graphics processor, which is Nvidia’s current generation of technology used in its AI chips, as well as its chips for computer games.

Nvidia said that its Jetson Thor chips are 7.5 times faster than its previous generation. That allows them to run generative AI models, including large language models and visual models that can interpret the world around them, which is essential for humanoid robots, Nvidia said. The Jetson Thor chips are equipped with 128GB of memory, which is essential for big AI models.

Companies including Agility Robotics, Amazon, Meta and Boston Dynamics are using its Jetson chips, Nvidia said. Nvidia has also invested in robotics companies such as Field AI.

However, robotics remains a small business for Nvidia, accounting for about 1% of the company’s total revenue, despite the fact that it has launched several new robot chips since 2014. But it’s growing fast.

Nvidia recently combined its business units to group its automotive and robotics divisions into the same line item. That unit reported $567 million in quarterly sales in May, which represented a 72% increase on an annual basis.

The company said its Jetson Thor chips can be used for self-driving cars as well, especially from Chinese brands. Nvidia calls its car chips Drive AGX, and while they are similar to its robotics chips, they run an operating system called Drive OS that’s been tuned for automotive purposes.

Don’t miss these insights from CNBC PRO

Nvidia's China risks loom large ahead of earnings

Continue Reading

Technology

Intel says Trump deal has risks for shareholders, international sales

Published

on

By

Intel says Trump deal has risks for shareholders, international sales

Intel’s CEO Lip-Bu Tan speaks at the company’s Annual Manufacturing Technology Conference in San Jose, California, U.S. April 29, 2025.

Laure Andrillon | Reuters

Intel on Monday warned of “adverse reactions” from investors, employees and others to the Trump administration taking a 10% stake in the company, in a filing citing risks involved with the deal.

A key concern area is international sales, with 76% of Intel’s revenue in its last fiscal year coming from outside the U.S., according to the filing with the Securities and Exchange Commission. The company had $53.1 billion in revenue for fiscal year 2024, down 2% from the year prior.

For Intel’s international customers, the company is now directly tied to President Donald Trump‘s ever-shifting tariff and trade policies.

“There could be adverse reactions, immediately or over time, from investors, employees, customers, suppliers, other business or commercial partners, foreign governments or competitors,” the company wrote in the filing. “There may also be litigation related to the transaction or otherwise and increased public or political scrutiny with respect to the Company.”

Intel also said that the potential for a changing political landscape in Washington could challenge or void the deal and create risks to current and future shareholders.

The deal, which was announced Friday, gives the Department of Commerce up to 433.3 million shares of the company, which is dilutive to existing shareholders. The purchase of shares is being funded largely by money already awarded to Intel under President Joe Biden‘s CHIPS Act.

Read more CNBC tech news

Intel has already received $2.2 billion from the program and is set for another $5.7 billion. A separate federal program awarded $3.2 billion, for a total of $11.1 billion, according to a release.

Trump called the agreement “a great Deal for America” and said the building of advanced chips “is fundamental to the future of our Nation.” 

Shares of Intel rallied as momentum built toward a deal in August, with the stock up about 25%.

The agreement requires the government to vote with Intel’s board of directors. In the Monday filing, the company noted that the government stake “reduces the voting and other governance rights of stockholders and may limit potential future transactions that may be beneficial to stockholders.”

Furthermore, the company acknowledged in the filing that it has not completed an analysis of all “financial, tax and accounting implications.”

Intel’s tumultuous fiscal year 2024 included the exit of CEO Pat Gelsinger in December after a four-year tenure during which the stock price tanked and the company lost ground to rivals in the artificial intelligence boom.

CEO Lip-Bu Tan took the helm in March.

NEC Director Kevin Hassett on Intel deal: It's possible government will take stake in more companies

Continue Reading

Trending