Connect with us

Published

on

Nvidia CEO Jensen Huang speaks during a press conference at The MGM during CES 2018 in Las Vegas on January 7, 2018.

Mandel Ngan | AFP | Getty Images

Software that can write passages of text or draw pictures that look like a human created them has kicked off a gold rush in the technology industry.

Companies like Microsoft and Google are fighting to integrate cutting-edge AI into their search engines, as billion-dollar competitors such as OpenAI and Stable Diffusion race ahead and release their software to the public.

Powering many of these applications is a roughly $10,000 chip that’s become one of the most critical tools in the artificial intelligence industry: The Nvidia A100.

The A100 has become the “workhorse” for artificial intelligence professionals at the moment, said Nathan Benaich, an investor who publishes a newsletter and report covering the AI industry, including a partial list of supercomputers using A100s. Nvidia takes 95% of the market for graphics processors that can be used for machine learning, according to New Street Research.

A.I. is the catalyst behind Nividia's earnings beat, says Susquehanna's Christopher Rolland

The A100 is ideally suited for the kind of machine learning models that power tools like ChatGPT, Bing AI, or Stable Diffusion. It’s able to perform many simple calculations simultaneously, which is important for training and using neural network models.

The technology behind the A100 was initially used to render sophisticated 3D graphics in games. It’s often called a graphics processor, or GPU, but these days Nvidia’s A100 is configured and targeted at machine learning tasks and runs in data centers, not inside glowing gaming PCs.

Big companies or startups working on software like chatbots and image generators require hundreds or thousands of Nvidia’s chips, and either purchase them on their own or secure access to the computers from a cloud provider.

Hundreds of GPUs are required to train artificial intelligence models, like large language models. The chips need to be powerful enough to crunch terabytes of data quickly to recognize patterns. After that, GPUs like the A100 are also needed for “inference,” or using the model to generate text, make predictions, or identify objects inside photos.

This means that AI companies need access to a lot of A100s. Some entrepreneurs in the space even see the number of A100s they have access to as a sign of progress.

“A year ago we had 32 A100s,” Stability AI CEO Emad Mostaque wrote on Twitter in January. “Dream big and stack moar GPUs kids. Brrr.” Stability AI is the company that helped develop Stable Diffusion, an image generator that drew attention last fall, and reportedly has a valuation of over $1 billion.

Now, Stability AI has access to over 5,400 A100 GPUs, according to one estimate from the State of AI report, which charts and tracks which companies and universities have the largest collection of A100 GPUs — although it doesn’t include cloud providers, which don’t publish their numbers publicly.

Nvidia’s riding the A.I. train

Nvidia stands to benefit from the AI hype cycle. During Wednesday’s fiscal fourth-quarter earnings report, although overall sales declined 21%, investors pushed the stock up about 14% on Thursday, mainly because the company’s AI chip business — reported as data centers — rose by 11% to more than $3.6 billion in sales during the quarter, showing continued growth.

Nvidia shares are up 65% so far in 2023, outpacing the S&P 500 and other semiconductor stocks alike.

Nvidia CEO Jensen Huang couldn’t stop talking about AI on a call with analysts on Wednesday, suggesting that the recent boom in artificial intelligence is at the center of the company’s strategy.

“The activity around the AI infrastructure that we built, and the activity around inferencing using Hopper and Ampere to influence large language models has just gone through the roof in the last 60 days,” Huang said. “There’s no question that whatever our views are of this year as we enter the year has been fairly dramatically changed as a result of the last 60, 90 days.”

Ampere is Nvidia’s code name for the A100 generation of chips. Hopper is the code name for the new generation, including H100, which recently started shipping.

More computers needed

Nvidia A100 processor

Nvidia

Compared to other kinds of software, like serving a webpage, which uses processing power occasionally in bursts for microseconds, machine learning tasks can take up the whole computer’s processing power, sometimes for hours or days.

This means companies that find themselves with a hit AI product often need to acquire more GPUs to handle peak periods or improve their models.

These GPUs aren’t cheap. In addition to a single A100 on a card that can be slotted into an existing server, many data centers use a system that includes eight A100 GPUs working together.

This system, Nvidia’s DGX A100, has a suggested price of nearly $200,000, although it comes with the chips needed. On Wednesday, Nvidia said it would sell cloud access to DGX systems directly, which will likely reduce the entry cost for tinkerers and researchers.

It’s easy to see how the cost of A100s can add up.

For example, an estimate from New Street Research found that the OpenAI-based ChatGPT model inside Bing’s search could require 8 GPUs to deliver a response to a question in less than one second.

At that rate, Microsoft would need over 20,000 8-GPU servers just to deploy the model in Bing to everyone, suggesting Microsoft’s feature could cost $4 billion in infrastructure spending.

“If you’re from Microsoft, and you want to scale that, at the scale of Bing, that’s maybe $4 billion. If you want to scale at the scale of Google, which serves 8 or 9 billion queries every day, you actually need to spend $80 billion on DGXs.” said Antoine Chkaiban, a technology analyst at New Street Research. “The numbers we came up with are huge. But they’re simply the reflection of the fact that every single user taking to such a large language model requires a massive supercomputer while they’re using it.”

The latest version of Stable Diffusion, an image generator, was trained on 256 A100 GPUs, or 32 machines with 8 A100s each, according to information online posted by Stability AI, totaling 200,000 compute hours.

At the market price, training the model alone cost $600,000, Stability AI CEO Mostaque said on Twitter, suggesting in a tweet exchange the price was unusually inexpensive compared to rivals. That doesn’t count the cost of “inference,” or deploying the model.

Huang, Nvidia’s CEO, said in an interview with CNBC’s Katie Tarasov that the company’s products are actually inexpensive for the amount of computation that these kinds of models need.

“We took what otherwise would be a $1 billion data center running CPUs, and we shrunk it down into a data center of $100 million,” Huang said. “Now, $100 million, when you put that in the cloud and shared by 100 companies, is almost nothing.”

Huang said that Nvidia’s GPUs allow startups to train models for a much lower cost than if they used a traditional computer processor.

“Now you could build something like a large language model, like a GPT, for something like $10, $20 million,” Huang said. “That’s really, really affordable.”

New competition

Nvidia isn’t the only company making GPUs for artificial intelligence uses. AMD and Intel have competing graphics processors, and big cloud companies like Google and Amazon are developing and deploying their own chips specially designed for AI workloads.

Still, “AI hardware remains strongly consolidated to NVIDIA,” according to the State of AI compute report. As of December, more than 21,000 open-source AI papers said they used Nvidia chips.

Most researchers included in the State of AI Compute Index used the V100, Nvidia’s chip that came out in 2017, but A100 grew fast in 2022 to be the third-most used Nvidia chip, just behind a $1500-or-less consumer graphics chip originally intended for gaming.

The A100 also has the distinction of being one of only a few chips to have export controls placed on it because of national defense reasons. Last fall, Nvidia said in an SEC filing that the U.S. government imposed a license requirement barring the export of the A100 and the H100 to China, Hong Kong, and Russia.

“The USG indicated that the new license requirement will address the risk that the covered products may be used in, or diverted to, a ‘military end use’ or ‘military end user’ in China and Russia,” Nvidia said in its filing. Nvidia previously said it adapted some of its chips for the Chinese market to comply with U.S. export restrictions.

The fiercest competition for the A100 may be its successor. The A100 was first introduced in 2020, an eternity ago in chip cycles. The H100, introduced in 2022, is starting to be produced in volume — in fact, Nvidia recorded more revenue from H100 chips in the quarter ending in January than the A100, it said on Wednesday, although the H100 is more expensive per unit.

The H100, Nvidia says, is the first one of its data center GPUs to be optimized for transformers, an increasingly important technique that many of the latest and top AI applications use. Nvidia said on Wednesday that it wants to make AI training over 1 million percent faster. That could mean that, eventually, AI companies wouldn’t need so many Nvidia chips.

Continue Reading

Technology

Elon Musk’s X temporarily down for tens of thousands of users

Published

on

By

Elon Musk's X temporarily down for tens of thousands of users

Elon Musk looks on as U.S. President Donald Trump meets South African President Cyril Ramaphosa in the Oval Office of the White House in Washington, D.C., U.S., May 21, 2025.

Kevin Lamarque | Reuters

The Elon Musk-owned social media platform X experienced a brief outage on Saturday morning, with tens of thousands of users reportedly unable to use the site.

About 25,000 users reported issues with the platform, according to the analytics platform Downdetector, which gathers data from users to monitor issues with various platforms.

Roughly 21,000 users reported issues just after 8:30 a.m. ET, per the analytics platform.

The issues appeared to be largely resolved by around 9:55 a.m., when about 2,000 users were reporting issues with the platform.

Read more CNBC politics coverage

X did not immediately respond to CNBC’s request for comment. Additional information on the outage was not available.

Musk, the billionaire owner of SpaceX and Tesla, acquired X, formerly known as Twitter in 2022.

The site has had a number of widespread outages since the acquisition.

The site experienced another outage in March, which Musk attributed at the time to a “massive cyberattack.”

“We get attacked every day, but this was done with a lot of resources,” Musk wrote in a post at the time.

This is breaking news. Check back for updates

Continue Reading

Technology

Companies turn to AI to navigate Trump tariff turbulence

Published

on

By

Companies turn to AI to navigate Trump tariff turbulence

Artificial intelligence robot looking at futuristic digital data display.

Yuichiro Chino | Moment | Getty Images

Businesses are turning to artificial intelligence tools to help them navigate real-world turbulence in global trade.

Several tech firms told CNBC say they’re deploying the nascent technology to visualize businesses’ global supply chains — from the materials that are used to form products, to where those goods are being shipped from — and understand how they’re affected by U.S. President Donald Trump’s reciprocal tariffs.

Last week, Salesforce said it had developed a new import specialist AI agent that can “instantly process changes for all 20,000 product categories in the U.S. customs system and then take action on them” as needed, to help navigate changes to tariff systems.

Engineers at the U.S. software giant used the Harmonized Tariff Schedule, a 4,400-page document of tariffs on goods imported to the U.S., to inform answers generated by the agent.

“The sheer pace and complexity of global tariff changes make it nearly impossible for most businesses to keep up manually,” Eric Loeb, executive vice president of government affairs at Salesforce, told CNBC. “In the past, companies might have relied on small teams of in-house experts to keep pace.”

Firms say that AI systems are enabling them to take decisions on adjustments to their global supply chains much faster.

Andrew Bell, chief product officer of supply chain management software firm Kinaxis, said that manufacturers and distributors looking to inform their response to tariffs are using his firm’s machine learning technology to assess their products and the materials that go into them, as well as external signals like news articles and macroeconomic data.

“With that information, we can start doing some of those simulations of, here is a particular part that is in your build material that has a significant tariff. If you switched to using this other part instead, what would the impact be overall?” Bell told CNBC.

‘AI’s moment to shine’

Trump’s tariffs list — which covers dozens of countries — has forced companies to rethink their supply chains and pricing, with the likes of Walmart and Nike already raising prices on some products. The U.S. imported about $3.3 trillion of goods in 2024, according to census data.

Uncertainty from the U.S. tariff measures “actually probably presents AI’s moment to shine,” Zack Kass, a futurist and former head of OpenAI’s go-to-market strategy, told CNBC’s Silvia Amaro at the Ambrosetti Forum in Italy last month.

Read more CNBC tech news

“If you wonder how hard things could get without AI vis-a-vis automation, and what would happen in a world where you can’t just employ a bunch of people overnight, AI presents this alternative proposal,” he added.

Nagendra Bandaru, managing partner and global head of technology services at Indian IT giant Wipro, said clients are using the company’s agentic AI solutions “to pivot supplier strategies, adjust trade lanes, and manage duty exposure dynamically as policy landscapes evolve.”

Wipro says it uses a range of AI systems — both proprietary and supplied by third parties — from large language models to traditional machine learning and computer vision techniques to inspect physical assets in cross-border transit.

‘Not a silver bullet’

While it preferred to keep company names confidential, Wipro said that firms using its AI products to navigate Trump’s tariffs range from a Fortune 500 electronics manufacturer with factories in Asia to an automotive parts supplier exporting to Europe and North America.

“AI is a powerful enabler — but not a silver bullet,” Bandaru told CNBC. “It doesn’t replace trade policy strategy, it enhances it by transforming global trade from a reactive challenge into a proactive, data-driven advantage.”

AI was already a key investment priority for global firms prior to Trump’s sweeping tariff announcements on April. Nearly three-quarters of business leaders ranked AI and generative AI in their top three technologies for investment in 2025, according to a report by Capgemini published in January.

“There are a number of ways AI can assist companies dealing with the tariffs and resulting uncertainty.  But any AI solution’s success will be predicated on the quality of the data it has access to,” Ajay Agarwal, partner at Bain Capital Ventures, told CNBC.

The venture capitalist said that one of his portfolio companies, FourKites, uses supply chain network data with AI to help firms understand the logistics impacts of adjusting suppliers due to tariffs.

“They are working with a number of Fortune 500 companies to leverage their agents for freight and ocean to provide this level of visibility and intelligence,” Agarwal said.

“Switching suppliers may reduce tariffs costs, but might increase lead times and transportation costs,” he added. “In addition, the volatility of the tariffs [has] severely impacted the rates and capacity available in both the ocean and the domestic freight networks.”

WATCH: Former OpenAI exec says tariffs ‘present AI’s moment to shine’

Former OpenAI exec says tariffs 'present AI's moment to shine'

Continue Reading

Technology

Amazon’s Zoox robotaxi unit issues second software recall in a month after San Francisco crash

Published

on

By

Amazon's Zoox robotaxi unit issues second software recall in a month after San Francisco crash

A Zoox autonomous robotaxi in San Francisco, California, US, on Wednesday, Dec. 4, 2024.

David Paul Morris | Bloomberg | Getty Images

Amazon‘s Zoox robotaxi unit issued a voluntary recall of its software for the second time in a month following a recent crash in San Francisco.

On May 8, an unoccupied Zoox robotaxi was turning at low speed when it was struck by an electric scooter rider after braking to yield at an intersection. The person on the scooter declined medical attention after sustaining minor injuries as a result of the collision, Zoox said.

“The Zoox vehicle was stopped at the time of contact,” the company said in a blog post. “The e-scooterist fell to the ground directly next to the vehicle. The robotaxi then began to move and stopped after completing the turn, but did not make further contact with the e-scooterist.”

Zoox said it submitted a voluntary software recall report to the National Highway Traffic Safety Administration on Thursday.

A Zoox spokesperson said the notice should be published on the NHTSA website early next week. The recall affected 270 vehicles, the spokesperson said.

The NHTSA said in a statement it had received the recall notice and that the agency “advises road users to be cautious in the vicinity of vehicles because drivers may incorrectly predict the travel path of a cyclist or scooter rider or come to an unexpected stop.”

If an autonomous vehicle continues to move after contact with any nearby vulnerable road user, it risks causing harm or further harm. In the AV industry, General Motors-backed Cruise exited the robotaxi business after a collision in which one of its vehicles injured a pedestrian who had been struck by a human-driven car and was then rolled over by the Cruise AV.

Zoox’s May incident comes roughly two weeks after the company announced a separate voluntary software recall following a recent Las Vegas crash. In that incident, an unoccupied Zoox robotaxi collided with a passenger vehicle, resulting in minor damage to both vehicles.

The company issued a software recall for 270 of its robotaxis in order to address a defect with its automated driving system that could cause it to inaccurately predict the movement of another car, increasing the “risk of a crash.”

Amazon acquired Zoox in 2020 for more than $1 billion, announcing at the time that the deal would help bring the self-driving technology company’s “vision for autonomous ride-hailing to reality.”

While Zoox is in a testing and development stage with its AVs on public roads in the U.S., Alphabet’s Waymo is already operating commercial, driverless ride-hailing services in Phoenix, San Francisco, Los Angeles and Austin, Texas, and is ramping up in Atlanta.

Tesla is promising it will launch its long-delayed robotaxis in Austin next month, and, if all goes well, plans to expand after that to San Francisco, Los Angeles and San Antonio, Texas.

— CNBC’s Lora Kolodny contributed to this report.

WATCH: Tesla’s decade-long journey to robotaxis

Tesla's decade-long journey to robotaxis

Continue Reading

Trending