Connect with us

Published

on

Jensen Huang, co-founder and chief executive officer of Nvidia Corp., during the Nvidia GPU Technology Conference (GTC) in San Jose, California, US, on Tuesday, March 19, 2024. 

David Paul Morris | Bloomberg | Getty Images

Nvidia’s 27% rally in May pushed its market cap to $2.7 trillion, behind only Microsoft and Apple among the most-valuable public companies in the world. The chipmaker reported a tripling in year-over-year sales for the third straight quarter driven by soaring demand for its artificial intelligence processors.

Mizuho Securities estimates that Nvidia controls between 70% and 95% of the market for AI chips used for training and deploying models like OpenAI’s GPT. Underscoring Nvidia’s pricing power is a 78% gross margin, a stunningly high number for a hardware company that has to manufacture and ship physical products.

Rival chipmakers Intel and Advanced Micro Devices reported gross margins in the latest quarter of 41% and 47%, respectively.

Nvidia’s position in the AI chip market has been described as a moat by some experts. Its flagship AI graphics processing units (GPUs), such as the H100, coupled with the company’s CUDA software led to such a head start on the competition that switching to an alternative can seem almost unthinkable.

Still, Nvidia CEO Jensen Huang, whose net worth has swelled from $3 billion to about $90 billion in the past five years, has said he’s “worried and concerned” about his 31-year-old company losing its edge. He acknowledged at a conference late last year that there are many powerful competitors on the rise.

“I don’t think people are trying to put me out of business,” Huang said in November. “I probably know they’re trying to, so that’s different.”

Nvidia has committed to releasing a new AI chip architecture every year, rather than every other year as was the case historically, and to putting out new software that could more deeply entrench its chips in AI software.

But Nvidia’s GPU isn’t alone in being able to run the complex math that underpins generative AI. If less powerful chips can do the same work, Huang might be justifiably paranoid.

The transition from training AI models to what’s called inference — or deploying the models — could also give companies an opportunity to replace Nvidia’s GPUs, especially if they’re less expensive to buy and run. Nvidia’s flagship chip costs roughly $30,000 or more, giving customers plenty of incentive to seek alternatives.

“Nvidia would love to have 100% of it, but customers would not love for Nvidia to have 100% of it,” said Sid Sheth, co-founder of aspiring rival D-Matrix. “It’s just too big of an opportunity. It would be too unhealthy if any one company took all of it.”

Founded in 2019, D-Matrix plans to release a semiconductor card for servers later this year that aims to reduce the cost and latency of running AI models. The company raised $110 million in September.

In addition to D-Matrix, companies ranging from multinational corporations to nascent startups are fighting for a slice of the AI chip market that could reach $400 billion in annual sales in the next five years, according to market analysts and AMD. Nvidia has generated about $80 billion in revenue over the past four quarters, and Bank of America estimates the company sold $34.5 billion in AI chips last year.

Many companies taking on Nvidia’s GPUs are betting that a different architecture or certain trade-offs could produce a better chip for particular tasks. Device makers are also developing technology that could end up doing a lot of the computing for AI that’s currently taking place in large GPU-based clusters in the cloud.

“Nobody can deny that today Nvidia is the hardware you want to train and run AI models,” Fernando Vidal, co-founder of 3Fourteen Research, told CNBC. “But there’s been incremental progress in leveling the playing field, from hyperscalers working on their own chips, to even little startups, designing their own silicon.”

AMD CEO Lisa Su wants investors to believe there’s plenty of room for many successful companies in the space.

“The key is that there are a lot of options there,” Su told reporters in December, when her company launched its most recent AI chip. “I think we’re going to see a situation where there’s not only one solution, there will be multiple solutions.”

Other big chipmakers

Lisa Su displays an AMD Instinct MI300 chip as she delivers a keynote address at CES 2023 in Las Vegas, Nevada, on Jan. 4, 2023.

David Becker | Getty Images

Nvidia’s top customers

How AWS is designing its own chips to help catch Microsoft and Google in generative A.I. race

One potential challenge for Nvidia is that it’s competing against some of its biggest customers. Cloud providers including Google, Microsoft and Amazon are all building processors for internal use. The Big Tech three, plus Oracle, make up over 40% of Nvidia’s revenue.

Amazon introduced its own AI-oriented chips in 2018, under the Inferentia brand name. Inferentia is now on its second version. In 2021, Amazon Web Services debuted Tranium targeted to training. Customers can’t buy the chips but they can rent systems through AWS, which markets the chips as more cost efficient than Nvidia’s.

Google is perhaps the cloud provider most committed to its own silicon. The company has been using what it calls Tensor Processing Units (TPUs) since 2015 to train and deploy AI models. In May, Google announced the sixth version of its chip, Trillium, which the company said was used to develop its models, including Gemini and Imagen.

Google also uses Nvidia chips and offers them through its cloud.

Microsoft isn’t as far along. The company said last year that it was building its own AI accelerator and processor, called Maia and Cobalt.

Meta isn’t a cloud provider, but the company needs massive amounts of computing power to run its software and website and to serve ads. While the Facebook parent company is buying billions of dollars worth of Nvidia processors, it said in April that some of its homegrown chips were already in data centers and enabled “greater efficiency” compared to GPUs.

JPMorgan analysts estimated in May that the market for building custom chips for big cloud providers could be worth as much as $30 billion, with potential growth of 20% per year.

Startups

Cerebras’ WSE-3 chip is one example of new silicon from upstarts designed to run and train artificial intelligence.

Cerebras Systems

Venture capitalists see opportunities for emerging companies to jump into the game. They invested $6 billion in AI semiconductor companies in 2023, up slightly from $5.7 billion a year earlier, according to data from PitchBook.

It’s a tough area for startups as semiconductors are expensive to design, develop and manufacture. But there are opportunities for differentiation.

For Cerebras Systems, an AI chipmaker in Silicon Valley, the focus is on basic operations and bottlenecks for AI, versus the more general purpose nature of a GPU. The company was founded in 2015 and was valued at $4 billion during its most recent fundraising, according to Bloomberg.

The Cerebras chip, WSE-2, puts GPU capabilities as well as central processing and additional memory into a single device, which is better for training large models, said CEO Andrew Feldman.

“We use a giant chip, they use a lot of little chips,” Feldman said. “They’ve got challenges of moving data around, we don’t.”

Feldman said his company, which counts Mayo Clinic, GlaxoSmithKline, and the U.S. Military as clients, is winning business for its supercomputing systems even going up against Nvidia.

“There’s ample competition and I think that’s healthy for the ecosystem,” Feldman said.

Sheth from D-Matrix said his company plans to release a card with its chiplet later this year that will allow for more computation in memory, as opposed to on a chip like a GPU. D-Matrix’s product can be slotted into an AI server along existing GPUs, but it takes work off of Nvidia chips, and helps to lower the cost of generative AI.

Customers “are very receptive and very incentivized to enable a new solution to come to market,” Sheth said.

Apple and Qualcomm

Apple iPhone 15 series devices are displayed for sale at The Grove Apple retail store on release day in Los Angeles, California, on September 22, 2023. 

Patrick T. Fallon | Afp | Getty Images

Don’t miss these exclusives from CNBC PRO

Nvidia and Microsoft has overshadowed weakness in tech stocks, says Solus' Dan Greenhaus

Continue Reading

Technology

iPhone 17 will drive record Apple shipments in 2025, IDC says

Published

on

By

iPhone 17 will drive record Apple shipments in 2025, IDC says

Apple’s latest iPhone models are shown on display at its Regent Street, London store on the launch day of the iPhone 17.

Arjun Kharpal | CNBC

Apple will hit a record level of iPhone shipments this year driven by its latest models and a resurgence in its key market of China, research firm IDC has forecast.

The company will ship 247.4 million iPhones in 2025, up just over 6% year-on-year, IDC forecast in a report on Tuesday. That’s more than the 236 million it sold in 2021, when the iPhone 13 was released.

Apple’s predicted surge is “thanks to the phenomenal success of its latest iPhone 17 series,” Nabila Popal, senior research director at IDC, said in a statement, adding that in China, “massive demand for iPhone 17 has significantly accelerated Apple’s performance.”

Shipments are a term used by analysts to refer to the number of devices sent by a vendor to its sales channels like e-commerce partners or stores. They do not directly equate to sales but indicate the demand expected by a company for their products.

When it launched in September, investors saw the iPhone 17 series as a key set of devices for Apple, which was facing increased competition in China and questions about its artificial intelligence strategy, as Android rivals were powering on.

Apple’s shipments are expected to jump 17% year-on-year in China in the fourth quarter, IDC said, leading the research firm to forecast 3% growth in the market this year versus a previous projection of a 1% decline.

In China, local players like Huawei have been taking away market share from Apple.

IDC’s report follows on from Counterpoint Research last week which forecast Apple to ship more smartphones than Samsung in 2025 for the first time in 14 years.

Bloomberg reported last month that Apple could delay the release of the base model of its next device, the iPhone 18, until 2027, which would break its regular cycle of releasing all of its phones in fall each year. IDC said this could mean Apple’s shipments may drop by 4.2% next year.

Continue Reading

Technology

Anthropic reportedly preparing for one of the largest IPOs ever in race with OpenAI: FT

Published

on

By

Anthropic reportedly preparing for one of the largest IPOs ever in race with OpenAI: FT

Nurphoto | Getty Images

Anthropic, the AI startup behind the popular Claude chatbot, is in early talks to launch one of the largest initial public offerings as early as next year, the Financial Times reported Wednesday. 

For the potential IPO, Anthropic has engaged law firm Wilson Sonsini Goodrich & Rosati, which has previously worked on high-profile tech IPOs such as Google, LinkedIn and Lyft, the FT said, citing two sources familiar with the matter.

The start-up, led by chief executive Dario Amodei, was also pursuing a private funding round that could value it above $300 billion, including a $15 billion combined commitment from Microsoft and Nvidia, per the report. 

It added that Anthropic has also discussed a potential IPO with major investment banks, but that sources characterized the discussions as preliminary and informal. 

If true, the news could position Anthropic in a race to market with rival ChatGPT-maker OpenAI, which is also reportedly laying the groundwork for a public offering. The potential listings would also test investors’ appetite for loss-making AI startups amid growing fears of a so-called AI bubble. 

However, an Anthropic spokesperson told the FT: “It’s fairly standard practice for companies operating at our scale and revenue level to effectively operate as if they are publicly traded companies,” adding that no decisions have been made on timing or whether to go public.

CNBC was unable to reach Anthropic and Wilson Sonsini, which has advised Anthropic for a few years, for comment. 

According to one of the FT’s sources, Anthropic has been working through internal preparations for a potential listing, though details were not provided. 

The FT report follows several notable changes at the company of late, including the hiring of former Airbnb executive Krishna Rao, who played a key role in the firm’s 2020 IPO.

CNBC also reported last month that Anthropic was recently valued to the range of $350 billion after receiving investments of up to $5 billion from Microsoft and $10 billion from Nvidia. 

In its race to overtake OpenAI in the AI space, the startup has also been expanding aggressively, recently announcing a $50 billion AI infrastructure build-out with data centers in Texas and New York, and tripling its international workforce.

According to the FT report, investors in the company are enthusiastic about Anthropic’s potential IPO, which could see it “seize the initiative” from OpenAI.

While OpenAI has been rumoured to be considering an IPO, its chief financial officer recently said the company is not pursuing a near-term listing, even as it closed a $6.6 billion share sale at a $500 billion valuation in October.

Continue Reading

Technology

We’re raising our CrowdStrike price target following a beat and raise quarter

Published

on

By

We're raising our CrowdStrike price target following a beat and raise quarter

Continue Reading

Trending