Connect with us

Published

on

Inside a sprawling lab at Google headquarters in Mountain View, California, hundreds of server racks hum across several aisles, performing tasks far less ubiquitous than running the world’s dominant search engine or executing workloads for Google Cloud’s millions of customers.

Instead, they’re running tests on Google’s own microchips, called Tensor Processing Units, or TPUs.

Originally trained for internal workloads, Google’s TPUs have been available to cloud customers since 2018. In July, Apple revealed it uses TPUs to train AI models underpinning Apple Intelligence. Google also relies on TPUs to train and run its Gemini chatbot.

“The world sort of has this fundamental belief that all AI, large language models, are being trained on Nvidia, and of course Nvidia has the lion’s share of training volume. But Google took its own path here,” said Futurum Group CEO Daniel Newman. He’s been covering Google’s custom cloud chips since they launched in 2015.

Google was the first cloud provider to make custom AI chips. Three years later, Amazon Web Services announced its first cloud AI chip, Inferentia. Microsoft‘s first custom AI chip, Maia, wasn’t announced until the end of 2023. 

But being first in AI chips hasn’t translated to a top spot in the overall rat race of generative AI. Google’s faced criticism for botched product releases, and Gemini came out more than a year after OpenAI’s ChatGPT.

Google Cloud, however, has gained momentum due in part to AI offerings. Google parent company Alphabet reported cloud revenue rose 29% in the most recent quarter, surpassing $10 billion in quarterly revenues for the first time.

“The AI cloud era has completely reordered the way companies are seen, and this silicon differentiation, the TPU itself, may be one of the biggest reasons that Google went from the third cloud to being seen truly on parity, and in some eyes, maybe even ahead of the other two clouds for its AI prowess,” Newman said.

‘A simple but powerful thought experiment’

In July, CNBC got the first on-camera tour of Google’s chip lab and sat down with the head of custom cloud chips, Amin Vahdat. He was already at Google when it first toyed with the idea of making chips in 2014. 

Amin Vahdat, VP of Machine Learning, Systems and Cloud AI at Google, holds up TPU Version 4 at Google headquarters in Mountain View, California, on July 23, 2024.

Marc Ganley

“It all started with a simple but powerful thought experiment,” Vahdat said. “A number of leads at the company asked the question: What would happen if Google users wanted to interact with Google via voice for just 30 seconds a day? And how much compute power would we need to support our users?”

The group determined Google would need to double the number of computers in its data centers. So they looked for a better solution.

“We realized that we could build custom hardware, not general purpose hardware, but custom hardware — Tensor Processing Units in this case — to support that much, much more efficiently. In fact, a factor of 100 more efficiently than it would have been otherwise,” Vahdat said.

Google data centers still rely on general-purpose central processing units, or CPUs, and Nvidia’s graphics processing units, or GPUs. Google’s TPUs are a different type of chip called an application-specific integrated circuit, or ASIC, which are custom-built for specific purposes. The TPU is focused on AI. Google makes another ASIC focused on video called a Video Coding Unit. 

Google also makes custom chips for its devices, similar to Apple’s custom silicon strategy. The Tensor G4 powers Google’s new AI-enabled Pixel 9, and its new A1 chip powers Pixel Buds Pro 2. 

The TPU, however, is what set Google apart. It was the first of its kind when it launched in 2015. Google TPUs still dominate among custom cloud AI accelerators, with 58% of the market share, according to The Futurum Group.

Google coined the term based on the algebraic term “tensor,” referring to the large-scale matrix multiplications that happen rapidly for advanced AI applications.

With the second TPU release in 2018, Google expanded the focus from inference to training and made them available for its cloud customers to run workloads, alongside market-leading chips such as Nvidia’s GPUs.

“If you’re using GPUs, they’re more programmable, they’re more flexible. But they’ve been in tight supply,” said Stacy Rasgon, senior analyst covering semiconductors at Bernstein Research.

The AI boom has sent Nvidia’s stock through the roof, catapulting the chipmaker to a $3 trillion market cap in June, surpassing Alphabet and jockeying with Apple and Microsoft for position as the world’s most valuable public company.

“Being candid, these specialty AI accelerators aren’t nearly as flexible or as powerful as Nvidia’s platform, and that is what the market is also waiting to see: Can anyone play in that space?” Newman said.

Now that we know Apple’s using Google’s TPUs to train its AI models, the real test will come as those full AI features roll out on iPhones and Macs next year.

Broadcom and TSMC

It’s no small feat to develop alternatives to Nvidia’s AI engines. Google’s sixth generation TPU, called Trillium, is set to come out later this year.

Google showed CNBC the sixth version of its TPU, Trillium, in Mountain View, California, on July 23, 2024. Trillium is set to come out later in 2024.

Marc Ganley

“It’s expensive. You need a lot of scale,” Rasgon said. “And so it’s not something that everybody can do. But these hyperscalers, they’ve got the scale and the money and the resources to go down that path.”

The process is so complex and costly that even the hyperscalers can’t do it alone. Since the first TPU, Google’s partnered with Broadcom, a chip developer that also helps Meta design its AI chips. Broadcom says it’s spent more than $3 billion to make these partnerships happen.  

“AI chips — they’re very complex. There’s lots of things on there. So Google brings the compute,” Rasgon said. “Broadcom does all the peripheral stuff. They do the I/O and the SerDes, all of the different pieces that go around that compute. They also do the packaging.”

Then the final design is sent off for manufacturing at a fabrication plant, or fab — primarily those owned by the world’s largest chipmaker, Taiwan Semiconductor Manufacturing Company, which makes 92% of the world’s most advanced semiconductors.

When asked if Google has any safeguards in place should the worst happen in the geopolitical sphere between China and Taiwan, Vahdat said, “It’s certainly something that we prepare for and we think about as well, but we’re hopeful that actually it’s not something that we’re going to have to trigger.”

Protecting against those risks is the primary reason the White House is handing out $52 billion in CHIPS Act funding to companies building fabs in the U.S. — with the biggest portions going to Intel, TSMC, and Samsung to date.

Processors and power

Google showed CNBC its new Axion CPU,

Marc Ganley

“Now we’re able to bring in that last piece of the puzzle, the CPU,” Vahdat said. “And so a lot of our internal services, whether it’s BigQuery, whether it’s Spanner, YouTube advertising and more are running on Axion.”

Google is late to the CPU game. Amazon launched its Graviton processor in 2018. Alibaba launched its server chip in 2021. Microsoft announced its CPU in November.

When asked why Google didn’t make a CPU sooner, Vahdat said, “Our focus has been on where we can deliver the most value for our customers, and there it has been starting with the TPU, our video coding units, our networking. We really thought that the time was now.”

All these processors from non-chipmakers, including Google’s, are made possible by Arm chip architecture — a more customizable, power-efficient alternative that’s gaining traction over the traditional x86 model from Intel and AMD. Power efficiency is crucial because, by 2027, AI servers are projected to use up as much power every year as a country like Argentina. Google’s latest environmental report showed emissions rose nearly 50% from 2019 to 2023 partly due to data center growth for powering AI.

“Without having the efficiency of these chips, the numbers could have wound up in a very different place,” Vahdat said. “We remain committed to actually driving these numbers in terms of carbon emissions from our infrastructure, 24/7, driving it toward zero.”

It takes a massive amount of water to cool the servers that train and run AI. That’s why Google’s third-generation TPU started using direct-to-chip cooling, which uses far less water. That’s also how Nvidia’s cooling its latest Blackwell GPUs.

Despite challenges, from geopolitics to power and water, Google is committed to its generative AI tools and making its own chips. 

“I’ve never seen anything like this and no sign of it slowing down quite yet,” Vahdat said. “And hardware is going to play a really important part there.”

Continue Reading

Technology

Nvidia positioned to weather Trump tariffs, chip demand ‘off the charts,’ says Altimeter’s Gerstner

Published

on

By

Nvidia positioned to weather Trump tariffs, chip demand 'off the charts,' says Altimeter's Gerstner

Altimeter CEO Brad Gerstner is buying Nvidia

Altimeter Capital CEO Brad Gerstner said Thursday that he’s moving out of the “bomb shelter” with Nvidia and into a position of safety, expecting that the chipmaker is positioned to withstand President Donald Trump’s widespread tariffs.

“The growth and the demand for GPUs is off the charts,” he told CNBC’s “Fast Money Halftime Report,” referring to Nvidia’s graphics processing units that are powering the artificial intelligence boom. He said investors just need to listen to commentary from OpenAI, Google and Elon Musk.

President Trump announced an expansive and aggressive “reciprocal tariff” policy in a ceremony at the White House on Wednesday. The plan established a 10% baseline tariff, though many countries like China, Vietnam and Taiwan are subject to steeper rates. The announcement sent stocks tumbling on Thursday, with the tech-heavy Nasdaq down more than 5%, headed for its worst day since 2022.

The big reason Nvidia may be better positioned to withstand Trump’s tariff hikes is because semiconductors are on the list of exceptions, which Gerstner called a “wise exception” due to the importance of AI.

Nvidia’s business has exploded since the release of OpenAI’s ChatGPT in 2022, and annual revenue has more than doubled in each of the past two fiscal years. After a massive rally, Nvidia’s stock price has dropped by more than 20% this year and was down almost 7% on Thursday.

Gerstner is concerned about the potential of a recession due to the tariffs, but is relatively bullish on Nvidia, and said the “negative impact from tariffs will be much less than in other areas.”

He said it’s key for the U.S. to stay competitive in AI. And while the company’s chips are designed domestically, they’re manufactured in Taiwan “because they can’t be fabricated in the U.S.” Higher tariffs would punish companies like Meta and Microsoft, he said.

“We’re in a global race in AI,” Gerstner said. “We can’t hamper our ability to win that race.”

WATCH: Brad Gerstner is buying Nvidia

Continue Reading

Technology

YouTube announces Shorts editing features amid potential TikTok ban

Published

on

By

YouTube announces Shorts editing features amid potential TikTok ban

Jaque Silva | Nurphoto | Getty Images

YouTube on Thursday announced new video creation tools for Shorts, its short-form video feed that competes against TikTok. 

The features come at a time when TikTok, which is owned by Chinese company ByteDance, is at risk of an effective ban in the U.S. if it’s not sold to an American owner by April 5.

Among the new tools is an updated video editor that allows creators to make precise adjustments and edits, a feature that automatically syncs video cuts to the beat of a song and AI stickers.

The creator tools will become available later this spring, said YouTube, which is owned by Google

Along with the new features, YouTube last week said it was changing the way view counts are tabulated on Shorts. Under the new guidelines, Shorts views will count the number of times the video is played or replayed with no minimum watch time requirement. 

Previously, views were only counted if a video was played for a certain number of seconds. This new tabulation method is similar to how views are counted on TikTok and Meta’s Reels, and will likely inflate view counts.

“We got this feedback from creators that this is what they wanted. It’s a way for them to better understand when their Shorts have been seen,” YouTube Chief Product Officer Johanna Voolich said in a YouTube video. “It’s useful for creators who post across multiple platforms.”

WATCH: TikTok is a digital Trojan horse, says Hayman Capital’s Kyle Bass

TikTok is a digital Trojan horse, says Hayman Capital's Kyle Bass

Continue Reading

Technology

Tech stocks sink after Trump tariff rollout — Apple heads for worst drop in 5 years

Published

on

By

Tech stocks sink after Trump tariff rollout — Apple heads for worst drop in 5 years

CEO of Meta and Facebook Mark Zuckerberg, Lauren Sanchez, Amazon founder Jeff Bezos, Google CEO Sundar Pichai, and Tesla and SpaceX CEO Elon Musk attend the inauguration ceremony before Donald Trump is sworn in as the 47th U.S. president in the U.S. Capitol Rotunda in Washington, Jan. 20, 2025.

Saul Loeb | Via Reuters

Technology stocks plummeted Thursday after President Donald Trump’s new tariff policies sparked widespread market panic.

Apple led the declines among the so-called “Magnificent Seven” group, dropping nearly 9%. The iPhone maker makes its devices in China and other Asian countries. The stock is on pace for its steepest drop since 2020.

Other megacaps also felt the pressure. Meta Platforms and Amazon fell more than 7% each, while Nvidia and Tesla slumped more than 5%. Nvidia builds its new chips in Taiwan and relies on Mexico for assembling its artificial intelligence systems. Microsoft and Alphabet both fell about 2%.

Semiconductor stocks also felt the pain, with Marvell Technology, Arm Holdings and Micron Technology falling more than 8% each. Broadcom and Lam Research dropped 6%, while Advanced Micro Devices declined more than 4% Software stocks ServiceNow and Fortinet fell more than 5% each.

Read more CNBC tech news

The drop in technology stocks came amid a broader market selloff spurred by fears of a global trade war after Trump unveiled a blanket 10% tariff on all imported goods and a range of higher duties targeting specific countries after the bell Wednesday. He said the new tariffs would be a “declaration of economic independence” for the U.S.

Companies and countries worldwide have already begun responding to the wide-sweeping policy, which included a 34% tariff on China stacked on a previous 20% tax, a 46% duty on Vietnam and a 20% levy on imports from the European Union.

China’s Ministry of Commerce urged the U.S. to “immediately cancel” the unilateral tariff measures and said it would take “resolute counter-measures.”

The tariffs come on the heels of a rough quarter for the tech-heavy Nasdaq and the worst period for the index since 2022. Stocks across the board have come under pressure over concerns of a weakening U.S. economy. The Nasdaq Composite dropped nearly 5% on Thursday, bringing its year-to-date loss to 13%.

Trump applauded some megacap technology companies for investing money into the U.S. during his speech, calling attention to Apple’s plan to spend $500 billion over the next four years.

Evercore ISI's Amit Daryanani on keeping Apple's outperform rating despite tariffs

Continue Reading

Trending