Connect with us

Published

on

Inside a sprawling lab at Google headquarters in Mountain View, California, hundreds of server racks hum across several aisles, performing tasks far less ubiquitous than running the world’s dominant search engine or executing workloads for Google Cloud’s millions of customers.

Instead, they’re running tests on Google’s own microchips, called Tensor Processing Units, or TPUs.

Originally trained for internal workloads, Google’s TPUs have been available to cloud customers since 2018. In July, Apple revealed it uses TPUs to train AI models underpinning Apple Intelligence. Google also relies on TPUs to train and run its Gemini chatbot.

“The world sort of has this fundamental belief that all AI, large language models, are being trained on Nvidia, and of course Nvidia has the lion’s share of training volume. But Google took its own path here,” said Futurum Group CEO Daniel Newman. He’s been covering Google’s custom cloud chips since they launched in 2015.

Google was the first cloud provider to make custom AI chips. Three years later, Amazon Web Services announced its first cloud AI chip, Inferentia. Microsoft‘s first custom AI chip, Maia, wasn’t announced until the end of 2023. 

But being first in AI chips hasn’t translated to a top spot in the overall rat race of generative AI. Google’s faced criticism for botched product releases, and Gemini came out more than a year after OpenAI’s ChatGPT.

Google Cloud, however, has gained momentum due in part to AI offerings. Google parent company Alphabet reported cloud revenue rose 29% in the most recent quarter, surpassing $10 billion in quarterly revenues for the first time.

“The AI cloud era has completely reordered the way companies are seen, and this silicon differentiation, the TPU itself, may be one of the biggest reasons that Google went from the third cloud to being seen truly on parity, and in some eyes, maybe even ahead of the other two clouds for its AI prowess,” Newman said.

‘A simple but powerful thought experiment’

In July, CNBC got the first on-camera tour of Google’s chip lab and sat down with the head of custom cloud chips, Amin Vahdat. He was already at Google when it first toyed with the idea of making chips in 2014. 

Amin Vahdat, VP of Machine Learning, Systems and Cloud AI at Google, holds up TPU Version 4 at Google headquarters in Mountain View, California, on July 23, 2024.

Marc Ganley

“It all started with a simple but powerful thought experiment,” Vahdat said. “A number of leads at the company asked the question: What would happen if Google users wanted to interact with Google via voice for just 30 seconds a day? And how much compute power would we need to support our users?”

The group determined Google would need to double the number of computers in its data centers. So they looked for a better solution.

“We realized that we could build custom hardware, not general purpose hardware, but custom hardware — Tensor Processing Units in this case — to support that much, much more efficiently. In fact, a factor of 100 more efficiently than it would have been otherwise,” Vahdat said.

Google data centers still rely on general-purpose central processing units, or CPUs, and Nvidia’s graphics processing units, or GPUs. Google’s TPUs are a different type of chip called an application-specific integrated circuit, or ASIC, which are custom-built for specific purposes. The TPU is focused on AI. Google makes another ASIC focused on video called a Video Coding Unit. 

Google also makes custom chips for its devices, similar to Apple’s custom silicon strategy. The Tensor G4 powers Google’s new AI-enabled Pixel 9, and its new A1 chip powers Pixel Buds Pro 2. 

The TPU, however, is what set Google apart. It was the first of its kind when it launched in 2015. Google TPUs still dominate among custom cloud AI accelerators, with 58% of the market share, according to The Futurum Group.

Google coined the term based on the algebraic term “tensor,” referring to the large-scale matrix multiplications that happen rapidly for advanced AI applications.

With the second TPU release in 2018, Google expanded the focus from inference to training and made them available for its cloud customers to run workloads, alongside market-leading chips such as Nvidia’s GPUs.

“If you’re using GPUs, they’re more programmable, they’re more flexible. But they’ve been in tight supply,” said Stacy Rasgon, senior analyst covering semiconductors at Bernstein Research.

The AI boom has sent Nvidia’s stock through the roof, catapulting the chipmaker to a $3 trillion market cap in June, surpassing Alphabet and jockeying with Apple and Microsoft for position as the world’s most valuable public company.

“Being candid, these specialty AI accelerators aren’t nearly as flexible or as powerful as Nvidia’s platform, and that is what the market is also waiting to see: Can anyone play in that space?” Newman said.

Now that we know Apple’s using Google’s TPUs to train its AI models, the real test will come as those full AI features roll out on iPhones and Macs next year.

Broadcom and TSMC

It’s no small feat to develop alternatives to Nvidia’s AI engines. Google’s sixth generation TPU, called Trillium, is set to come out later this year.

Google showed CNBC the sixth version of its TPU, Trillium, in Mountain View, California, on July 23, 2024. Trillium is set to come out later in 2024.

Marc Ganley

“It’s expensive. You need a lot of scale,” Rasgon said. “And so it’s not something that everybody can do. But these hyperscalers, they’ve got the scale and the money and the resources to go down that path.”

The process is so complex and costly that even the hyperscalers can’t do it alone. Since the first TPU, Google’s partnered with Broadcom, a chip developer that also helps Meta design its AI chips. Broadcom says it’s spent more than $3 billion to make these partnerships happen.  

“AI chips — they’re very complex. There’s lots of things on there. So Google brings the compute,” Rasgon said. “Broadcom does all the peripheral stuff. They do the I/O and the SerDes, all of the different pieces that go around that compute. They also do the packaging.”

Then the final design is sent off for manufacturing at a fabrication plant, or fab — primarily those owned by the world’s largest chipmaker, Taiwan Semiconductor Manufacturing Company, which makes 92% of the world’s most advanced semiconductors.

When asked if Google has any safeguards in place should the worst happen in the geopolitical sphere between China and Taiwan, Vahdat said, “It’s certainly something that we prepare for and we think about as well, but we’re hopeful that actually it’s not something that we’re going to have to trigger.”

Protecting against those risks is the primary reason the White House is handing out $52 billion in CHIPS Act funding to companies building fabs in the U.S. — with the biggest portions going to Intel, TSMC, and Samsung to date.

Processors and power

Google showed CNBC its new Axion CPU,

Marc Ganley

“Now we’re able to bring in that last piece of the puzzle, the CPU,” Vahdat said. “And so a lot of our internal services, whether it’s BigQuery, whether it’s Spanner, YouTube advertising and more are running on Axion.”

Google is late to the CPU game. Amazon launched its Graviton processor in 2018. Alibaba launched its server chip in 2021. Microsoft announced its CPU in November.

When asked why Google didn’t make a CPU sooner, Vahdat said, “Our focus has been on where we can deliver the most value for our customers, and there it has been starting with the TPU, our video coding units, our networking. We really thought that the time was now.”

All these processors from non-chipmakers, including Google’s, are made possible by Arm chip architecture — a more customizable, power-efficient alternative that’s gaining traction over the traditional x86 model from Intel and AMD. Power efficiency is crucial because, by 2027, AI servers are projected to use up as much power every year as a country like Argentina. Google’s latest environmental report showed emissions rose nearly 50% from 2019 to 2023 partly due to data center growth for powering AI.

“Without having the efficiency of these chips, the numbers could have wound up in a very different place,” Vahdat said. “We remain committed to actually driving these numbers in terms of carbon emissions from our infrastructure, 24/7, driving it toward zero.”

It takes a massive amount of water to cool the servers that train and run AI. That’s why Google’s third-generation TPU started using direct-to-chip cooling, which uses far less water. That’s also how Nvidia’s cooling its latest Blackwell GPUs.

Despite challenges, from geopolitics to power and water, Google is committed to its generative AI tools and making its own chips. 

“I’ve never seen anything like this and no sign of it slowing down quite yet,” Vahdat said. “And hardware is going to play a really important part there.”

Continue Reading

Technology

Amazon was questioned by House China committee over ‘dangerous and unwise’ TikTok partnership

Published

on

By

Amazon was questioned by House China committee over 'dangerous and unwise' TikTok partnership

Amazon logo on a brick building exterior, San Francisco, California, August 20, 2024.

Smith Collection | Gado | Archive Photos | Getty Images

Amazon representatives met with the House China committee in recent months to discuss lawmaker concerns over the company’s partnership with TikTok, CNBC confirmed.

A spokesperson for the House Select Committee on the Chinese Communist Party confirmed the meeting, which centered on a shopping deal between Amazon and TikTok announced in August. The agreement allows users of TikTok, owned by China’s ByteDance, to link their account with Amazon and make purchases from the site without leaving TikTok.

“The Select Committee conveyed to Amazon that it is dangerous and unwise for Amazon to partner with TikTok given the grave national security threat the app poses,” the spokesperson said. The parties met in September, according to Bloomberg, which first reported the news.

Representatives from Amazon and TikTok did not immediately respond to CNBC’s request for comment.

TikTok’s future viability in the U.S. is uncertain. In April, President Joe Biden signed a law that requires ByteDance to sell TikTok by Jan. 19. If TikTok fails to cut ties with its parent company, app stores and internet hosting services would be prohibited from offering the app.

President-elect Donald Trump could rescue TikTok from a potential U.S. ban. He promised on the campaign trail that he would “save” TikTok, and said in a March interview with CNBC’s “Squawk Box” that “there’s a lot of good and there’s a lot of bad” with the app.

In his first administration, Trump had tried to implement a TikTok ban. He changed his stance around the time he met with billionaire Jeff Yass. The Republican megadonor’s trading firm, Susquehanna International Group, owns a 15% stake in ByteDance, while Yass has a 7% stake in the company, NBC and CNBC reported in March.

— CNBC’s Jonathan Vanian contributed to this report.

Don’t miss these insights from CNBC PRO

TikTok is 'digital nicotine' for young people, says D.C. Attorney General Brian Schwalb

Continue Reading

Technology

Amazon launches fixed pricing for treatment of conditions such as hair loss. Hims & Hers stock drops 15%

Published

on

By

Amazon launches fixed pricing for treatment of conditions such as hair loss. Hims & Hers stock drops 15%

A worker delivers Amazon packages in San Francisco on Oct. 24, 2024.

David Paul Morris | Bloomberg | Getty Images

Amazon on Thursday announced Prime members can access new fixed pricing for treatment of conditions like erectile dysfunction and men’s hair loss, its latest effort to compete with other direct-to-consumer marketplaces such as Hims & Hers Health and Ro.

Shares of Hims & Hers fell as much as 17% on Thursday, on pace for its worst day.

Amazon said in a blog post that Prime members can see the cost of a telehealth visit and their desired treatment before they decide to proceed with care for five common issues. Patients can access treatment for anti-aging skin care starting at $10 a month; motion sickness for $2 per use; erectile dysfunction at $19 a month; eyelash growth at $43 a month, and men’s hair loss for $16 a month by using Amazon’s savings benefit Prime Rx at checkout.

Amazon acquired primary care provider One Medical for roughly $3.9 billion in July 2022, and Thursday’s announcement builds on its existing pay-per-visit telehealth offering. Video visits through the service cost $49, and messaging visits cost $29 where available. Users can get treatment for more than 30 common conditions, including sinus infection and pink eye.

Medications filled through Amazon Pharmacy are eligible for discounted pricing and will be delivered to patients’ doors in standard Amazon packaging. Prime members will pay for the consultation and medication, but there are no additional fees, the blog post said.

Amazon has been trying to break into the lucrative health-care sector for years. The company launched its own online pharmacy in 2020 following its acquisition of PillPack in 2018. Amazon introduced, and later shuttered, a telehealth service called Amazon Care, as well as a line of health and wellness devices.

The company has also discontinued a secretive effort to develop an at-home fertility tracker, CNBC reported Wednesday.

— CNBC’s Annie Palmer contributed to this report.

Continue Reading

Technology

WikiLeaks whistleblower Chelsea Manning says censorship is still ‘a dominant threat’

Published

on

By

WikiLeaks whistleblower Chelsea Manning says censorship is still 'a dominant threat'

Chelsea Manning: Censorship still a dominant threat

Former U.S. Army intelligence analyst Chelsea Manning says censorship is still “a dominant threat,” advocating for a more decentralized internet to help better protect individuals online.

Her comments come amid ongoing tension linked to online safety rules, with some tech executives recently seeking to push back over content moderation concerns.

Speaking to CNBC’s Karen Tso at the Web Summit tech conference in Lisbon, Portugal, on Wednesday, Manning said that one way to ensure online privacy could be “decentralized identification,” which gives individuals the ability to control their own data.

“Censorship is a dominant threat. I think that it is a question of who’s doing the censoring, and what the purpose is — and also censorship in the 21st century is more about whether or not you’re boosted through like an algorithm, and how the fine-tuning of that seems to work,” Manning said.

“I think that social media and the monopolies of social media have sort of gotten us used to the fact that certain things that drive engagement will be attractive,” she added.

“One of the ways that we can sort of countervail that is to go back to the more decentralized and distribute the internet of the early ’90s, but make that available to more people.”

Nym Technologies Chief Security Officer Chelsea Manning at a press conference held with Nym Technologies CEO Harry Halpin in the Media Village to present NymVPN during the second day of Web Summit on November 13, 2024 in Lisbon, Portugal. 

Horacio Villalobos | Getty Images News | Getty Images

Asked how tech companies could make money in such a scenario, Manning said there would have to be “a better social contract” put in place to determine how information is shared and accessed.

“One of the things about distributed or decentralized identification is that through encryption you’re able to sort of check the box yourself, instead of having to depend on the company to provide you with a check box or an accept here, you’re making that decision from a technical perspective,” Manning said.

‘No longer secrecy versus transparency’

Manning, who works as a security consultant at Nym Technologies, a company that specializes in online privacy and security, was convicted of espionage and other charges at a court-martial in 2013 for leaking a trove of secret military files to online media publisher WikiLeaks.

She was sentenced to 35 years in prison, but was later released in 2017, when former U.S. President Barack Obama commuted her sentence.

Asked to what extent the environment has changed for whistleblowers today, Manning said, “We’re at an interesting time because information is everywhere. We have more information than ever.”

She added, “Countries and governments no longer seem to invest the same amount of time and effort in hiding information and keeping secrets. What countries seem to be doing now is they seem to be spending more time and energy spreading misinformation and disinformation.”

Manning said the challenge for whistleblowers now is to sort through the information to understand what is verifiable and authentic.

“It’s no longer secrecy versus transparency,” she added.

Continue Reading

Trending