Elon Musk recently walked back his impossible extremely ambitious robotaxi goals, shifting from a target of reaching “half the U.S. population by the end of 2025” to a more “modest” goal of launching in “eight to 10 U.S. metro areas” within the next two months.
Now, in a development that should surprise no one, a new report suggests that even this heavily scaled-back timeline is facing significant obstacles.
According to a report from The Information, Tesla is lagging on the most basic regulatory front. The company has reportedly not yet completed the necessary paperwork to begin offering robotaxi rides in Arizona and Nevada, two of the three additional states Musk has targeted for expansion by the end of 2025.
The third state, Florida, is expected to be an easier lift due to its looser regulations, which fits Tesla’s pattern of prioritizing optics over navigating real regulatory scrutiny.
Advertisement – scroll for more content
Meanwhile, in Tesla’s original home state of California, the company is still testing its service with a human safety driver sitting in the driver’s seat. This is because its current permit only allows a human to “drive a traditional vehicle.” Therefore, the “robotaxi” service in California is simply a ride-hailing service with the Tesla drivers using “Supervised Full Self-Driving.”
To operate a true autonomous service, with or without a safety driver present, Tesla would need to apply for a separate permit—one that its chief rival, Alphabet’s Waymo, already possesses and uses to offer rides in San Francisco and Los Angeles. As we’ve noted previously, Tesla has tellingly not even applied for this autonomous vehicle permit, which would likely require it to disclose critical disengagement and safety data it is unwilling to make public.
These regulatory and bureaucratic slowdowns clash with the grand vision Musk is selling. The CEO has stated that the success of Robotaxi is “critical” to his plan to pivot Tesla into an “autonomous driving and humanoid robotics company.”
This vision is also directly tied to his unprecedented new compensation package, which shareholders are set to vote on next week. That package hinges on massive goals, including putting 1 million Robotaxis into service and lifting Tesla’s market capitalization to an astronomical $8.5 trillion.
For now, the robotaxi service continues to use a version of the Model Y. The purpose-built “Cybercab,” a two-seater vehicle with no steering wheel, isn’t planned for mass production until the second quarter of 2026.
Tesla threw cold water on that program too this week as it said that it might add a steering wheel to the vehicle, which would facilitate the regulatory approval.
Electrek’s Take
This is predictable, but still frustrating. We’ve been saying for years that the technology is only half the battle; it’s far from solved. The other half is the massive, state-by-state, and even city-by-city, regulatory grind.
Tesla still has a lot of work to do to make the technology safe enough to remove its safety monitor without negatively affecting safety.
On the other front, it’s one thing for Musk to set an “Elon time” goal, but it’s another to seemingly ignore the basic bureaucratic legwork required to operate in new states. To hear that Tesla hasn’t even filed the paperwork in places like Arizona and Nevada is a significant failure. It suggests the bottleneck isn’t just performance.
This whole endeavor continues to look like a dangerous game of smoke and mirrors, prioritizing optics to justify a compensation package and stock valuation that are completely detached from the reality of the technology or the regulatory hurdles ahead.
In short, Tesla either doesn’t really believe it is ready to scale Robotaxi, unlike what it has been claiming to shareholders, or it doesn’t want to release critical data to regulators, which would suggest the same thing.
In the meantime, they will deploy ride-hailing services with drivers and call it “Robotaxi”, like they do in the Bay Area.
FTC: We use income earning auto affiliate links.More.
Pentagon-backed MP Materials warned investors this week to approach other rare earths projects with caution, pointing to the industry’s difficult economics.
Stocks of U.S. rare earth companies have had wild swings in recent months as investors have speculated that the Trump administration might strike more deals along the lines of its landmark agreement with MP. Smaller retail traders have gotten involved in the stocks with the VanEck Rare Earth and Strategic Metals ETF up 60% this year.
The Defense Department in July took an equity stake in MP, set a price floor for the company, and inked an offtake agreement with the rare earth miner and magnet maker in an effort to roll back China’s dominance of the industry.
CEO James Litinsky said he didn’t want “people to get burned” amid the speculation. Litinsky cautioned investors “to just be very clear-eyed about what the actual structural economics are amidst all the excitement.”
“The vast majority of projects being promoted today simply will not work at virtually any price,” Litinksy said on the company’s third-quarter earnings call Thursday evening.
Stock Chart IconStock chart icon
VanEck Rare Earth and Strategic Metals ETF, YTD
MP views itself as “America’s national champion,” Litinsky said. MP is the only active rare earth miner in the U.S. and has offtake agreements with Apple and General Motors in addition to the Pentagon.
“We have structural advantage because we’re fully vertically integrated,” the CEO said. “We’re years and billions ahead of others.”
It takes years for the best rare earth producers to ramp up and stabilize their output and economics “despite what some promoters might suggest,” Litinksy said. Australia’s Lynas took about a decade and MP will reach normalized production in about three years from the start of commissioning, he said.
The White House is “not ruling out other deals with equity stakes or price floors as we did with MP Materials, but that doesn’t mean every initiative we take would be in the shape of the MP deal,” a Trump administration official told CNBC in September.
Litinsky described the rare earth industry as close to a “structural oligopoly,” a system where there are just a few major players. The government investing in a dozens of sites and businesses wouldn’t necessarily set up a supply chain, he said.
The Trump administration should continue to encourage private capital to flow into the industry through loans, grants and other support, Litinsky said. There is room for “a lot of other players and supply” but the market will require “materially higher prices” for the industry’s structural challenges to change, he said.
“If X dollars of capital can stimulate two or three X in private capital, they should be doing that as much as possible,” Litinsky said.
The CEO indicated that he views MP as a forerunner that will help create the conditions for a broader market that is not dependent on China over time.
“In the very short term the administration has made sure that we have a successful national champion in MP,” Litinsky said. “We are going to sort of pave the path if you will to then figure out how there’s much broader supply coming online.”
Rare earths are crucial for making magnets that are key inputs in U.S. weapons platforms, semiconductor manufacturing, electric vehicles, clean energy technology and consumer electronics. Beijing dominates the global supply chain and the U.S. is dependent on China for imports.
This week on Electrek’s Wheel-E podcast, we discuss the most popular news stories from the world of electric bikes and other nontraditional electric vehicles. This time, that includes a new e-bike model from Tenways, California kills off its e-bike voucher program, a review of the new VMAX VX2 Hub e-scooter, Zero launches a scooter, NIU’s got a new micro-car, and more.
The Wheel-E podcast returns every two weeks on Electrek’s YouTube channel, Facebook, Linkedin, and Twitter.
As a reminder, we’ll have an accompanying post, like this one, on the site with an embedded link to the live stream. Head to the YouTube channel to get your questions and comments in.
After the show ends, the video will be archived on YouTube and the audio on all your favorite podcast apps:
We also have a Patreon if you want to help us to avoid more ads and invest more in our content. We have some awesome gifts for our Patreons and more coming.
Here are a few of the articles that we will discuss during the Wheel-E podcast today:
Here’s the live stream for today’s episode starting at 9:00 a.m. ET (or the video after 10:00 a.m. ET):
FTC: We use income earning auto affiliate links.More.
Nvidia has established itself as the undisputed leader in artificial intelligence chips, selling large quantities of silicon to most of the world’s biggest tech companies en route to a $4.5 trillion market cap.
One of Nvidia’s key clients is Google, which has been loading up on the chipmaker’s graphics processing units, or GPUs, to try and keep pace with soaring demand for AI compute power in the cloud.
While there’s no sign that Google will be slowing its purchases of Nvidia GPUs, the internet giant is increasingly showing that it’s not just a buyer of high-powered silicon. It’s also a developer.
On Thursday, Google announced that its most powerful chip yet, called Ironwood, is being made widely available in the coming weeks. It’s the seventh generation of Google’s Tensor Processing Unit, or TPU, the company’s custom silicon that’s been in the works for more than a decade.
TPUs are application-specific integrated circuits, or ASICs, which play a crucial role in AI by providing highly specialized and efficient hardware for particular tasks. Google says Ironwood is designed to handle the heaviest AI workloads, from training large models to powering real-time chatbots and AI agents, and is more than four times faster than its predecessor. AI startup Anthropic plans to use up to 1 million of them to run its Claude model.
For Google, TPUs offer a competitive edge at a time when all the hyperscalers are rushing to build mammoth data centers, and AI processors can’t get manufactured fast enough to meet demand. Other cloud companies are taking a similar approach, but are well behind in their efforts.
Amazon Web Services made its first cloud AI chip, Inferentia, available to customers in 2019, followed by Trainium three years later. Microsoft didn’t announce its first custom AI chip, Maia, until the end of 2023.
“Of the ASIC players, Google’s the only one that’s really deployed this stuff in huge volumes,” said Stacy Rasgon, an analyst covering semiconductors at Bernstein. “For other big players, it takes a long time and a lot of effort and a lot of money. They’re the furthest along among the other hyperscalers.”
Google didn’t provide a comment for this story.
Originally trained for internal workloads, Google’s TPUs have been available to cloud customers since 2018. Of late, Nvidia has shown some level of concern. When OpenAI signed its first cloud contract with Google earlier this year, the announcement spurred Nvidia CEO Jensen Huang to initiate further talks with the AI startup and its CEO, Sam Altman, according to reporting by The Wall Street Journal.
Unlike Nvidia, Google isn’t selling its chips as hardware, but rather providing access to TPUs as a service through its cloud, which has emerged as one of the company’s big growth drivers. In its third-quarter earnings report last week, Google parent Alphabet said cloud revenue increased 34% from a year earlier to $15.15 billion, beating analyst estimates. The company ended the quarter with a business backlog of $155 billion.
“We are seeing substantial demand for our AI infrastructure products, including TPU-based and GPU-based solutions,” CEO Sundar Pichai said on the earnings call. “It is one of the key drivers of our growth over the past year, and I think on a going-forward basis, I think we continue to see very strong demand, and we are investing to meet that.”
Google doesn’t break out the size of its TPU business within its cloud segment. Analysts at D.A. Davidson estimated in September that a “standalone” business consisting of TPUs and Google’s DeepMind AI division could be valued at about $900 billion, up from an estimate of $717 billion in January. Alphabet’s current market cap is more than $3.4 trillion.
‘Tightly targeted’ chips
Customization is a major differentiator for Google. One critical advantage, analysts say, is the efficiency TPUs offer customers relative to competitive products and services.
“They’re really making chips that are very tightly targeted for their workloads that they expect to have,” said James Sanders, an analyst at Tech Insights.
Rasgon said that efficiency is going to become increasingly important because with all the infrastructure that’s being built, the “likely bottleneck probably isn’t chip supply, it’s probably power.”
On Tuesday, Google announced Project Suncatcher, which explores “how an interconnected network of solar-powered satellites, equipped with our Tensor Processing Unit (TPU) AI chips, could harness the full power of the Sun.”
As a part of the project, Google said it plans to launch two prototype solar-powered satellites carrying TPUs by early 2027.
“This approach would have tremendous potential for scale, and also minimizes impact on terrestrial resources,” the company said in the announcement. “That will test our hardware in orbit, laying the groundwork for a future era of massively-scaled computation in space.”
Dario Amodei, co-founder and chief executive officer of Anthropic, at the World Economic Forum in 2025.
Stefan Wermuth | Bloomberg | Getty Images
Google’s largest TPU deal on record landed late last month, when the company announced a massive expansion of its agreement with OpenAI rival Anthropic valued in the tens of billions of dollars. With the partnership, Google is expected to bring well over a gigawatt of AI compute capacity online in 2026.
“Anthropic’s choice to significantly expand its usage of TPUs reflects the strong price-performance and efficiency its teams have seen with TPUs for several years,” Google Cloud CEO Thomas Kurian said at the time of the announcement.
Google has invested $3 billion in Anthropic. And while Amazon remains Anthropic’s most deeply embedded cloud partner, Google is now providing the core infrastructure to support the next generation of Claude models.
“There is such demand for our models that I think the only way we would have been able to serve as much as we’ve been able to this year is this multi-chip strategy,” Anthropic Chief Product Officer Mike Krieger told CNBC.
That strategy spans TPUs, Amazon Trainium and Nvidia GPUs, allowing the company to optimize for cost, performance and redundancy. Krieger said Anthropic did a lot of up-front work to make sure its models can run equally well across the silicon providers.
“I’ve seen that investment pay off now that we’re able to come online with these massive data centers and meet customers where they are,” Krieger said.
Hefty spending is coming
Two months before the Anthropic deal, Google forged a six-year cloud agreement with Meta worth more than $10 billion, though it’s not clear how much of the arrangement includes use of TPUs. And while OpenAI said it will start using Google’s cloud as it diversifies away from Microsoft, the company told Reuters it’s not deploying GPUs.
Alphabet CFO Anat Ashkenazi attributed Google’s cloud momentum in the latest quarter to rising enterprise demand for Google’s full AI stack. The company said it signed more billion-dollar cloud deals in the first nine months of 2025 than in the previous two years combined.
“In GCP, we see strong demand for enterprise AI infrastructure, including TPUs and GPUs,” Ashkenazi said, adding that users are also flocking to the company’s latest Gemini offerings as well as services “such as cybersecurity and data analytics.”
Amazon, which reported 20% growth in its market-leading cloud infrastructure business last quarter, is expressing similar sentiment.
AWS CEO Matt Garman told CNBC in a recent interview that the company’s Trainium chip series is gaining momentum. He said “every Trainium 2 chip we land in our data centers today is getting sold and used,” and he promised further performance gains and efficiency improvements with Trainium 3.
Shareholders have shown a willingness to stomach hefty investments.
Google just raised the high end of its capital expenditures forecast for the year to $93 billion, up from prior guidance of $85 billion, with an even steeper ramp expected in 2026. The stock price soared 38% in the third quarter, its best performance for any period in 20 years, and is up another 17% in the fourth quarter.
Mizuho recently pointed to Google’s distinct cost and performance advantage with TPUs, noting that while the chips were originally built for internal use, Google is now winning external customers and bigger workloads.
Morgan Stanley analysts wrote in a report in June that while Nvidia’s GPUs will likely remain the dominant chip provider in AI, growing developer familiarity with TPUs could become a meaningful driver of Google Cloud growth.
And analysts at D.A. Davidson said in September that they see so much demand for TPUs that Google should consider selling the systems “externally to customers,” including frontier AI labs.
“We continue to believe that Google’s TPUs remain the best alternative to Nvidia, with the gap between the two closing significantly over the past 9-12 months,” they wrote. “During this time, we’ve seen growing positive sentiment around TPUs.”