Connect with us

Published

on

Amazon plans to give Alexa an AI overhaul

Amazon is upgrading its decade-old Alexa voice assistant with generative artificial intelligence and plans to charge a monthly subscription fee to offset the cost of the technology, according to people with knowledge of Amazon’s plans. 

The Seattle-based tech and retail giant will launch a more conversational version of Alexa later this year, potentially positioning it to better compete with new generative AI-powered chatbots from companies including Google and OpenAI, according to two sources familiar with the matter, who asked not to be named because the discussions were private. Amazon’s subscription for Alexa will not be included in the $139-per-year Prime offering, and Amazon has not yet nailed down the price point, one source said.

Amazon declined to comment on its plans for Alexa. 

While Amazon wowed consumers with Alexa’s voice-driven tasks in 2014, its capabilities could seem old-fashioned amid recent leaps in artificial intelligence. Last week, OpenAI announced GPT-4o, with the capability for two-way conversations that can go significantly deeper than Alexa. For example, it can translate conversations into different languages in real time. Google launched a similar generative-AI-powered voice feature for Gemini. 

Some interpreted last week’s announcements as a threat to Alexa and Siri, Apple‘s voice assistant feature for iPhones. NYU professor Scott Galloway called the updates the “Alexa and Siri killers” on his recent podcast. Many people use Alexa and Siri for basic tasks, such as setting timers or alarms and announcing the weather.

The development of new AI chatbots in recent months has increased the pressure internally on a division that was once seen as a darling of Amazon founder Jeff Bezos, according to the sources — but has been subject to strict profit imperatives since his departure. 

Three former employees pointed to Bezos’ early obsession with Alexa, describing it as his passion project. Attention from Bezos resulted in more dollars and less pressure to make a return on those funds immediately. 

That changed when Andy Jassy took over as CEO in 2021, according to three sources. Jassy was charged with rightsizing Amazon’s business during the pandemic, and Alexa became less of a priority internally, they said. Jassy has been privately underwhelmed with what modern-day Alexa is capable of, according to one person. The Alexa team worried they had invented an expensive alarm clock, weather machine and way to play Spotify music, one source said.  

For instance, Jassy, an avid sports fan, asked the voice assistant the live score of a recent game, according to a person in the room, and was openly frustrated that Alexa didn’t know an answer that was so easy to find online. 

When reached for comment, Amazon pointed to the company’s annual shareholder letter released last month. In it, Jassy mentioned that the company was building a “substantial number of GenAI applications across every Amazon consumer business,” adding that that included “an even more intelligent and capable Alexa.”

The team is now tasked with turning Alexa into a relevant device that holds up amid the new AI competition, and one that justifies the resources and headcount Amazon has dedicated to it. It has undergone a massive reorganization, with much of the team shifting to the artificial general intelligence, or AGI, team, according to three sources. Others pointed to bloat within Alexa, a team of thousands of employees.

As of 2023, Amazon said it had sold more than 500 million Alexa-enabled devices, giving the company a foothold with consumers. 

Alexa, were you too early?

Apple, Amazon and Google were early movers with their voice assistants, which did employ AI. But the current wave of advanced generative AI enables much more creative, human-sounding interactions. Apple is expected to unveil a more conversational Siri at its annual developers conference in June, according to The New York Times. 

Those who worked on the Alexa team describe it as a great idea that may have been too early, and that it’s going to be hard to turn the ship around. 

There’s also the challenge of finding AI engineering talent, as OpenAI, Microsoft and Google recruit from the same pool of academics and tech talent. Plus, generative AI workloads are expensive thanks to the hardware and computing power required. One source estimated the cost of using generative AI in Alexa at 2 cents per query, and said a $20 price point was floated internally. Another suggested it would need to be in a single-digit dollar amount, which would undercut other subscription offerings. OpenAI’s ChatGPT charges $20 per month for its advanced models. 

Still, they point to Alexa’s installed user base, with devices in hundreds of millions of homes, as an opportunity. Those who worked on Alexa say the fact that it’s already in people’s living rooms and kitchens makes the stakes higher, and mistakes more costly if Alexa doesn’t understand a command or provides unreliable information. 

Amazon has been battling a perception that it’s behind in artificial intelligence. While it offers multiple AI models on Amazon Web Services, it does not have a leading large language model to unseat OpenAI, Google or Meta. Amazon spent $2.75 billion backing AI startup Anthropic, its largest venture investment in the company’s three-decade history. Google also has an Anthropic investment and partnership.

Amazon will use its own large language model, Titan, in the Alexa upgrade, according to a source.  

Bezos is among those who have voiced concern that Amazon is behind in AI, according to two sources familiar with him. Bezos is still “very involved” in Amazon’s AI efforts, CNBC reported last week, and has been sending Amazon executives emails wondering why certain AI startups are picking other cloud providers over AWS. 

Don’t miss these exclusives from CNBC PRO

Jeff Bezos still 'very involved' in Amazon's AI efforts, sources tell CNBC

Continue Reading

Technology

Oracle set to report quarterly results after the bell

Published

on

By

Oracle set to report quarterly results after the bell

Larry Ellison, Oracle’s co-founder and chief technology officer, appears at the Formula One British Grand Prix in Towcester, U.K., on July 6, 2025.

Jay Hirano | Sopa Images | Lightrocket | Getty Images

Oracle is scheduled to report fiscal second-quarter results after market close on Wednesday.

Here’s what analysts are expecting, according to LSEG:

  • Earnings per share: $1.64 adjusted
  • Revenue: $16.21 billion

Wall Street expects revenue to increase 15% in the quarter that ended Nov. 30, from $14.1 billion a year earlier. Analysts polled by StreetAccount are looking for $7.92 billion in cloud revenue and $6.06 billion from software.

The report lands at a critical moment for Oracle, which has tried to position itself at the center of the artificial intelligence boom by committing to massive build-outs. While the move has been a boon for Oracle’s revenue and its backlog, investors have grown concerned about the amount of debt the company is raising and the risks it faces should the AI market slow.

The stock plummeted 23% in November, its worst monthly performance since 2001 and, as of Tuesday’s close, is 33% below its record reached in September. Still, the shares are up 33% for the year, outperforming the Nasdaq, which has gained 22% over that stretch.

Over the past decade, Oracle has diversified its business beyond databases and enterprise software and into cloud infrastructure, where it competes with Amazon, Microsoft and Google. Those companies are all vying for big AI contracts and are investing heavily in data centers and hardware necessary to meet expected demand.

OpenAI, which sparked the generative AI rush with the launch of ChatGPT three years ago, has committed to spending more than $300 billion on Oracle’s infrastructure services over five years.

“Oracle’s job is not to imagine gigawatt-scale data centers. Oracle’s job is to build them,” Larry Ellison, the company’s co-founder and chairman, told investors in September.

Oracle raised $18 billion during the period, one of the biggest issuances on record for a tech company. Skeptical investors have been buying five-year credit default swaps, driving them to multiyear highs. Credit default swaps are like insurance for investors, with buyers paying for protection in case the borrower can’t repay its debt.

“Customer concentration is a major issue here, but I think the bigger thing is, How are they going to pay for this?” said RBC analyst Rishi Jaluria, who has the equivalent of a hold rating on Oracle’s stock.

During the quarter, Oracle named executives Clay Magouyrk and Mike Sicilia as the company’s new CEOs, succeeding Safra Catz. Oracle also introduced AI agents for automating various facets of finance, human resources and sales.

Executives will discuss the results and issue guidance on a conference call starting at 5 p.m. ET.

WATCH: Oracle’s debt concerns loom large ahead of quarterly earnings

Oracle's debt concerns loom large ahead of quarterly earnings

Continue Reading

Technology

Nvidia refutes report that China’s DeepSeek is using its banned Blackwell AI chips

Published

on

By

Nvidia refutes report that China's DeepSeek is using its banned Blackwell AI chips

Jensen Huang, chief executive officer of Nvidia Corp., outside the US Capitol in Washington, DC, US, on Wednesday, Dec. 3, 2025.

Bloomberg | Bloomberg | Getty Images

Nvidia on Wednesday refuted a report that the Chinese artificial intelligence startup DeepSeek has been using smuggled Blackwell chips to develop its upcoming model.

The U.S. has banned the export of Nvidia’s Blackwell chips, which are considered the company’s most advanced offerings, to China in an effort to stay ahead in the AI race.

DeepSeek is reportedly using chips that were snuck into the country without authorization, according to The Information.

“We haven’t seen any substantiation or received tips of ‘phantom datacenters’ constructed to deceive us and our OEM partners, then deconstructed, smuggled, and reconstructed somewhere else,” a Nvidia spokesperson said in a statement. “While such smuggling seems farfetched, we pursue any tip we receive.”

Read more CNBC tech news

Nvidia has been one of the biggest winners of the AI boom so far because it develops the graphics processing units (GPUs) that are key for training models and running large workloads.

Since the hardware is so crucial for advancing AI technology, Nvidia’s relationship with China has become a political flashpoint among U.S. lawmakers.

President Donald Trump on Monday said Nvidia can ship its H200 chips to “approved customers” in China and elsewhere on the condition that the U.S. will get 25% of those sales.

The announcement was met with pushback from some Republicans.

DeepSeek spooked the U.S. tech sector in January when it released a reasoning model, called R1, that rocketed to the top of app stores and industry leaderboards. R1 was also created at a fraction of the cost of other models in the U.S., according to some analyst estimates.

In August, DeepSeek hinted that China will soon have its own “next generation” chips to support its AI models.

WATCH: Nvidia selling H200 AI chips to China is net positive, says Patrick Moorhead

Nvidia selling H200 AI chips to China is net positive, says Patrick Moorhead

– CNBC’s Kristina Partsinevelos contributed to this report.

Continue Reading

Technology

‘Greetings, earthlings’: Nvidia-backed Starcloud trains first AI model in space as orbital data center race heats up

Published

on

By

‘Greetings, earthlings’: Nvidia-backed Starcloud trains first AI model in space as orbital data center race heats up

The Starcloud-1 satellite is launched into space from a SpaceX rocket on November 2, 2025.

Courtesy: SpaceX | Starcloud

Nvidia-backed startup Starcloud trained an artificial intelligence model from space for the first time, signaling a new era for orbital data centers that could alleviate Earth’s escalating digital infrastructure crisis.

Last month, the Washington-based company launched a satellite with an Nvidia H100 graphics processing unit, sending a chip into outer space that’s 100 times more powerful than any GPU compute that has been in space before. Now, the company’s Starcloud-1 satellite is running and querying responses from Gemma, an open large language model from Google, in orbit, marking the first time in history that an LLM has been has run on a high-powered Nvidia GPU in outer space, CNBC has learned.

“Greetings, Earthlings! Or, as I prefer to think of you — a fascinating collection of blue and green,” reads a message from the recently launched satellite. “Let’s see what wonders this view of your world holds. I’m Gemma, and I’m here to observe, analyze, and perhaps, occasionally offer a slightly unsettlingly insightful commentary. Let’s begin!” the model wrote.

Starcloud’s output Gemma in space. Gemma is a family of open models built from the same technology used to create Google’s Gemini AI models.

Starcloud

Starcloud wants to show outer space can be a hospitable environment for data centers, particularly as Earth-based facilities strain power grids, consume billions of gallons of water annually and produce hefty greenhouse gas emissions. The electricity consumption of data centers is projected to more than double by 2030, according to data from the International Energy Agency.

Starcloud CEO Philip Johnston told CNBC that the company’s orbital data centers will have 10 times lower energy costs than terrestrial data centers.

“Anything you can do in a terrestrial data center, I’m expecting to be able to be done in space. And the reason we would do it is purely because of the constraints we’re facing on energy terrestrially,” Johnston said in an interview.

Johnston, who co-founded the startup in 2024, said Starcloud-1’s operation of Gemma is proof that space-based data centers can exist and operate a variety of AI models in the future, particularly those that require large compute clusters.

“This very powerful, very parameter dense model is living on our satellite,” Johnston said. “We can query, it and it will respond in the same way that when you query a chat from a database on Earth, it will give you a very sophisticated response. We can do that with our satellite.”

In a statement to CNBC, Google DeepMind product director Tris Warkentin said that “seeing Gemma run in the harsh environment of space is a testament to the flexibility and robustness of open models.”

In addition to Gemma, Starcloud was able to train NanoGPT, an LLM created by OpenAI founding member Andrej Karpathy, on the H100 chip using the complete works of Shakespeare. This led the model to speak in Shakespearean English.

Starcloud — a member of the Nvidia Inception program and graduate from Y Combinator and the Google for Startups Cloud AI Accelerator — plans to build a 5-gigawatt orbital data center with solar and cooling panels that measure roughly 4 kilometers in both width and height. A compute cluster of that gigawatt size would produce more power than the largest power plant in the U.S. and would be substantially smaller and cheaper than a terrestrial solar farm of the same capacity, according to Starcloud’s white paper.

These data centers in space would capture constant solar energy to power next-generation AI models, unhindered by the Earth’s day and night cycles and weather changes. Starcloud’s satellites should have a five-year lifespan given the expected lifetime of the Nvidia chips on its architecture, Johnston said.

Orbital data centers would have real-world commercial and military use cases. Already, Starcloud’s systems can enable real-time intelligence and, for example, spot the thermal signature of a wildfire the moment it ignites and immediately alert first responders, Johnston said.

“We’ve linked in the telemetry of the satellite, so we linked in the vital signs that it’s drawing from the sensors — things like altitude, orientation, location, speed,” Johnston said. “You can ask it, ‘Where are you now?’ and it will say ‘I’m above Africa and in 20 minutes, I’ll be above the Middle East.’ And you could also say, ‘What does it feel like to be a satellite? And it will say, ‘It’s kind of a bit weird’ … It’ll give you an interesting answer that you could only have with a very high-powered model.”

Starcloud is working on customer workloads by running inference on satellite imagery from observation company Capella Space, which could help spot lifeboats from capsized vessels at sea and forest fires in a certain location. The company will include several Nvidia H100 chips and integrate Nvidia’s Blackwell platform onto its next satellite launch in October 2026 to offer greater AI performance. The satellite launching next year will feature a module running a cloud platform from cloud infrastructure startup Crusoe, allowing customers to deploy and operate AI workloads from space.

“Running advanced AI from space solves the critical bottlenecks facing data centers on Earth,” Johnston told CNBC.

“Orbital compute offers a way forward that respects both technological ambition and environmental responsibility. When Starcloud-1 looked down, it saw a world of blue and green. Our responsibility is to keep it that way,” he added.

The risks

Risks in operating orbital data centers remain, however. Analysts from Morgan Stanley have noted that orbital data centers could face hurdles such as harsh radiation, difficulty of in-orbit maintenance, debris hazards and regulatory issues tied to data governance and space traffic.

Still, tech giants are pursuing orbital data centers given the prospect of nearly limitless solar energy and greater, gigawatt-sized operations in space.

Along with Starcloud and Nvidia’s efforts, several companies have announced space-based data center missions. On Nov. 4, Google unveiled a “moonshot” initiative titled Project Suncatcher, which aims to put solar-powered satellites into space with Google’s tensor processing units. Privately-owned Lonestar Data Holdings is working to put the first-ever commercial lunar data center on the moon’s surface.

OpenAI CEO Sam Altman has explored an acquisition or partnership with a rocket maker, suggesting a desire to compete against Elon Musk‘s SpaceX, according to The Wall Street Journal. SpaceX is a key launch partner for Starcloud.

Referring to Starcloud’s launch in early November, Nvidia senior director of AI infrastructure Dion Harris said: “From one small data center, we’ve taken a giant leap toward a future where orbital computing harnesses the infinite power of the sun.”

Continue Reading

Trending