Connect with us

Published

on

Baidu CEO Robin Li speaks during the company’s Create conference in Shenzhen, China, on April 16, 2024.

Bloomberg | Bloomberg | Getty Images

SHENZHEN, China – One year after Chinese search engine operator Baidu released its ChatGPT-like Ernie bot, the company this week announced tools to encourage locals to develop artificial intelligence applications.

“In China today, there are 1 billion internet users, strong foundation models, sufficient AI application scenarios and the most complete industrial system in the world,” CEO Robin Li said in his opening speech at Baidu’s annual AI developers conference on Tuesday.

“Everyone can be a developer,” he said in Mandarin, according to a CNBC translation.

While many point out how China lags behind the U.S. in artificial intelligence capabilities, others emphasize how the strength of the Chinese market lies more in technological application. Take next-day e-commerce and 30-minute food delivery, for example.

Baidu’s newly announced AI tools allow people with no coding knowledge to create generative AI-powered chatbots for specific functions, which can then be integrated in a website, Baidu search engine results or other online portals. That’s different from a similar tool called GPTs that OpenAI launched earlier this year, since those custom-built chatbots — for everything from suggesting movies to fixing code — sit within the ChatGPT interface.

Expect AI to become as universal as email: HSBC

The basic Baidu tools are generally available to try for free, up until a certain usage limit, similar to some of Google’s cloud and AI functions. OpenAI charges a monthly fee for the latest version of ChatGPT and the ability to use it for computer programs. The older ChatGPT 3.5 model is free to use, but without access to the custom-built GPTs.

Baidu this week also announced three new versions of its Ernie AI model — called “Speed,” “Lite” and “Tiny” — that coders can selectively access, based on the complexity of the task.  

“It feels like their focus is on building the entire native AI development ecosystem, providing a full set of development tools and platform solutions,” said Bo Du, managing director at WestSummit Capital Management. That’s according to a CNBC translation of the Chinese remarks.

Baidu said this week that Ernie bot has accumulated more than 200 million users since its launch in March last year, and that computer programs are accessing the underlying AI model 200 million times a day. The company said more than 85,000 business clients have used its AI cloud platform to create 190,000 AI applications.

How the tech is being used

Many of the use cases Baidu showed off this week centered on consumer-facing applications: tourism and creation of content such as picture books and scheduling meetings.

In a demonstration hall, Baidu business departments showed off how the AI tools could be integrated with virtual people doing livestreams, or directing search engine traffic to an AI-based interactive buying guide.

Buysmart.AI, which won Baidu’s AI competition last year, uses the tech for an online shopping assistant connected to Chinese social media platform Weibo. The startup said it is using ChatGPT for a standalone interactive e-commerce app in the U.S.

“Personally I think that Ernie 4.0 has a better grasp of Chinese than ChatGPT 3.5,” Buysmart.AI co-founder Andy Qiu said in an interview. That’s according to a CNBC translation of his Mandarin-language remarks.

Consumers in the U.S. are currently more interested in AI products than users in China are, Qiu said. But he said that overall there is still room for improvement when it comes to building consumers’ trust of AI assistants and convincing users to place an order.

Also on display was a humanoid robot developed by Shenzhen-based UBTech Robotics that used Baidu’s Ernie AI model for understanding commands and reading written words.

It’s not immediately clear how such AI applications can significantly change business at this point. But Baidu is the latest to roll out more tools for people to experiment more easily and cheaply with.

Customer service, voice assistants and internet-connected devices can use smaller AI models to respond quickly to users, pointed out Helen Chai, managing director at CIC Consulting.

She added that in scenarios such as legal consultation or medical diagnosis, small AI models can be trained on specific data to achieve performance that’s comparable to larger AI models.

In the future, big AI-based applications will be based on a mixture of models, Baidu CEO Li said, using the technical term of “mixture of experts” or MoE.

He also promoted Baidu’s capabilities in AI-produced code, one of the areas in which Silicon Valley tech companies see the most potential for generative AI.

Baidu said since it deployed its “Comate” AI coding assistant a year ago, the tool has contributed to 27% of the tech company’s newly generated code. Audio streaming app Ximalaya, IT services company iSoftStone and Shanghai Mitsubishi Elevator are among more than 10,000 corporate Comate users, and have adopted nearly half of the code the tool generates, according to Baidu.

The global rush for developing generative AI has created a shortage in the semiconductors needed to provide the computing power. Chinese companies face added constraints due to U.S. restrictions on chip exports.

Baidu did not specifically discuss a shortage in computing power during the main conference session. In his speech, Dou Shen, head of AI cloud at Baidu, noted “uncertainties” in the chip supply chain and announced that Baidu has a platform that can access the power of several different kinds of chips.

Back in February, Li said on an earnings call that Baidu’s AI chip reserve “enables us to continue enhancing Ernie for the next one or two years.” The company is set to release first-quarter results on May 16.

Continue Reading

Technology

Oracle set to report quarterly results after the bell

Published

on

By

Oracle set to report quarterly results after the bell

Larry Ellison, Oracle’s co-founder and chief technology officer, appears at the Formula One British Grand Prix in Towcester, U.K., on July 6, 2025.

Jay Hirano | Sopa Images | Lightrocket | Getty Images

Oracle is scheduled to report fiscal second-quarter results after market close on Wednesday.

Here’s what analysts are expecting, according to LSEG:

  • Earnings per share: $1.64 adjusted
  • Revenue: $16.21 billion

Wall Street expects revenue to increase 15% in the quarter that ended Nov. 30, from $14.1 billion a year earlier. Analysts polled by StreetAccount are looking for $7.92 billion in cloud revenue and $6.06 billion from software.

The report lands at a critical moment for Oracle, which has tried to position itself at the center of the artificial intelligence boom by committing to massive build-outs. While the move has been a boon for Oracle’s revenue and its backlog, investors have grown concerned about the amount of debt the company is raising and the risks it faces should the AI market slow.

The stock plummeted 23% in November, its worst monthly performance since 2001 and, as of Tuesday’s close, is 33% below its record reached in September. Still, the shares are up 33% for the year, outperforming the Nasdaq, which has gained 22% over that stretch.

Over the past decade, Oracle has diversified its business beyond databases and enterprise software and into cloud infrastructure, where it competes with Amazon, Microsoft and Google. Those companies are all vying for big AI contracts and are investing heavily in data centers and hardware necessary to meet expected demand.

OpenAI, which sparked the generative AI rush with the launch of ChatGPT three years ago, has committed to spending more than $300 billion on Oracle’s infrastructure services over five years.

“Oracle’s job is not to imagine gigawatt-scale data centers. Oracle’s job is to build them,” Larry Ellison, the company’s co-founder and chairman, told investors in September.

Oracle raised $18 billion during the period, one of the biggest issuances on record for a tech company. Skeptical investors have been buying five-year credit default swaps, driving them to multiyear highs. Credit default swaps are like insurance for investors, with buyers paying for protection in case the borrower can’t repay its debt.

“Customer concentration is a major issue here, but I think the bigger thing is, How are they going to pay for this?” said RBC analyst Rishi Jaluria, who has the equivalent of a hold rating on Oracle’s stock.

During the quarter, Oracle named executives Clay Magouyrk and Mike Sicilia as the company’s new CEOs, succeeding Safra Catz. Oracle also introduced AI agents for automating various facets of finance, human resources and sales.

Executives will discuss the results and issue guidance on a conference call starting at 5 p.m. ET.

WATCH: Oracle’s debt concerns loom large ahead of quarterly earnings

Oracle's debt concerns loom large ahead of quarterly earnings

Continue Reading

Technology

Nvidia refutes report that China’s DeepSeek is using its banned Blackwell AI chips

Published

on

By

Nvidia refutes report that China's DeepSeek is using its banned Blackwell AI chips

Jensen Huang, chief executive officer of Nvidia Corp., outside the US Capitol in Washington, DC, US, on Wednesday, Dec. 3, 2025.

Bloomberg | Bloomberg | Getty Images

Nvidia on Wednesday refuted a report that the Chinese artificial intelligence startup DeepSeek has been using smuggled Blackwell chips to develop its upcoming model.

The U.S. has banned the export of Nvidia’s Blackwell chips, which are considered the company’s most advanced offerings, to China in an effort to stay ahead in the AI race.

DeepSeek is reportedly using chips that were snuck into the country without authorization, according to The Information.

“We haven’t seen any substantiation or received tips of ‘phantom datacenters’ constructed to deceive us and our OEM partners, then deconstructed, smuggled, and reconstructed somewhere else,” a Nvidia spokesperson said in a statement. “While such smuggling seems farfetched, we pursue any tip we receive.”

Read more CNBC tech news

Nvidia has been one of the biggest winners of the AI boom so far because it develops the graphics processing units (GPUs) that are key for training models and running large workloads.

Since the hardware is so crucial for advancing AI technology, Nvidia’s relationship with China has become a political flashpoint among U.S. lawmakers.

President Donald Trump on Monday said Nvidia can ship its H200 chips to “approved customers” in China and elsewhere on the condition that the U.S. will get 25% of those sales.

The announcement was met with pushback from some Republicans.

DeepSeek spooked the U.S. tech sector in January when it released a reasoning model, called R1, that rocketed to the top of app stores and industry leaderboards. R1 was also created at a fraction of the cost of other models in the U.S., according to some analyst estimates.

In August, DeepSeek hinted that China will soon have its own “next generation” chips to support its AI models.

WATCH: Nvidia selling H200 AI chips to China is net positive, says Patrick Moorhead

Nvidia selling H200 AI chips to China is net positive, says Patrick Moorhead

– CNBC’s Kristina Partsinevelos contributed to this report.

Continue Reading

Technology

‘Greetings, earthlings’: Nvidia-backed Starcloud trains first AI model in space as orbital data center race heats up

Published

on

By

‘Greetings, earthlings’: Nvidia-backed Starcloud trains first AI model in space as orbital data center race heats up

The Starcloud-1 satellite is launched into space from a SpaceX rocket on November 2, 2025.

Courtesy: SpaceX | Starcloud

Nvidia-backed startup Starcloud trained an artificial intelligence model from space for the first time, signaling a new era for orbital data centers that could alleviate Earth’s escalating digital infrastructure crisis.

Last month, the Washington-based company launched a satellite with an Nvidia H100 graphics processing unit, sending a chip into outer space that’s 100 times more powerful than any GPU compute that has been in space before. Now, the company’s Starcloud-1 satellite is running and querying responses from Gemma, an open large language model from Google, in orbit, marking the first time in history that an LLM has been has run on a high-powered Nvidia GPU in outer space, CNBC has learned.

“Greetings, Earthlings! Or, as I prefer to think of you — a fascinating collection of blue and green,” reads a message from the recently launched satellite. “Let’s see what wonders this view of your world holds. I’m Gemma, and I’m here to observe, analyze, and perhaps, occasionally offer a slightly unsettlingly insightful commentary. Let’s begin!” the model wrote.

Starcloud’s output Gemma in space. Gemma is a family of open models built from the same technology used to create Google’s Gemini AI models.

Starcloud

Starcloud wants to show outer space can be a hospitable environment for data centers, particularly as Earth-based facilities strain power grids, consume billions of gallons of water annually and produce hefty greenhouse gas emissions. The electricity consumption of data centers is projected to more than double by 2030, according to data from the International Energy Agency.

Starcloud CEO Philip Johnston told CNBC that the company’s orbital data centers will have 10 times lower energy costs than terrestrial data centers.

“Anything you can do in a terrestrial data center, I’m expecting to be able to be done in space. And the reason we would do it is purely because of the constraints we’re facing on energy terrestrially,” Johnston said in an interview.

Johnston, who co-founded the startup in 2024, said Starcloud-1’s operation of Gemma is proof that space-based data centers can exist and operate a variety of AI models in the future, particularly those that require large compute clusters.

“This very powerful, very parameter dense model is living on our satellite,” Johnston said. “We can query, it and it will respond in the same way that when you query a chat from a database on Earth, it will give you a very sophisticated response. We can do that with our satellite.”

In a statement to CNBC, Google DeepMind product director Tris Warkentin said that “seeing Gemma run in the harsh environment of space is a testament to the flexibility and robustness of open models.”

In addition to Gemma, Starcloud was able to train NanoGPT, an LLM created by OpenAI founding member Andrej Karpathy, on the H100 chip using the complete works of Shakespeare. This led the model to speak in Shakespearean English.

Starcloud — a member of the Nvidia Inception program and graduate from Y Combinator and the Google for Startups Cloud AI Accelerator — plans to build a 5-gigawatt orbital data center with solar and cooling panels that measure roughly 4 kilometers in both width and height. A compute cluster of that gigawatt size would produce more power than the largest power plant in the U.S. and would be substantially smaller and cheaper than a terrestrial solar farm of the same capacity, according to Starcloud’s white paper.

These data centers in space would capture constant solar energy to power next-generation AI models, unhindered by the Earth’s day and night cycles and weather changes. Starcloud’s satellites should have a five-year lifespan given the expected lifetime of the Nvidia chips on its architecture, Johnston said.

Orbital data centers would have real-world commercial and military use cases. Already, Starcloud’s systems can enable real-time intelligence and, for example, spot the thermal signature of a wildfire the moment it ignites and immediately alert first responders, Johnston said.

“We’ve linked in the telemetry of the satellite, so we linked in the vital signs that it’s drawing from the sensors — things like altitude, orientation, location, speed,” Johnston said. “You can ask it, ‘Where are you now?’ and it will say ‘I’m above Africa and in 20 minutes, I’ll be above the Middle East.’ And you could also say, ‘What does it feel like to be a satellite? And it will say, ‘It’s kind of a bit weird’ … It’ll give you an interesting answer that you could only have with a very high-powered model.”

Starcloud is working on customer workloads by running inference on satellite imagery from observation company Capella Space, which could help spot lifeboats from capsized vessels at sea and forest fires in a certain location. The company will include several Nvidia H100 chips and integrate Nvidia’s Blackwell platform onto its next satellite launch in October 2026 to offer greater AI performance. The satellite launching next year will feature a module running a cloud platform from cloud infrastructure startup Crusoe, allowing customers to deploy and operate AI workloads from space.

“Running advanced AI from space solves the critical bottlenecks facing data centers on Earth,” Johnston told CNBC.

“Orbital compute offers a way forward that respects both technological ambition and environmental responsibility. When Starcloud-1 looked down, it saw a world of blue and green. Our responsibility is to keep it that way,” he added.

The risks

Risks in operating orbital data centers remain, however. Analysts from Morgan Stanley have noted that orbital data centers could face hurdles such as harsh radiation, difficulty of in-orbit maintenance, debris hazards and regulatory issues tied to data governance and space traffic.

Still, tech giants are pursuing orbital data centers given the prospect of nearly limitless solar energy and greater, gigawatt-sized operations in space.

Along with Starcloud and Nvidia’s efforts, several companies have announced space-based data center missions. On Nov. 4, Google unveiled a “moonshot” initiative titled Project Suncatcher, which aims to put solar-powered satellites into space with Google’s tensor processing units. Privately-owned Lonestar Data Holdings is working to put the first-ever commercial lunar data center on the moon’s surface.

OpenAI CEO Sam Altman has explored an acquisition or partnership with a rocket maker, suggesting a desire to compete against Elon Musk‘s SpaceX, according to The Wall Street Journal. SpaceX is a key launch partner for Starcloud.

Referring to Starcloud’s launch in early November, Nvidia senior director of AI infrastructure Dion Harris said: “From one small data center, we’ve taken a giant leap toward a future where orbital computing harnesses the infinite power of the sun.”

Continue Reading

Trending