LISBON, Portugal — Tech giants are increasingly investing in the development of so-called “sovereign” artificial intelligence models as they seek to boost competitiveness by focusing more on local infrastructure.
Data sovereignty refers to the idea that people’s data should be stored on infrastructure within the country or continent they reside in.
“Sovereign AI is a relatively new term that’s emerged in the last year or so,” Chris Gow, IT networking giant Cisco’s Brussels-based EU public policy lead, told CNBC.
Currently, many of the biggest large language models (LLMs), like OpenAI’s ChatGPT and Anthropic’s Claude, use data centers based in the U.S. to store data and process requests via the cloud.
This has led to concern from politicians and regulators in Europe, who see dependence on U.S. technology as harmful to the continent’s competitiveness — and, more worryingly, technological resilience.
Where did ‘AI sovereignty’ come from?
The notion of data and technological sovereignty is something that has previously been on Europe’s agenda. It came about, in part, as a result of businesses reacting to new regulations.
The European Union’s General Data Protection Regulation, for example, requires companies to handle user data in a secure, compliant way that respects their right to privacy. High-profile cases in the EU have also raised doubts over whether data on European citizens can be transferred across borders safely.
The European Court of Justice in 2020 invalidated an EU-U.S. data-sharing framework, on the grounds that the pact did not afford the same level of protection as guaranteed within the EU by the General Data Protection Regulation (GDPR). Last year the EU-U.S. Data Privacy Framework was formed to ensure that data can flow safely between the EU and U.S.
These political development have ultimately resulted in a push toward localization of cloud infrastructure, where data is stored and processed for many online services.
Filippo Sanesi, global head of marketing and operations at OVHCloud, said the French cloud firm is seeing lots of demand for its European-located infrastructure, as they “understand the value of having their data in Europe, which are subject to European legislation.”
“As this concept of data sovereignty becomes more mature and people understand what it means, we see more and more companies understanding the importance of having your data locally and under a specific jurisdiction and governance,” Sanesi told CNBC. “We have a lot of data,” he added. “This data is sovereign in specific countries, under specific regulations.”
“Now, with this data, you can actually make products and services for AI, and those services should then be sovereign, should be controlled, deployed and developed locally by local talent for the local population or businesses.”
The AI sovereignty push hasn’t been driven forward by regulators — at least, not yet, according to Cisco’s Gow. Rather, it’s come from private companies, which are opening more data centers — facilities containing vast amounts of computing equipment to enable cloud-based AI tools — in Europe, he said.
Sovereign AI is “more driven by the industry naming it that, than it is from the policymakers’ side,” Gow said. “You don’t see the ‘AI sovereignty’ terminology used on the regulator side yet.”
Countries are pushing the idea of AI sovereignty because they recognize AI is “the future” and a “massively strategic technology,” Gow said.
Governments are focusing on boosting their domestic tech companies and ecosystems, as well as the all-important backend infrastructure that enables AI services.
“The AI workload uses 20 times the bandwidth of a traditional workload,” Gow said. It’s also about enabling the workforce, according to Gow, as firms need skilled workers to be successful.
Most important of all, however, is the data. “What you’re seeing is quite a few attempts from that side to think about training LLMs on localized data, in language,” Gow said.
The aim of the Italia project is to store results in a given jurisdiction and rely on data from citizens within that region so that results produced by the AI systems there are more grounded in local languages, culture and history.
“Sovereign AI is about reflecting the values of an organization or, equally, the country that you’re in and the values and the language,” David Hogan, EMEA head of enterprise sales for chipmaking giant Nvidia, told CNBC.
“The core challenge is that most of the frontier models today have been trained primarily on Western data generally,” Hogan added.
In Denmark for example, where Nvidia has a major presence, officials are concerned about vital services such as health care and telecoms being delivered by AI systems that aren’t “reflective” of local Danish culture and values, according to Hogan.
On Wednesday, Denmark laid out a landmark white paper outlining how companies can use AI in compliance with the incoming EU AI Act — the world’s first major AI law. The document is meant to serve as a blueprint for other EU nations to follow and adopt.
“If you’re in a European country that’s not one of the major language countries that’s spoken internationally, probably less than 2% of the data is trained on your language — let alone your culture,” Hogan said.
How regulation fueled a mindset shift
That’s not to say regulations haven’t proven an important factor in getting tech giants to think more about building localized AI infrastructure within Europe.
OVHCloud’s Sanesi said regulations like the EU’s GDPR catalyzed a lot of the interest in onshoring the processing of data in a given region.
The concept of AI sovereignty is also getting buy-in from local European tech firms.
Earlier this week, Berlin-headquartered search engine Ecosia and its Paris-based peer Qwant announced a joint venture to develop a European search index from scratch, aiming to serve improved French and German language results.
Meanwhile, French telecom operator Orange has said it’s in discussions with a number of foundational AI model companies about building a smartphone-based “sovereign AI” model for its customers that more accurately reflects their own language and culture.
“It wouldn’t make sense to build our own LLMs. So there’s a lot of discussion right now about, how do we partner with existing providers to make it more local and safer?” Bruno Zerbib, Orange’s chief technology officer, told CNBC.
“There are a lot of use cases where [AI data] can be processed locally [on a phone] instead of processed on the cloud,” Zerbib added. Orange hasn’t yet selected a partner for these sovereign AI model ambitions.
Thomas Fuller | SOPA Images | Lightrocket | Getty Images
Ambarella shares popped 19% after a report that the chip designer is currently working with bankers on a potential sale.
Bloomberg reported the news, citing sources familiar with the matter.
While no deal is imminent, the sources told Bloomberg that the firm may draw interest from semiconductor companies looking to improve their automotive business. Private equity firms have already expressed interest, according to the report.
Read more CNBC tech news
The Santa Clara, California-based company is known for its system-on-chip semiconductors and software used for edge artificial intelligence. Ambarella chips are used in the automotive sector for electronic mirrors and self-driving assistance systems.
Shares have slumped about 18% year to date. The company’s market capitalization last stood at nearly $2.6 billion.
Nvidia CEO Jensen Huang attends a roundtable discussion at the Viva Technology conference dedicated to innovation and startups at Porte de Versailles exhibition center in Paris on June 11, 2025.
The sales are worth nearly $15 million at Tuesday’s opening price.
The transactions are the first sale in Huang’s plan to sell as many as 600,000 shares of Nvidia through the end of 2025. It’s a plan that was announced in March, and it’d be worth $873 million at Tuesday’s opening price.
The Nvidia founder still owns more than 800 million Nvidia shares, according to Monday’s SEC filing. Huang has a net worth of about $126 billion, ranking him 12th on the Bloomberg Billionaires Index.
Nvidia stock is up more than 800% since December 2022 after OpenAI’s ChatGPT was first released to the public. That launch drew attention to Nvidia’s graphics processing units, or GPUs, which were needed to develop and power the artificial intelligence service.
The company’s chips remain in high demand with the majority of the AI chip market, and Nvidia has introduced two subsequent generations of its AI GPU technology.
Nvidia continues to grow. Its stock is up 9% this year, even as the company faces export control issues that could limit foreign markets for its AI chips.
In May, the company reported first-quarter earnings that showed the chipmaker’s revenue growing 69% on an annual basis to $44 billion during the quarter.
Dario Amodei, Anthropic CEO, speaking on CNBC’s Squawk Box outside the World Economic Forum in Davos, Switzerland on Jan. 21st, 2025.
Gerry Miller | CNBC
Anthropic‘s use of books to train its artificial intelligence model Claude was “fair use” and “transformative,” a federal judge ruled late on Monday.
Amazon-backed Anthropic’s AI training did not violate the authors’ copyrights since the large language models “have not reproduced to the public a given work’s creative elements, nor even one author’s identifiable expressive style,” wrote U.S. District Judge William Alsup.
“The purpose and character of using copyrighted works to train LLMs to generate new text was quintessentially transformative,” Alsup wrote. “Like any reader aspiring to be a writer.”
The decision was a significant win for AI companies as legal battles play out over the use and application of copyrighted works in developing and training LLMs. Alsup’s ruling begins to establish the legal limits and opportunities for the industry going forward.
Read more CNBC reporting on AI
A spokesperson for Anthropic said in a statement that the company was “pleased” with the ruling and that the decision was, “Consistent with copyright’s purpose in enabling creativity and fostering scientific progress.”
CNBC has reached out to the plaintiffs for comment.
The lawsuit, filed in the U.S. District Court for the Northern District of California, was brought by authors Andrea Bartz, Charles Graeber and Kirk Wallace Johnson in August. The suit alleged that Anthropic built a “multibillion-dollar business by stealing hundreds of thousands of copyrighted books.”
Alsup did, however, order a trial on the pirated material that Anthropic put into its central library of content, even though the company did not use it for AI training.
“That Anthropic later bought a copy of a book it earlier stole off the internet will not absolve it of liability for the theft, but it may affect the extent of statutory damages,” the judge wrote.