Connect with us

Published

on

David Silver, leader of the reinforcement learning research group at DeepMind, being awarded an honorary “ninth dan” professional ranking for AlphaGo.
JUNG YEON-JE | AFP | Getty Images

Computer scientists are questioning whether DeepMind, the Alphabet-owned U.K. firm that’s widely regarded as one of the world’s premier AI labs, will ever be able to make machines with the kind of “general” intelligence seen in humans and animals.

In its quest for artificial general intelligence, which is sometimes called human-level AI, DeepMind is focusing a chunk of its efforts on an approach called “reinforcement learning.”

This involves programming an AI to take certain actions in order to maximize its chance of earning a reward in a certain situation. In other words, the algorithm “learns” to complete a task by seeking out these preprogrammed rewards. The technique has been successfully used to train AI models how to play (and excel at) games like Go and chess. But they remain relatively dumb, or “narrow.” DeepMind’s famous AlphaGo AI can’t draw a stickman or tell the difference between a cat and a rabbit, for example, while a seven-year-old can.

Despite this, DeepMind, which was acquired by Google in 2014 for around $600 million, believes that AI systems underpinned by reinforcement learning could theoretically grow and learn so much that they break the theoretical barrier to AGI without any new technological developments.

Researchers at the company, which has grown to around 1,000 people under Alphabet’s ownership, argued in a paper submitted to the peer-reviewed Artificial Intelligence journal last month that “Reward is enough” to reach general AI. The paper was first reported by VentureBeat last week.

In the paper, the researchers claim that if you keep “rewarding” an algorithm each time it does something you want it to, which is the essence of reinforcement learning, then it will eventually start to show signs of general intelligence.

“Reward is enough to drive behavior that exhibits abilities studied in natural and artificial intelligence, including knowledge, learning, perception, social intelligence, language, generalization and imitation,” the authors write.

“We suggest that agents that learn through trial and error experience to maximize reward could learn behavior that exhibits most if not all of these abilities, and therefore that powerful reinforcement learning agents could constitute a solution to artificial general intelligence.”

Not everyone is convinced, however.

Samim Winiger, an AI researcher in Berlin, told CNBC that DeepMind’s “reward is enough” view is a “somewhat fringe philosophical position, misleadingly presented as hard science.”

He said the path to general AI is complex and that the scientific community is aware that there are countless challenges and known unknowns that “rightfully instill a sense of humility” in most researchers in the field and prevent them from making “grandiose, totalitarian statements” such as “RL is the final answer, all you need is reward.”

DeepMind told CNBC that while reinforcement learning has been behind some of its most well-known research breakthroughs, the AI technique accounts for only a fraction of the overall research it carries out. The company said it thinks it’s important to understand things at a more fundamental level, which is why it pursues other areas such as “symbolic AI” and “population-based training.”

“In somewhat typical DeepMind fashion, they chose to make bold statements that grabs attention at all costs, over a more nuanced approach,” said Winiger. “This is more akin to politics than science.”

Stephen Merity, an independent AI researcher, told CNBC that there’s “a difference between theory and practice.” He also noted that “a stack of dynamite is likely enough to get one to the moon, but it’s not really practical.”

Ultimately, there’s no proof either way to say whether reinforcement learning will ever lead to AGI.

Rodolfo Rosini, a tech investor and entrepreneur with a focus on AI, told CNBC: “The truth is nobody knows and that DeepMind’s main product continues to be PR and not technical innovation or products.”

Entrepreneur William Tunstall-Pedoe, who sold his Siri-like app Evi to Amazon, told CNBC that even if the researchers are correct “that doesn’t mean we will get there soon, nor does it mean that there isn’t a better, faster way to get there.”

DeepMind’s “Reward is enough” paper was co-authored by DeepMind heavyweights Richard Sutton and David Silver, who met DeepMind CEO Demis Hassabis at the University of Cambridge in the 1990s.

“The key problem with the thesis put forth by ‘Reward is enough’ is not that it is wrong, but rather that it cannot be wrong, and thus fails to satisfy Karl Popper’s famous criterion that all scientific hypotheses be falsifiable,” said a senior AI researcher at a large U.S. tech firm, who wished to remain anonymous due to the sensitive nature of the discussion.

“Because Silver et al. are speaking in generalities, and the notion of reward is suitably underspecified, you can always either cherry pick cases where the hypothesis is satisfied, or the notion of reward can be shifted such that it is satisfied,” the source added.

“As such, the unfortunate verdict here is not that these prominent members of our research community have erred in any way, but rather that what is written is trivial. What is learned from this paper, in the end? In the absence of practical, actionable consequences from recognizing the unalienable truth of this hypothesis, was this paper enough?”

What is AGI?

While AGI is often referred to as the holy grail of the AI community, there’s no consensus on what AGI actually is. One definition is it’s the ability of an intelligent agent to understand or learn any intellectual task that a human being can.

But not everyone agrees with that and some question whether AGI will ever exist. Others are terrified about its potential impacts and whether AGI would build its own, even more powerful, forms of AI, or so-called superintelligences.

Ian Hogarth, an entrepreneur turned angel investor, told CNBC that he hopes reinforcement learning isn’t enough to reach AGI. “The more that existing techniques can scale up to reach AGI, the less time we have to prepare AI safety efforts and the lower the chance that things go well for our species,” he said.

Winiger argues that we’re no closer to AGI today than we were several decades ago. “The only thing that has fundamentally changed since the 1950/60s, is that science-fiction is now a valid tool for giant corporations to confuse and mislead the public, journalists and shareholders,” he said.

Fueled with hundreds of millions of dollars from Alphabet every year, DeepMind is competing with the likes of Facebook and OpenAI to hire the brightest people in the field as it looks to develop AGI. “This invention could help society find answers to some of the world’s most pressing and fundamental scientific challenges,” DeepMind writes on its website.

DeepMind COO Lila Ibrahim said on Monday that trying to “figure out how to operationalize the vision” has been the biggest challenge since she joined the company in April 2018.

Continue Reading

Technology

Here are 4 major moments that drove the stock market last week

Published

on

By

Here are 4 major moments that drove the stock market last week

Continue Reading

Technology

Oracle says there have been ‘no delays’ in OpenAI arrangement after stock slide

Published

on

By

Oracle says there have been 'no delays' in OpenAI arrangement after stock slide

Oracle CEO Clay Magouyrk appears on a media tour of the Stargate AI data center in Abilene, Texas, on Sept. 23, 2025.

Kyle Grillot | Bloomberg | Getty Images

Oracle on Friday pushed back against a report that said the company will complete data centers for OpenAI, one of its major customers, in 2028, rather than 2027.

The delay is due to a shortage of labor and materials, according to the Friday report from Bloomberg, which cited unnamed people. Oracle shares fell to a session low of $185.98, down 6.5% from Thursday’s close.

“Site selection and delivery timelines were established in close coordination with OpenAI following execution of the agreement and were jointly agreed,” an Oracle spokesperson said in an email to CNBC. “There have been no delays to any sites required to meet our contractual commitments, and all milestones remain on track.”

The Oracle spokesperson did not specify a timeline for turning on cloud computing infrastructure for OpenAI. In September, OpenAI said it had a partnership with Oracle worth more than $300 billion over the next five years.

“We have a good relationship with OpenAI,” Clay Magouyrk, one of Oracle’s two newly appointed CEOs, said at an October analyst meeting.

Doing business with OpenAI is relatively new to 48-year-old Oracle. Historically, Oracle grew through sales of its database software and business applications. Its cloud infrastructure business now contributes over one-fourth of revenue, although Oracle remains a smaller hyperscaler than Amazon, Microsoft and Google.

OpenAI has also made commitments to other companies as it looks to meet expected capacity needs.

In September, Nvidia said it had signed a letter of intent with OpenAI to deploy at least 10 gigawatts of Nvidia equipment for the San Francisco artificial intelligence startup. The first phase of that project is expected in the second half of 2026.

Nvidia and OpenAI said in a September statement that they “look forward to finalizing the details of this new phase of strategic partnership in the coming weeks.”

But no announcement has come yet.

In a November filing, Nvidia said “there is no assurance that we will enter into definitive agreements with respect to the OpenAI opportunity.”

OpenAI has historically relied on Nvidia graphics processing units to operate ChatGPT and other products, and now it’s also looking at designing custom chips in a collaboration with Broadcom.

On Thursday, Broadcom CEO Hock Tan laid out a timeline for the OpenAI work, which was announced in October. Broadcom and OpenAI said they had signed a term sheet.

“It’s more like 2027, 2028, 2029, 10 gigawatts, that was the OpenAI discussion,” Tan said on Broadcom’s earnings call. “And that’s, I call it, an agreement, an alignment of where we’re headed with respect to a very respected and valued customer, OpenAI. But we do not expect much in 2026.”

OpenAI declined to comment.

WATCH: Oracle says there have been ‘no delays’ in OpenAI arrangement after stock slide

Oracle says there have been 'no delays' in OpenAI arrangement after stock slide

Continue Reading

Technology

AI order from Trump might be ‘illegal,’ Democrats and consumer advocacy groups claim

Published

on

By

AI order from Trump might be ‘illegal,’ Democrats and consumer advocacy groups claim

“This is the wrong approach — and most likely illegal,” Sen. Amy Klobuchar, D-Minn., said in a post on X Thursday.

“We need a strong federal safety standard, but we should not remove the few protections Americans currently have from the downsides of AI,” Klobuchar said.

Trump’s executive order directs Attorney General Pam Bondi to create a task force to challenge state laws regulating AI.

The Commerce Department was also directed to identify “onerous” state regulations aimed at AI.

The order is a win for tech companies such as OpenAI and Google and the venture firm Andreessen Horowitz, which have all lobbied against state regulations they view as burdensome. 

It follows a push by some Republicans in Congress to impose a moratorium on state AI laws. A recent plan to tack on that moratorium to the National Defense Authorization Act was scuttled.

Collin McCune, head of government affairs at Andreessen Horowitz, celebrated Trump’s order, calling it “an important first step” to boost American competition and innovation. But McCune urged Congress to codify a national AI framework.

“States have an important role in addressing harms and protecting people, but they can’t provide the long-term clarity or national direction that only Congress can deliver,” McCune said in a statement.

Sriram Krishnan, a White House AI advisor and former general partner at Andreessen Horowitz, during an interview Friday on CNBC’s “Squawk Box,” said that Trump is was looking to partner with Congress to pass such legislation.

“The White House is now taking a firm stance where we want to push back on ‘doomer’ laws that exist in a bunch of states around the country,” Krishnan said.

He also said that the goal of the executive order is to give the White House tools to go after state laws that it believes make America less competitive, such as recently passed legislation in Democratic-led states like California and Colorado.

The White House will not use the executive order to target state laws that protect the safety of children, Krishnan said.

Robert Weissman, co-president of the consumer advocacy group Public Citizen, called Trump’s order “mostly bluster” and said the president “cannot unilaterally preempt state law.”

“We expect the EO to be challenged in court and defeated,” Weissman said in a statement. “In the meantime, states should continue their efforts to protect their residents from the mounting dangers of unregulated AI.”

Weissman said about the order, “This reward to Big Tech is a disgraceful invitation to reckless behavior
by the world’s largest corporations and a complete override of the federalist principles that Trump and MAGA claim to venerate.”

In the short term, the order could affect a handful of states that have already passed legislation targeting AI. The order says that states whose laws are considered onerous could lose federal funding.

One Colorado law, set to take effect in June, will require AI developers to protect consumers from reasonably foreseeable risks of algorithmic discrimination.

Some say Trump’s order will have no real impact on that law or other state regulations.

“I’m pretty much ignoring it, because an executive order cannot tell a state what to do,” said Colorado state Rep. Brianna Titone, a Democrat who co-sponsored the anti-discrimination law.

In California, Gov. Gavin Newsom recently signed a law that, starting in January, will require major AI companies to publicly disclose their safety protocols. 

That law’s author, state Sen. Scott Wiener, said that Trump’s stated goal of having the United States dominate the AI sector is undercut by his recent moves. 

“Of course, he just authorized chip sales to China & Saudi Arabia: the exact opposite of ensuring U.S. dominance,” Wiener wrote in an X post on Thursday night. The Bay Area Democrat is seeking to succeed Speaker-emerita Nancy Pelosi in the U.S. House of Representatives.

Trump on Monday said he will Nvidia to sell its advanced H200 chips to “approved customers” in China, provided that U.S. gets a 25% cut of revenues.

Continue Reading

Trending