Connect with us

Published

on

David Silver, leader of the reinforcement learning research group at DeepMind, being awarded an honorary “ninth dan” professional ranking for AlphaGo.
JUNG YEON-JE | AFP | Getty Images

Computer scientists are questioning whether DeepMind, the Alphabet-owned U.K. firm that’s widely regarded as one of the world’s premier AI labs, will ever be able to make machines with the kind of “general” intelligence seen in humans and animals.

In its quest for artificial general intelligence, which is sometimes called human-level AI, DeepMind is focusing a chunk of its efforts on an approach called “reinforcement learning.”

This involves programming an AI to take certain actions in order to maximize its chance of earning a reward in a certain situation. In other words, the algorithm “learns” to complete a task by seeking out these preprogrammed rewards. The technique has been successfully used to train AI models how to play (and excel at) games like Go and chess. But they remain relatively dumb, or “narrow.” DeepMind’s famous AlphaGo AI can’t draw a stickman or tell the difference between a cat and a rabbit, for example, while a seven-year-old can.

Despite this, DeepMind, which was acquired by Google in 2014 for around $600 million, believes that AI systems underpinned by reinforcement learning could theoretically grow and learn so much that they break the theoretical barrier to AGI without any new technological developments.

Researchers at the company, which has grown to around 1,000 people under Alphabet’s ownership, argued in a paper submitted to the peer-reviewed Artificial Intelligence journal last month that “Reward is enough” to reach general AI. The paper was first reported by VentureBeat last week.

In the paper, the researchers claim that if you keep “rewarding” an algorithm each time it does something you want it to, which is the essence of reinforcement learning, then it will eventually start to show signs of general intelligence.

“Reward is enough to drive behavior that exhibits abilities studied in natural and artificial intelligence, including knowledge, learning, perception, social intelligence, language, generalization and imitation,” the authors write.

“We suggest that agents that learn through trial and error experience to maximize reward could learn behavior that exhibits most if not all of these abilities, and therefore that powerful reinforcement learning agents could constitute a solution to artificial general intelligence.”

Not everyone is convinced, however.

Samim Winiger, an AI researcher in Berlin, told CNBC that DeepMind’s “reward is enough” view is a “somewhat fringe philosophical position, misleadingly presented as hard science.”

He said the path to general AI is complex and that the scientific community is aware that there are countless challenges and known unknowns that “rightfully instill a sense of humility” in most researchers in the field and prevent them from making “grandiose, totalitarian statements” such as “RL is the final answer, all you need is reward.”

DeepMind told CNBC that while reinforcement learning has been behind some of its most well-known research breakthroughs, the AI technique accounts for only a fraction of the overall research it carries out. The company said it thinks it’s important to understand things at a more fundamental level, which is why it pursues other areas such as “symbolic AI” and “population-based training.”

“In somewhat typical DeepMind fashion, they chose to make bold statements that grabs attention at all costs, over a more nuanced approach,” said Winiger. “This is more akin to politics than science.”

Stephen Merity, an independent AI researcher, told CNBC that there’s “a difference between theory and practice.” He also noted that “a stack of dynamite is likely enough to get one to the moon, but it’s not really practical.”

Ultimately, there’s no proof either way to say whether reinforcement learning will ever lead to AGI.

Rodolfo Rosini, a tech investor and entrepreneur with a focus on AI, told CNBC: “The truth is nobody knows and that DeepMind’s main product continues to be PR and not technical innovation or products.”

Entrepreneur William Tunstall-Pedoe, who sold his Siri-like app Evi to Amazon, told CNBC that even if the researchers are correct “that doesn’t mean we will get there soon, nor does it mean that there isn’t a better, faster way to get there.”

DeepMind’s “Reward is enough” paper was co-authored by DeepMind heavyweights Richard Sutton and David Silver, who met DeepMind CEO Demis Hassabis at the University of Cambridge in the 1990s.

“The key problem with the thesis put forth by ‘Reward is enough’ is not that it is wrong, but rather that it cannot be wrong, and thus fails to satisfy Karl Popper’s famous criterion that all scientific hypotheses be falsifiable,” said a senior AI researcher at a large U.S. tech firm, who wished to remain anonymous due to the sensitive nature of the discussion.

“Because Silver et al. are speaking in generalities, and the notion of reward is suitably underspecified, you can always either cherry pick cases where the hypothesis is satisfied, or the notion of reward can be shifted such that it is satisfied,” the source added.

“As such, the unfortunate verdict here is not that these prominent members of our research community have erred in any way, but rather that what is written is trivial. What is learned from this paper, in the end? In the absence of practical, actionable consequences from recognizing the unalienable truth of this hypothesis, was this paper enough?”

What is AGI?

While AGI is often referred to as the holy grail of the AI community, there’s no consensus on what AGI actually is. One definition is it’s the ability of an intelligent agent to understand or learn any intellectual task that a human being can.

But not everyone agrees with that and some question whether AGI will ever exist. Others are terrified about its potential impacts and whether AGI would build its own, even more powerful, forms of AI, or so-called superintelligences.

Ian Hogarth, an entrepreneur turned angel investor, told CNBC that he hopes reinforcement learning isn’t enough to reach AGI. “The more that existing techniques can scale up to reach AGI, the less time we have to prepare AI safety efforts and the lower the chance that things go well for our species,” he said.

Winiger argues that we’re no closer to AGI today than we were several decades ago. “The only thing that has fundamentally changed since the 1950/60s, is that science-fiction is now a valid tool for giant corporations to confuse and mislead the public, journalists and shareholders,” he said.

Fueled with hundreds of millions of dollars from Alphabet every year, DeepMind is competing with the likes of Facebook and OpenAI to hire the brightest people in the field as it looks to develop AGI. “This invention could help society find answers to some of the world’s most pressing and fundamental scientific challenges,” DeepMind writes on its website.

DeepMind COO Lila Ibrahim said on Monday that trying to “figure out how to operationalize the vision” has been the biggest challenge since she joined the company in April 2018.

Continue Reading

Technology

A little-known startup just used AI to make a moon dust battery for Blue Origin

Published

on

By

A little-known startup just used AI to make a moon dust battery for Blue Origin

Istari Digital CEO Will Roper talks about the AI technology that built the Blue Origin moon vacuum

Artificial intelligence has created a device that turns moon dust into energy.

The moon vacuum, which was unveiled on Wednesday by Blue Origin at Amazon‘s re:Invent 2025 conference in Las Vegas, was built using critical technology from startup Istari Digital.

“So what it does is sucks up moon dust and it extracts the heat from it so it can be used as an energy source, like turning moon dust into a battery,” Istari CEO Will Roper told CNBC’s Morgan Brennan.

Spacecraft carrying out missions on the lunar surface are typically constrained by lunar night, the two-week period every 28 days during which the moon is cast in darkness and temperatures experience extreme drops, crippling hardware and rendering it useless unless a strong, long-lasting power source is present.

“Kind of like vacuuming at home, but creating your own electricity while you do it,” he added.

The battery was completely designed by AI, said Roper, who was assistant secretary of the Air Force under President Donald Trump‘s first term and is known for transforming the acquisition process at both the Air Force and, at the time, the newly created Space Force.

Read more CNBC tech news

A major part of the breakthrough in Istari’s technology is the way in which it handles and limits AI hallucinations.

Roper said the platform takes all the requirements a part needs and creates guardrails or a “fence around the playground” that the AI can’t leave while coming up with designs.

“Within that playground, AI can generate to its heart’s content,” he said.

“In the case of Blue Origin’s moon battery, [it] doesn’t tell you the design was a good one, but it tells us that all of the requirements were met, the standards were met, things like that that you got to check before you go operational,” he added.

Istari is backed by former Google CEO Eric Schmidt and already works with the U.S. government, including as a prime contractor with Lockheed Martin on the experimental x-56A unmanned aircraft.

Watch the full interview above and go deeper into the business of the stars with the Manifest Space podcast.

X-Energy’s Kam Ghaffarian on Nuclear Power, AI, and the Space Tech Race

Continue Reading

Technology

Nvidia CEO Jensen Huang talks chip restrictions with Trump, blasts state-by-state AI regulations

Published

on

By

Nvidia CEO Jensen Huang talks chip restrictions with Trump, blasts state-by-state AI regulations

Jensen Huang: State-by-state AI regulation would drag industry to a halt

Nvidia CEO Jensen Huang said he met with President Donald Trump on Wednesday and that the two men discussed chip export restrictions, as lawmakers consider a proposal to limit exports of advanced artificial intelligence chips to nations like China.

“I’ve said it repeatedly that we support export controls, and that we should ensure that American companies have the best and the most and first,” Huang told reporters on Capitol Hill.

Lawmakers were considering including the Guaranteeing Access and Innovation for National Artificial Intelligence Act in a major defense package, known as the National Defense Authorization Act. The GAIN AI Act would require chipmakers like Nvidia and Advanced Micro Devices to give U.S. companies first pick on their AI chips before selling them in countries like China.

The proposal isn’t expected to be part of the NDAA, Bloomberg reported, citing a person familiar with the matter.

Huang said it was “wise” that the proposal is being left out of the annual defense policy bill.

“The GAIN AI Act is even more detrimental to the United States than the AI Diffusion Act,” Huang said.

Nvidia’s CEO also criticized the idea of establishing a patchwork of state laws regulating AI. The notion of state-by-state regulation has generated pushback from tech companies and spurred the creation of a super PAC called “Leading the Future,” which is backed by the AI industry.

“State-by-state AI regulation would drag this industry into a halt and it would create a national security concern, as we need to make sure that the United States advances AI technology as quickly as possible,” Huang said. “A federal AI regulation is the wisest.”

Trump last month urged legislators to include a provision in the NDAA that would preempt state AI laws in favor of “one federal standard.”

But House Majority Leader Steve Scalise (R-LA) told CNBC’s Emily Wilkins on Tuesday the provision won’t make it into the bill, citing a lack of sufficient support. He and other lawmakers will continue to look for ways to establish a national standard on AI, Scalise added.

WATCH: Nvidia currying favor to be able to sell chips in China

Nvidia obviously currying favor to be able to sell chips in China, says Niles Investment's Dan Niles

Continue Reading

Technology

Design executive behind ‘Liquid Glass’ is leaving Apple

Published

on

By

Design executive behind 'Liquid Glass' is leaving Apple

File: Then Apple Creative Director Alan Dye celebrates the launch of the July Issue at the new WIRED office on June 24, 2015 in San Francisco, California.

Kimberly White | Getty Images

Apple‘s head of user interface design, Alan Dye, will join Meta, in a notable shift of executive talent in Silicon Valley.

The iPhone maker confirmed Dye’s departure on Wednesday and Apple CEO Tim Cook said in a statement that the company prioritizes design and has a strong team. The statement said that veteran designer Stephen Lemay will succeed Dye.

“Steve Lemay has played a key role in the design of every major Apple interface since 1999,” Cook said in a statement.

Meta CEO Mark Zuckerberg in a Wednesday social media post said that Dye would lead up a new creative studio that brings together design, fashion and technology.

“We plan to elevate design within Meta,” wrote Zuckerberg, who did not say what specific products Dye will work on.

Compared to other Silicon Valley companies, Apple has always emphasized design to customers and investors as one of its strengths. Apple prominently features its design executives to discuss interface changes at the company’s launch events.

In June, Dye revealed a redesign of Apple’s software interface for iPhones, Macs and the Apple Watch called Liquid Glass. The company described it as an “elegant” new design with translucent buttons, updated app icons and fluid animations.

Dye said it was the “next chapter” of the company’s software and said it “sets the stage” for the next era of Apple products.

“Our new design blurs the lines between hardware and software to create an experience that’s more delightful than ever while still familiar and easy to use,” Dye said at the launch.

Reviews were mixed on the Liquid Glass update, which shipped with new iPhones in September.

Apple announces liquid glass during the Apple Worldwide Developers Conference (WWDC) on June 9, 2025 in Cupertino, California.

Justin Sullivan | Getty Images

For years, Apple design was embodied by executive Jony Ive, who left Apple in 2019 and is now working with OpenAI on artificial intelligence hardware alongside Sam Altman.

Dye took over user interface design and became one of the design studio’s leads in 2015 when Ive stepped back from a day-to-day role. Dye started at Apple in 2006 and worked on software for the iPhone, iPad, Mac, Apple Watch, Apple TV and Vision Pro, according to his LinkedIn profile.

He was also partly responsible for the first iPhone in 2017 that did away with the home screen button at the bottom of the device and replaced it with a software-based swipe-up motion.

Meta has said in recent years that it wants to be a major developer of hardware and Zuckerberg has said Apple is one of his company’s biggest competitors.

The social media company currently makes several virtual reality headsets under its Quest brand, and recently scored its first hardware hit with Ray-Ban Meta smart glasses, which are stylish sunglasses equipped with cameras and the ability to run an AI model that can answer questions. Sales of the device tripled over the past year, Ray-Ban parent company EssilorLuxottica said in July.

“We’re entering a new era where AI glasses and other devices will change how we connect with technology and each other,” Zuckerberg wrote.

Bloomberg first reported the move.

WATCH: Apple AI chief steps down: Here’s what to know

Apple AI chief steps down: Here's what to know

Continue Reading

Trending