Connect with us

Published

on

Dario Amodei, Anthropic CEO, speaking on CNBC’s Squawk Box outside the World Economic Forum in Davos, Switzerland on Jan. 21st, 2025.

Gerry Miller | CNBC

Anthropic on Monday announced Claude for Life Sciences, a new offering for researchers to use the company’s artificial intelligence technology in the advancement of scientific discovery. 

Claude for Life Sciences is built around Anthropic’s existing AI models, but supports new connections with other scientific tools that are commonly used in labs during research and development.

It will be able to help researchers through all stages of the discovery process, from carrying out literature reviews to developing hypotheses, analyzing data, drafting regulatory submissions and more, Anthropic said.

The launch of Claude for Life Sciences marks Anthropic’s first formal entry into the sector, and comes just months after the company hired longtime industry executive Eric Kauderer-Abrams as its head of biology and life sciences. 

“Now is the threshold moment for us where we’ve decided this is a big investment area,” Kauderer-Abrams told CNBC in an interview. “We want a meaningful percentage of all of the life science work in the world to run on Claude, in the same way that that happens today with coding.”

Read more CNBC tech news

Anthropic, which is one of the companies at the center of the AI boom, develops a family of large language models called Claude. It was founded in 2021 by a group of former OpenAI executives and researchers, and its valuation has swelled to $183 billion in just four years.

The company launched a new model, Claude Sonnet 4.5, late last month and said it is “significantly better” at life sciences tasks like understanding laboratory protocols.  

Kauderer-Abrams said researchers have already been engaging with Anthropic’s models to help with isolated parts of the scientific process, so the company decided to formally build out Claude for Life Sciences as a way to support them from start to finish. 

That meant Anthropic had to establish integrations with key players in the life sciences ecosystem, including Benchling, PubMed, 10x Genomics and Synapse.org, among others. Anthropic has also partnered with companies that can help life sciences organizations adopt AI, like Caylent, KPMG, Deloitte, and cloud providers AWS and Google Cloud, the company said.

“We’re willing and enthusiastic about doing that grind to make sure that all the pieces come together,” Kauderer-Abrams said.

In a prerecorded demo, Anthropic showed how a scientist working on preclinical studies could use Claude for Life Sciences to compare two study designs that test different dosing strategies. 

The scientist was able to query her lab’s data directly from Benchling, generate a summary and tables of key differences with links back to the original material. After reviewing the results, the scientist generated a study report that could be included in a regulatory submission. 

Anthropic said an analysis like this used to require “days” of validating and compiling information, but now, it can be done in minutes. 

Kauderer-Abrams said the company believes AI can bring about real efficiency gains for the life sciences sector, but it’s also under “no illusions” that it will magically overcome the physical limitations of conducting scientific research. Clinical trials that take three years are not suddenly going to take one month, he said.

Instead, Anthropic is focused on exploring the time-consuming, expensive parts of the discovery process “piece by piece” to determine where AI could be most useful.

“We’re here to make sure that this transformation happens and that it’s done responsibly,” Kauderer-Abrams said.

WATCH: Anthropic launches Claude Sonnet 4.5, its latest AI model

Anthropic launches Claude Sonnet 4.5, its latest AI model

Continue Reading

Technology

OpenAI investor Reid Hoffman spars with AI czar Sacks, calls Anthropic ‘one of the good guys’

Published

on

By

OpenAI investor Reid Hoffman spars with AI czar Sacks, calls Anthropic 'one of the good guys'

Reid Hoffman, Partner at Greylock and co-founder LinkedIn, speaks during the WSJ Tech Live conference hosted by the Wall Street Journal at the Montage Laguna Beach in Laguna Beach, California, on October 21, 2024.

Frederic J. Brown | Afp | Getty Images

Two of the main members of the PayPal mafia are sparring again — this time over artificial intelligence.

Billionaire tech investor and LinkedIn co-founder Reid Hoffman on Monday called Anthropic “one of the good guys” after the AI startup was criticized last week by David Sacks, the venture capitalist serving as President Donald Trump’s AI and crypto czar.

“Anthropic, along with some others (incl Microsoft, Google, and OpenAI) are trying to deploy AI the right way, thoughtfully, safely, and enormously beneficial for society,” Hoffman wrote on X. That’s why I am intensely rooting for their success.”

Hoffman has served on Microsoft’s board since 2017, shortly after selling LinkedIn to the software giant. Microsoft is a key OpenAI investor and partner. Hoffman was also an early investor in OpenAI, Anthropic’s larger rival, and remains a shareholder. He revealed on Monday that Greylock, where he’s a partner, has invested in Anthropic.

Greylock and Anthropic didn’t respond to requests for comment.

In a series of posts, Hoffman said he tries to avoid commenting directly about companies like OpenAI and Anthropic, but that “in all industries, especially in AI, it’s important to back the good guys.”

Read more CNBC tech news

Hoffman and Sacks were both early employees at PayPal, joining in 1999 and assuming major roles at the payments company. Along with Peter Thiel, Elon Musk, Max Levchin and a group of other high-profile techies, they were part of what became known as the PayPal mafia because of the number of successful companies they went on to build.

But Hoffman and Sacks have been public antagonists recently, due mostly to their political differences. Hoffman is a major Democratic donor, contributing millions of dollars to Kamala Harris’ unsuccessful presidential bid.

Sacks emerged as a vocal Trump supporter ahead of the 2024 election before joining the administration. He hosted a fundraiser for Trump at his San Francisco mansion.

Politics of AI

David Sacks, U.S. President Donald Trump’s “AI and Crypto Czar”, speaks to President Trump as he signs a series of executive orders in the Oval Office of the White House on Jan. 23, 2025 in Washington, DC.

Anna Moneymaker | Getty Images

Sacks criticized the essay and, in a post on X, accused Anthropic of “running a sophisticated regulatory capture strategy based on fear-mongering.” He said the company is “principally responsible for the state regulatory frenzy that is damaging the startup ecosystem.”

Anthropic has repeatedly pushed back against efforts by the federal government to hinder state-level regulation of AI, including a Trump-backed provision that would have blocked those rules for 10 years.

After Hoffman shared his thoughts about Anthropic on Monday, Sacks and Musk, who owns a competing AI company called xAI and was also a major early figure in the second Trump administration, were quick to respond.

“The leading funder of lawfare and dirty tricks against President Trump wants you to know that ‘Anthropic is one of the good guys,'” Sacks wrote in response to Hoffman on Monday. “Thanks for clarifying that. All we needed to know.”

“Indeed,” Musk said in a reply.

The chirping went back and forth on Monday.

“Shows you didn’t read the post (not shocked),” Hoffman wrote. “When you are ready to have a professional conversation about AI’s impact on America, I’m here to chat.”

Jason Calacanis, who co-hosts the All-In podcast, along with Sacks and two other tech friends, wrote in response to Hoffman that he should “come on the pod,” inviting him this week. Hoffman previously joined for an episode at the end of August, roughly two months before the presidential election.

Hoffman wrote that he is “open to coming back on” but that “this week is packed.”

— CNBC’s MacKenzie Sigalos contributed to this report

WATCH: Anthropic’s Mike Krieger on new model release and the race to build real-world AI agents

Anthropic’s Mike Krieger on new model release and the race to build real-world AI agents

Continue Reading

Technology

OpenAI cracks down on Sora 2 deepfakes after pressure from Bryan Cranston, SAG-AFTRA

Published

on

By

OpenAI cracks down on Sora 2 deepfakes after pressure from Bryan Cranston, SAG-AFTRA

Mike Windle/Getty Images

OpenAI announced on Monday in a joint statement that it will be working with Bryan Cranston, SAG-AFTRA, and other actor unions to protect against deepfakes on its artificial intelligence video creation app Sora.

The “Breaking Bad” and “Malcolm in the Middle” actor expressed concern after unauthorized AI-generated clips using his voice and likeness appeared on the app following the Sora 2 launch at the end of September, the Screen Actors Guild-American Federation of Television and Radio Artists said in a post on X.

“I am grateful to OpenAI for its policy and for improving its guardrails, and hope that they and all of the companies involved in this work, respect our personal and professional right to manage replication of our voice and likeness,” Cranston said in a statement.

Along with SAG-AFTRA, OpenAI said it will collaborate with United Talent Agency, which represents Cranston, the Association of Talent Agents and Creative Artists Agency to strengthen guardrails around unapproved AI generations.

The CAA and UTA previously slammed OpenAI for its usage of copyrighted materials, calling Sora a risk to their clients and intellectual property.

Read more CNBC tech news

OpenAI had to block videos of Martin Luther King Jr. on Sora last week at the request of King’s estate after users created “disrespectful depictions” of the civil rights leader.

Zelda Williams, the daughter for late comedian Robin Williams, asked people to stop sending her AI-generated videos of her father shortly after the Sora 2 release.

OpenAI’s approach to copyright restrictions and other issues related to likeness have evolved since the Sora 2 launch Sept. 30.

On Oct. 3, CEO Sam Altman updated Sora’s opt-out policy, which previously allowed the use of IP unless studios specifically requested that their material not be used, to allow rightsholders “more granular control over generation of characters.”

At launch, Sora required an opt-in for the use of an individual’s voice and likeness, though OpenAI said that it is now also committing to “responding expeditiously to any complaints it may receive.”

The company reiterated its support of the NO FAKES Act, a federal bill passed designed to protect against unauthorized AI-generated replicas of people’s voice or visual likeness.

“OpenAI is deeply committed to protecting performers from the misappropriation of their voice and likeness,” Altman said in a statement. “We were an early supporter of the NO FAKES Act when it was introduced last year, and will always stand behind the rights of performers.”

We tested OpenAI’s Sora 2 AI-video app to find out why Hollywood is worried

Continue Reading

Technology

Jim Cramer: Patient Apple bulls are vindicated, and the stock is just getting started

Published

on

By

Jim Cramer: Patient Apple bulls are vindicated, and the stock is just getting started

Continue Reading

Trending