Alphabet has cut contractual ties with Appen, the artificial intelligence data firm that helped train Google’s chatbot Bard, Google Search results and other AI products.
After a “strategic review process,” Alphabet notified Appen over the weekend of the termination, which will go into effect March 19, according to a filing from Appen. The company said it had “no prior knowledge of Google’s decision to terminate the contract.”
Alphabet accounted for roughly one-third of Appen’s revenue, meaning the decision to end the relationship will impact “at least two thousand subcontracted Alphabet workers,” according to a statement Monday from the Alphabet Workers Union.
Appen, based in Australia, has helped train AI models for a star-studded list of tech behemoths. Five customers — Microsoft, Apple, Meta, Google and Amazon — have in the past accounted for 80% of Appen’s revenue. Appen has a platform of about 1 million freelance workers in more than 170 countries.
In 2023, revenue from work with Alphabet totaled $82.8 million of Appen’s $273 million in sales for the year, according to Monday’s filing.
Despite Appen’s enviable client list and its nearly 30-year history, the company has struggled in recent years with a loss of customers, a string of executive departures and plummeting financials — even as generative AI tools increased demand for training data. Revenue dropped 30% in 2023, after declining 13% a year earlier, which the company attributed in part to “challenging external operating and macro conditions.”
In August 2020, Appen’s shares peaked at AU$42.44 ($27.08) on the Australian Securities Exchange, sending its market cap to the equivalent of $4.3 billion. Now, the stock is trading at around AU$0.28, down more than 99% since its peak.
Former employees, who asked not to be named for fear of retaliation, told CNBC in September that the company’s current struggle to pivot to generative AI reflects years of weak quality controls and a disjointed organizational structure.
Appen’s past work for tech companies has been on projects like evaluating the relevance of search results, helping AI assistants understand requests in different accents, categorizing e-commerce images using AI, and building out map locations of electric vehicle charging stations, according to public information and interviews conducted by CNBC.
Appen has also touted its work on search relevance for Adobe and on translation services for Microsoft, as well as in providing training data for lidar companies, security applications and automotive manufacturers.
But large language models of today operate differently. The underlying LLMs behind OpenAI’s ChatGPT and Google’s Bard are scouring the digital universe to provide sophisticated answers and advanced images in response to simple text queries. Companies are spending far more on processors from Nvidia and less on Appen.
Google and Appen have had conflicts in the past, namely a dispute about wages. In 2019, Google said its contractors would need to pay their workers $15 an hour. Appen didn’t meet that requirement, according to public letters written by some workers.
In January 2023, after months of organizing, raises went into effect for Appen freelancers working on the Bard chatbot and other Google products. The rates went up to between $14 and $14.50 per hour.
But labor issues persisted. In June, Appen faced charges from the U.S. National Labor Relations Board after allegedly firing six freelancers who spoke out publicly about frustrations with workplace conditions. The workers were later reinstated.
Appen wrote in Monday’s filing that it will focus on managing costs, turning the business around and providing customers with quality AI data.
“Appen will immediately adjust its strategic priorities following the notification of the Google contract termination and provide further details in its FY23 full year results on 27 February 2024,” the company wrote.
When Meta CEO Mark Zuckerbergpoached Scale AI founder Alexandr Wang last week as part of a $14.3 billion investment in the artificial intelligence startup, he was apparently just getting started.
Zuckerberg’s multibillion-dollar AI hiring spree has now turned to Daniel Gross, the CEO of Ilya Sutskever’s startup Safe Superintelligence, and former GitHub CEO Nat Friedman, according to sources with knowledge of the matter.
It’s not how Zuckerberg planned for a deal to go down.
Earlier this year, sources said, Meta tried to acquire Safe Superintelligence, which was reportedly valued at $32 billion in a fundraising round in April. Sutskever, who just launched the startup a year ago, shortly after leaving OpenAI, rebuffed Meta’s efforts, as well as the company’s attempt to hire him, said the sources, who asked not to be named because the information is confidential.
Soon after those talks ended, Zuckerberg started negotiating with Gross, the sources said. In addition to his role at Safe Superintelligence, Gross runs a venture capital firm with Friedman called NFDG, their combined initials.
Both men are joining Meta as part of the transaction, and will work on products under Wang, one source said. Meta, meanwhile, will get a stake in NFDG, according to multiple sources.
The Information was first to report on Meta’s plans to hire Gross and Friedman.
Gross, Friedman and Sutskever didn’t respond to CNBC’s requests for comment.
A Meta spokesperson said the company “will share more about our superintelligence effort and the great people joining this team in the coming weeks.”
Zuckerberg’s aggressive hiring tactics escalate an AI talent war that’s reached new heights of late. Meta, Google and OpenAI, along with a host of other big companies and high-valued startups, are racing to develop the most powerful large language models, and pushing towards artificial general intelligence (AGI), or AI that’s considered equal to or greater than human intelligence.
Last week, Meta agreed to pump $14.3 billion into Scale AI to bring on Wang and a few other top engineers while getting a 49% stake in the startup.
Altman said on the latest episode of the “Uncapped” podcast, which is hosted by his brother, that Meta has tried to lure OpenAI employees by offering signing bonuses as high as $100 million, with even larger annual compensation packages. Altman said “none of our best people have decided to take them up on that.”
“I’ve heard that Meta thinks of us as their biggest competitor,” Altman said on the podcast. “Their current AI efforts have not worked as well as they have hoped and I respect being aggressive and continuing to try new things.”
Meta didn’t respond to a request for comment on Altman’s remarks.
OpenAI, for its part, has gone to similar lengths, paying about $6.5 billion to hire iPhone designer Jony Ive and to acquire his nascent devices startup io.
Elsewhere, the founders of AI startup Character.AI were recruited back to Google last year in a multibillion-dollar deal, while DeepMind co-founder Mustafa Suleyman was brought on by Microsoft in a $650 million purchase of talent from Inflection AI.
In Gross, Zuckerberg is getting a longtime entrepreneur and AI investor. Gross founded the search engine Cue, which was acquired by Apple in 2013. He was a top executive at Apple and helped lead machine learning efforts and the development of Siri. He was later a partner at startup accelerator Y Combinator, before co‑founding Safe Superintelligence alongside Sutskever.
Friedman co-founded two startups before becoming the CEO of GitHub following Microsoft’s acquisition of the code-sharing platform in 2018.
NFDG has backed Coinbase, Figma, CoreWeave, Perplexity and Character.ai over the years, according to Pitchbook. It’s unclear what happens to its investment portfolio in a Meta deal, a source said.
Silhouettes of laptop and mobile device users are seen next to a screen projection of the YouTube logo.
Dado Ruvic | Reuters
Google is using its expansive library of YouTube videos to train its artificial intelligence models, including Gemini and the Veo 3 video and audio generator, CNBC has learned.
The tech company is turning to its catalog of 20 billion YouTube videos to train these new-age AI tools, according to a person who was not authorized to speak publicly about the matter. Google confirmed to CNBC that it relies on its vault of YouTube videos to train its AI models, but the company said it only uses a subset of its videos for the training and that it honors specific agreements with creators and media companies.
“We’ve always used YouTube content to make our products better, and this hasn’t changed with the advent of AI,” said a YouTube spokesperson in a statement. “We also recognize the need for guardrails, which is why we’ve invested in robust protections that allow creators to protect their image and likeness in the AI era — something we’re committed to continuing.”
Such use of YouTube videos has the potential to lead to an intellectual property crisis for creators and media companies, experts said.
While YouTube says it has shared this information previously, experts who spoke with CNBC said it’s not widely understood by creators and media organizations that Google is training its AI models using its video library.
YouTube didn’t say how many of the 20 billion videos on its platform or which ones are used for AI training. But given the platform’s scale, training on just 1% of the catalog would amount to 2.3 billion minutes of content, which experts say is more than 40 times the training data used by competing AI models.
The company shared in a blog post published in September that YouTube content could be used to “improve the product experience … including through machine learning and AI applications.” Users who have uploaded content to the service have no way of opting out of letting Google train on their videos.
“It’s plausible that they’re taking data from a lot of creators that have spent a lot of time and energy and their own thought to put into these videos,” said Luke Arrigoni, CEO of Loti, a company that works to protect digital identity for creators. “It’s helping the Veo 3 model make a synthetic version, a poor facsimile, of these creators. That’s not necessarily fair to them.”
CNBC spoke with multiple leading creators and IP professionals, none were aware or had been informed by YouTube that their content could be used to train Google’s AI models.
Google DeepMind Veo 3.
Courtesy: Google DeepMind
The revelation that YouTube is training on its users’ videos is noteworthy after Google in May announced Veo 3, one of the most advanced AI video generators on the market. In its unveiling, Google showcased cinematic-level video sequences, including a scene of an old man on a boat and another showing Pixar-like animals talking with one another. The entirety of the scenes, both the visual and the audio, were entirely AI generated.
According to YouTube, an average of 20 million videos are uploaded to the platform each day by independent creators by nearly every major media company. Many creators say they are now concerned they may be unknowingly helping to train a system that could eventually compete with or replace them.
“It doesn’t hurt their competitive advantage at all to tell people what kind of videos they train on and how many they trained on,” Arrigoni said. “The only thing that it would really impact would be their relationship to creators.”
Even if Veo 3’s final output does not directly replicate existing work, the generated content fuels commercial tools that could compete with the creators who made the training data possible, all without credit, consent or compensation, experts said.
When uploading a video to the platform, the user is agreeing that YouTube has a broad license to the content.
“By providing Content to the Service, you grant to YouTube a worldwide, non-exclusive, royalty-free, sublicensable and transferable license to use that Content,” the terms of service read.
“We’ve seen a growing number of creators discover fake versions of themselves circulating across platforms — new tools like Veo 3 are only going to accelerate the trend,” said Dan Neely, CEO of Vermillio, which helps individuals protect their likeness from being misused and also facilitates secure licensing of authorized content.
Neely’s company has challenged AI platforms for generating content that allegedly infringes on its clients’ intellectual property, both individual and corporate. Neely says that although YouTube has the right to use this content, many of the content creators who post on the platform are unaware that their videos are being used to train video-generating AI software.
Vermillio uses a proprietary tool called Trace ID to asses whether an AI-generated video has significant overlap with a human-created video. Trace ID assigns scores on a scale of zero to 100. Any score over 10 for a video with audio is considered meaningful, Neely said.
A video from YouTube creator Brodie Moss closely matched content generated by Veo 3. Using Vermillio’s Trace ID tool, the system attributed a score of 71 to the original video with the audio alone scoring over 90.
Some creators told CNBC they welcome the opportunity to use Veo 3, even if it may have been trained on their content.
“I try to treat it as friendly competition more so than these are adversaries,” said Sam Beres, a creator with 10 million subscribers on YouTube. “I’m trying to do things positively because it is the inevitable —but it’s kind of an exciting inevitable.”
Google includes an indemnification clause for its generative AI products, including Veo, which means that if a user faces a copyright challenge over AI-generated content, Google will take on legal responsibility and cover the associated costs.
YouTube announced a partnership with Creative Artists Agency in December to develop access for top talent to identify and manage AI-generated content that features their likeness. YouTube also has a tool for creators to request a video to be taken down if they believe it abuses their likeness.
However, Arrigoni said that the tool hasn’t been reliable for his clients.
YouTube also allows creators to opt out of third party training from select AI companies including Amazon, Apple and Nvidia, but users are not able to stop Google from training for its own models.
The Walt Disney Company and Universal filed a joint lawsuit last Wednesday against the AI image generator Midjourney, alleging copyright infringement, the first lawsuit of its kind out of Hollywood.
“The people who are losing are the artists and the creators and the teenagers whose lives are upended,” said Sen. Josh Hawley, R-Mo., in May at a Senate hearing about the use of AI to replicate the likeness of humans. “We’ve got to give individuals powerful enforceable rights and their images in their property in their lives back again or this is just never going to stop.”
Disclosure: Universal is part of NBCUniversal, the parent company of CNBC.
Samsung launched the Galaxy Z Fold6 at its Galaxy Unpacked event in Paris. The tech giant said the foldable device is thinner and lighter than its predecessor.
Arjun Kharpal | CNBC
Samsung will unveil a thinner version of its flagship foldable smartphone at a launch likely set to take place next month, as it battles Chinese rivals to deliver the slimmest devices to the market.
Folding phones, which have a single screen that can fold in half, came in focus when Samsung first launched such a device in 2019. But Chinese players, in particular Honor and Oppo, have since aggressively released foldables that are thinner and lighter than Samsung’s offerings.
Why are slim foldables important?
“With foldables, thinness has become more critical than ever because people aren’t prepared to accept the compromise for a thicker and heavier phone to get the real estate that a folding phone can deliver,” Ben Wood, chief analyst at CCS Insight, told CNBC on Thursday.
Honor, Oppo and other Chinese players have used their slim designs to differentiate themselves from Samsung.
Let’s look at a comparison: Samsung’s last foldable from 2024, the Galaxy Z Fold6, is 12.1 millimeter ~(0.48 inches) thick when folded and weighs 239 grams (8.43 oz). Oppo’s Find N5, which was released earlier this year, is 8.93 millimeters thick when closed and weighs 229 grams. The Honor Magic V3, which was launched last year, is 9.2 millimeters when folded and weighs 226 grams.
“Samsung needs to step up” in foldables, Wood said.
And that’s what the South Korean tech giant is planning to do at its upcoming launch, which is likely to take place next month.
“The newest Galaxy Z series is the thinnest, lightest and most advanced foldable yet – meticulously crafted and built to last,” Samsung said in a preview blog post about the phone earlier this month.
But the competition is not letting up. Honor is planning a launch on July 2 in China for its latest folding phone, the Magic V5.
“The interesting thing for Samsung, if they can approach the thinness that Honor has achieved it is will be a significant step up from predecessor, it will be a tangible step up in design,” Wood said.
Despite these advances by way of foldables, the market for the devices has not been as exciting as many had hoped.
CCS Insight said that foldables will account for just 2% of the overall smartphone market this year. Thinner phones may be one way to address the sluggish market, but consumer preferences would also need to change.
“There is a chance that by delivering much thinner foldables that are more akin to the traditional monoblock phone, it will provide an opportunity to turn consumer heads and get them to revisit the idea of having a folding device,” Wood said.
“However, I would caution foldables do remain problematic because in many cases consumers struggle to see why they need a folding device.”
Although the market remains small for foldables compared to traditional smartphones, noted analyst Ming-Chi Kuo of TF International Securities on Wednesday said Apple — which has been notably absent from this product line-up — plans to make a folding iPhone starting next year.