Google has been facing a wave of litigation recently as the implications of generative artificial intelligence (AI) on copyright and privacy rights become clearer.
Amid the ever-intensifying debate, Google has not only defended its AI training practices but also pledged to shield users of its generative AI products from accusations of copyright violations.
However, Google’s protective umbrella only spans seven specified products with generative AI attributes and conspicuously leaves out Google’s Bard search tool. The move, although a solace to some, opens a Pandora’s box of questions around accountability, the protection of creative rights and the burgeoning field of AI.
Moreover, the initiative is also being perceived as more than just a mere reactive measure from Google, but rather a meticulously crafted strategy to indemnify the blossoming AI landscape.
AI’s legal cloud
The surge of generative AI over the last couple of years has rekindled the age-old flame of copyright debates with a modern twist. The bone of contention currently pivots around whether the data used to train AI models and the output generated by them violate propriety intellectual property (IP) affiliated with private entities.
In this regard, the accusations against Google consist of just this and, if proven, could not only cost Google a lot of money but also set a precedent that could throttle the growth of generative AI as a whole.
Google’s legal strategy, meticulously designed to instill confidence among its clientele, stands on two primary pillars, i.e., the indemnification of its training data and its generated output. To elaborate, Google has committed to bearing legal responsibility should the data employed to devise its AI models face allegations of IP violations.
Not only that, but the tech giant is also looking to protect users against claims that the text, images or other content engendered by its AI services do not infringe upon anyone else’s personal data — encapsulating a wide array of its services, including Google Docs, Slides and Cloud Vertex AI.
Google has argued that the utilization of publicly available information for training AI systems is not tantamount to stealing, invasion of privacy or copyright infringement.
However, this assertion is under severe scrutiny as a slew of lawsuits accuse Google of misusing personal and copyrighted information to feed its AI models. One of the proposed class-action lawsuits even alleges that Google has built its entire AI prowess on the back of secretly purloined data from millions of internet users.
Therefore, the legal battle seems to be more than just a confrontation between Google and the aggrieved parties; it underlines a much larger ideological conundrum, namely: “Who truly owns the data on the internet? And to what extent can this data be used to train AI models, especially when these models churn out commercially lucrative outputs?”
An artist’s perspective
The dynamic between generative AI and protecting intellectual property rights is a landscape that seems to be evolving rapidly.
Nonfungible token artist Amitra Sethi told Cointelegraph that Google’s recent announcement is a significant and welcome development, adding:
“Google’s policy, which extends legal protection to users who may face copyright infringement claims due to AI-generated content, reflects a growing awareness of the potential challenges posed by AI in the creative field.”
However, Sethi believes that it is important to have a nuanced understanding of this policy. While it acts as a shield against unintentional infringement, it might not cover all possible scenarios. In her view, the protective efficacy of the policy could hinge on the unique circumstances of each case.
When an AI-generated piece loosely mirrors an artist’s original work, Sethi believes the policy might offer some recourse. But in instances of “intentional plagiarism through AI,” the legal scenario could get murkier. Therefore, she believes that it is up to the artists themselves to remain proactive in ensuring the full protection of their creative output.
Sethi said that she recently copyrighted her unique art genre, “SoundBYTE,” so as to highlight the importance of artists taking active measures to secure their work. “By registering my copyright, I’ve established a clear legal claim to my creative expressions, making it easier to assert my rights if they are ever challenged,” she added.
In the wake of such developments, the global artist community seems to be coming together to raise awareness and advocate for clearer laws and regulations governing AI-generated content.
Tools like Glaze and Nightshade have also appeared to protect artists’ creations. Glaze applies minor changes to artwork that, while practically imperceptible to the human eye, feeds incorrect or bad data to AI art generators. Similarly, Nightshade lets artists add invisible changes to the pixels within their pieces, thereby “poisoning the data” for AI scrapers.
Examples of how “poisoned” artworks can produce an incorrect image from an AI query. Source: MIT
Industry-wide implications
The existing narrative is not limited to Google and its product suite. Other tech majors like Microsoft and Adobe have also made overtures to protect their clients against similar copyright claims.
Microsoft, for instance, has put forth a robust defense strategy to shield users of its generative AI tool, Copilot. Since its launch, the company has staunchly defended the legality of Copilot’s training data and its generated information, asserting that the system merely serves as a means for developers to write new code in a more efficient fashion.
Adobe has incorporated guidelines within its AI tools to ensure users are not unwittingly embroiled in copyright disputes and is also offering AI services bundled with legal assurances against any external infringements.
The inevitable court cases that will appear regarding AI will undoubtedly shape not only legal frameworks but also the ethical foundations upon which future AI systems will operate.
Tomi Fyrqvist, co-founder and chief financial officer for decentralized social app Phaver, told Cointelegraph that in the coming years, it would not be surprising to see more lawsuits of this nature coming to the fore:
“There is always going to be someone suing someone. Most likely, there will be a lot of lawsuits that are opportunistic, but some will be legit.”
Collect this article as an NFT to preserve this moment in history and show your support for independent journalism in the crypto space.
Sir Keir Starmer’s been on the other side of the world for most of the week – at the COP30 climate summit in Brazil, his 40th foreign trip in 16 months.
Back home, his government’s credibility has continued its painful unravelling.
Five days on from David Lammy’s disastrous stand-in performance at PMQS, the justice secretary’s ministerial colleagues are still struggling to explain why he repeatedly failed to answer questions on whether another migrant criminal had been released from prison by mistake.
Image: Prime Minister Sir Keir Starmer at the Remembrance Sunday service at the Cenotaph in London. Pic: PA
Yes, Conservative MP James Cartlidge got the question wrong, as Brahim Kaddour-Cherif was an illegal migrant, not an asylum seeker.
But Mr Cartlidge argued that because the deputy prime minister failed to divulge the information he did have, he failed to act with full transparency and should be investigated by the PM’s ethics advisor for a possible breach of the ministerial code.
Please use Chrome browser for a more accessible video player
10:17
Lammy not sharing facts is ‘shocking’
She told Sunday Morning with Trevor Phillips she doesn’t accept that he was being evasive, insisting Mr Lammy had been carefully weighing his words to ensure that “when we do speak about matters of such significance to the public… we do so with care and make sure the full facts are presented”.
At that time, rather extraordinarily, we’re told the justice secretary did not have the full facts of the case, even though the Metropolitan Police had been informed the day before (six days after Kaddour-Cherif was accidentally freed).
Please use Chrome browser for a more accessible video player
5:14
In full: Moment sex offender arrested
The combination of wrongly-freed prisoners and illegal migrants is a conjunction of two of the most toxic issues in British politics – the overflowing prison system and the dysfunctional asylum system.
Both are vast, chaotic problems the government is struggling to get a grip on, as the Conservatives also found, to their cost.
But ministers’ ongoing failure to bring both issues under control has only been highlighted by Mr Lammy’s sloppy handling of the situation.
Football regulator donations row
Ms Nandy has herself been at the heart of another government controversy this week – over the appointment of the new football regulator, David Kogan.
Please use Chrome browser for a more accessible video player
1:20
‘I didn’t want to mislead MPs on prisoner release’
An independent investigation found she “unknowingly” breached the code on public appointments by failing to declare that Mr Kogan had previously donated £2,900 to her Labour leadership campaign – and also criticised her department for not highlighting his status as a Labour donor who had previously given £33,410 to the party.
The culture secretary has apologised and explained she had been unaware of the donations.
She also pointed out that Mr Kogan was a candidate originally put forward by the Conservatives. But again, it’s messy.
It’s yet another story which chips away at the government’s promises to clear up politics and act with full transparency and accountability.
Please use Chrome browser for a more accessible video player
2:00
Political fallout analysed
Budget blues?
The ultimate breach of trust looks set to come with the budget on 26 November, however.
In an extraordinary early morning speech this week, Chancellor Rachel Reeves signalled that she’s likely to raise taxes in two and a half weeks – and thus breach the core promise of the Labour Party manifesto.
The rationale for her dire warnings on Tuesday was to start explaining why she will probably have to do so – getting in her excuses early about the languishing state of the economy as a result of Brexit, Donald Trump’s tariffs and her inheritance from the Conservatives.
The Tories claim Ms Reeves could sort out the finances by cutting welfare spending – something ministers dramatically failed to do when their efforts at reform were scuttled by angry backbenchers.
Governments breach their manifesto commitments all the time.
But if the chancellor goes ahead and puts up income tax, as expected (even if that’s offset, for some, by a corresponding cut to national insurance), it will be a shock – and the first such increase in 50 years.
The new deputy leader of the party, Lucy Powell, pointedly warned the government this week about the risks of breaching trust in politics by breaking manifesto promises.
Lisa Nandy didn’t shoot her comments down when Sir Trevor asked for her response, arguing instead that while “we take our promises very, very seriously”, they [Labour] “were also elected on a promise to change this country”, with a particular focus on fixing the NHS.
The impossibility of doing both – protecting taxes while also increasing government spending in such a challenging economic climate – highlights the folly of making such restrictive promises.
United States President Donald Trump announced on Sunday that most Americans will receive a $2,000 “dividend” from the tariff revenue and criticized the opposition to his sweeping tariff policies.
“A dividend of at least $2000 a person, not including high-income people, will be paid to everyone,” Trump said on Truth Social.
The US Supreme Court is currently hearing arguments about the legality of the tariffs, with the overwhelming majority of prediction market traders betting against a court approval.
Kalshi traders place the odds of the Supreme Court approving the policy at just 23%, while Polymarket traders have the odds at 21%. Trump asked:
“The president of the United States is allowed, and fully approved by Congress, to stop all trade with a foreign country, which is far more onerous than a tariff, and license a foreign country, but is not allowed to put a simple tariff on a foreign country, even for purposes of national security?”
Investors and market analysts celebrated the announcement as economic stimulus that will boost cryptocurrency and other asset prices as portions of the stimulus flow into the markets, but also warned of the long-term negative effects of the proposed dividend.
While a portion of the stimulus will flow into markets and raise asset prices, Kobeissi Letter warned that the ultimate long-term effect of any economic stimulus will be fiat currency inflation and the loss of purchasing power.
The proposed economic stimulus checks will add to the national debt and result in higher inflation over time. Source: The Kobeissi Letter
“If you don’t put the $2,000 in assets, it is going to be inflated away or just service some interest on debt and sent to banks,” Bitcoin analyst, author, and advocate Simon Dixon said.
“Stocks and Bitcoin only know to go higher in response to stimulus,” investor and market analyst Anthony Pompliano said in response to Trump’s announcement.
Opinion by: Agata Ferreira, assistant professor at the Warsaw University of Technology
A new consensus is forming across the Web3 world. For years, privacy was treated as a compliance problem, liability for developers and at best, a niche concern. Now it is becoming clear that privacy is actually what digital freedom is built on.
The Ethereum Foundation’s announcement of the Privacy Cluster — a cross-team effort focused on private reads and writes, confidential identities and zero-knowledge proofs — is a sign of a philosophical redefinition of what trust, consensus and truth mean in the digital age and a more profound realization that privacy must be built into infrastructure.
Regulators should pay attention. Privacy-preserving designs are no longer just experimental; they are now a standard approach. They are becoming the way forward for decentralized systems. The question is whether law and regulation will adopt this shift or remain stuck in an outdated logic that equates visibility with safety.
From shared observation to shared verification
For a long time, digital governance has been built on a logic of visibility. Systems were trustworthy because they could be observed by regulators, auditors or the public. This “shared observation” model is behind everything from financial reporting to blockchain explorers. Transparency was the means of ensuring integrity.
In cryptographic systems, however, a more powerful paradigm is emerging: shared verification. Instead of every actor seeing everything, zero-knowledge proofs and privacy-preserving designs enable verifying that a rule was followed without revealing the underlying data. Truth becomes something you can prove, not something you must expose.
This shift might seem technical, but it has profound consequences. It means we no longer need to pick between privacy and accountability. Both can coexist, embedded directly into the systems we rely on. Regulators, too, must adapt to this logic rather than battle against it.
Privacy as infrastructure
The industry is realizing the same thing: Privacy is not a niche. It’s infrastructure. Without it, the Web3 openness becomes its weakness, and transparency collapses into surveillance.
Emerging architectures across ecosystems demonstrate that privacy and modularity are finally converging. Ethereum’s Privacy Cluster focuses on confidential computation and selective disclosure at the smart-contract level.
Others are going deeper, integrating privacy into the network consensus itself: sender-unlinkable messaging, validator anonymity, private proof-of-stake and self-healing data persistence. These designs are rebuilding the digital stack from the ground up, aligning privacy, verifiability and decentralization as mutually reinforcing properties.
This is not an incremental improvement. It is a new way of thinking about freedom in the digital network age.
Policy is lagging behind the technology
Current regulatory approaches still reflect the logic of shared observation. Privacy-preserving technologies are scrutinized or restricted, while visibility is mistaken for safety and compliance. Developers of privacy protocols face regulatory pressure, and policymakers continue to think that encryption is an obstacle to observability.
This perspective is outdated and dangerous. In a world where everyone is being watched, and where data is harvested on an unprecedented scale, bought, sold, leaked and exploited, the absence of privacy is the actual systemic risk. It undermines trust, puts people at risk and makes democracies weaker. By contrast, privacy-preserving designs make integrity provable and enable accountability without exposure.
Lawmakers must begin to view privacy as an ally, not an adversary — a tool for enforcing fundamental rights and restoring confidence in digital environments.
Stewardship, not just scrutiny
The next phase of digital regulation must move from scrutiny to support. Legal and policy frameworks should protect privacy-preserving open source systems as critical public goods. Stewardship stance is a duty, not a policy choice.
It means providing legal clarity for developers and distinguishing between acts and architecture. Laws should punish misconduct, not the existence of technologies that enable privacy. The right to maintain private digital communication, association and economic exchange must be treated as a fundamental right, enforced by both law and infrastructure.
Such an approach would demonstrate regulatory maturity, recognizing that resilient democracies and legitimate governance rely on privacy-preserving infrastructure.
The architecture of freedom
The Ethereum Foundation’s privacy initiative and other new privacy-first network designs share the idea that freedom in the digital age is an architectural principle. It cannot depend solely on promises of good governance or oversight; it must be built into protocols that shape our lives.
These new systems, private rollups, state-separated architectures and sovereign zones represent the practical synthesis of privacy and modularity. They enable communities to build independently while remaining verifiably connected, thereby combining autonomy with accountability.
Policymakers should view this as an opportunity to support the direct embedding of fundamental rights into the technical foundation of the internet. Privacy-by-design should be embraced as legality-by-design, a way to enforce fundamental rights through code, not just through constitutions, charters and conventions.
The blockchain industry is redefining what “consensus” and “truth” mean, replacing shared observation with shared verification, visibility with verifiability, and surveillance with sovereignty. As this new dawn for privacy takes shape, regulators face a choice: Limit it under the old frameworks of control, or support it as the foundation of digital freedom and a more resilient digital order.
The tech is getting ready. The laws need to catch up.
Opinion by: Agata Ferreira, assistant professor at the Warsaw University of Technology.
This article is for general information purposes and is not intended to be and should not be taken as legal or investment advice. The views, thoughts, and opinions expressed here are the author’s alone and do not necessarily reflect or represent the views and opinions of Cointelegraph.