A Californian lawmaker has introduced a bill to protect actors, artists, and entertainers from artificial intelligence, by mandating employment contracts to include informed consent when it comes to digital replicates.
Ash Kalra, an Assembly Member of the State of California — the lawmaker behind the bill — argued that generative AI presents a “real threat” to workers in the entertainment industry and its use should be banned unless a bargaining agreement between the parties is made.
The legislation, Assembly Bill 459, will eventually be assigned to a committee whose members will research, discuss, and make changes to the bill proposed by Kalra, which will then be put before the chamber for a vote.
Kalra explained in a Sept. 13 statement that “common sense requirements” like those in bill AB 459 need to be implemented to protect these workers:
“Mandating informed consent and representation will help ensure workers are not unknowingly at risk of losing the right to their digital self, and with it, their careers and livelihoods.”
The bill was also supported by Screen Actors Guild and American Federation of Television and Radio Artists (SAG-AFTRA) — a United States-based labor union representing over 100,000 media professionals around the world.
Kalra’s office said AB 459 will be in print later today. Here’s the draft language of the legislation. pic.twitter.com/J2rIEUBvuh
Duncan Crabtree-Ireland, the guild’s national executive director and chief negotiator stressed that protecting an actor’s digital image through consent-based laws is now a necessity.
“We believe that our members must maintain full control over the use of their digital selves through informed consent, and this is key if they are to be able to build and sustain a career,” he explained.
Crabtree-Ireland said AI copycats can lead to “abusive” and “exploitative” practices and that legislation will play a key role in curtailing such conduct:
“We see protection against the unjust transfers of these rights to be an imperative against potential abusive or exploitative practices. We are deeply concerned by the proliferation of AI-created audio and video content without full consent, and this legislation is an important step to ending these dangerous practices.”
The Screen Actors Guild has been behind a nearly four-month-long ongoing strike action in Hollywood over base compensation, better working conditions and other contentious issues.
One of the issues raised revolves around AI and its use, with the workers demanding stricter protections against artificial intelligence and larger royalties for their work, known in the industry as residuals.
In a recent interview with Variety, U.S. actor Sean Penn took a shot at many studios’ desire to utilize actors’ likenesses and voices for future AI use.
“So you want my scans and voice data and all that. OK, here’s what I think is fair: I want your daughter’s, because I want to create a virtual replica of her and invite my friends over to do whatever we want in a virtual party right now.
“Would you please look at the camera and tell me you think that’s cool?” he asked.
US President Donald Trump renewed his criticism of Federal Reserve Chair Jerome Powell, accusing him of being too slow to cut interest rates and escalating a long-running conflict that risks undermining the central bank’s political independence.
With the European Central Bank (ECB) cutting interest rates again on April 17, “Too Late” Powell has failed to act appropriately in the United States, even with inflation falling, Trump said on Truth Social on April 17.
“Powell’s termination cannot come fast enough!” Trump said.
Florida Senator Rick Scott agreed with the president, saying, “it’s time for new leadership at the Federal Reserve.”
Trump’s public criticism of the Fed breaks a decades-long convention in American politics that sought to safeguard the central bank from political scrutiny, which includes any executive decision to replace the chair.
In an April 16 address at the Economic Club of Chicago, Powell said Fed independence is “a matter of law.” Powell previously signaled his intent to serve out the remainder of his tenure, which expires in May 2026.
The Federal Reserve wields significant influence over financial markets, with its monetary policy decisions affecting US dollar liquidity and shaping investor sentiment.
Since the COVID-19 pandemic, crypto markets have increasingly come under the Fed’s sphere of influence due to the rising correlation between dollar liquidity and asset prices.
This was further corroborated by a 2024 academic paper written by Kingston University of London professors Jinsha Zhao and J Miao, which concluded that liquidity conditions now account for more than 65% of Bitcoin’s (BTC) price movements.
As inflation moderates and market turmoil intensifies amid the trade war, Fed officials are facing mounting pressure to cut interest rates. However, Powell has reiterated the central bank’s wait-and-see approach as officials evaluate the potential impact of tariffs.
A measure of real-time inflation known as “truflation” suggests that cost pressures are much weaker than the Fed’s primary indicators, which are several months out of date. Source: Truflation
The Fed is expected to maintain its wait-and-see policy approach at its next meeting in May, with Fed Fund futures prices implying a less than 10% chance of a rate cut. However, rate cut bets have increased to more than 65% for the Fed’s June policy meeting.
The Wyoming Stable Token Commission, a body authorized by the US state to issue a stablecoin, has suggested that it may clarify its language to better comply with potential guidelines from the Securities and Exchange Commission (SEC).
In an April 17 meeting in the extension of the Wyoming Capitol building, Commissioner Joel Revill suggested the body could reduce the risk of the state’s proposed WYST stablecoin qualifying as a security under SEC rules. The discussion among the commissioners and Executive Director Anthony Apollo followed the SEC issuing guidelines that certain “covered stablecoins” were considered” non-securities” and largely not subject to reporting requirements.
Wyoming Stable Token Commission Executive Director Anthony Apollo with Senator Cynthia Lummis. Source: LinkedIn
“We’re looking to kind of create our own vernacular around some of this, to clarify, and then use that as a jumping off point of discussion for the commission,” said Apollo, adding there were internal discussions regarding the SEC guidance but the commission was scheduled to address the matter in a May memo.
The commission, established after Wyoming passed a law to issue a state-issued stablecoin pegged to the US dollar and redeemable for fiat currency, has been exploring issues surrounding WYST. Wyoming Governor Mark Gordon said in August that the government initially planned a launch in the first quarter of 2025 for the stablecoin, later amending the timeline to potentially launch in July.
Looking to the US Congress for guidance
The commission said it would be monitoring efforts by the federal government to establish a regulatory framework for stablecoins. Among the proposed legislation was the Guiding and Establishing National Innovation for US Stablecoins, or GENIUS Act, in the Senate, and the Stablecoin Transparency and Accountability for a Better Ledger Economy, or STABLE Act, in the House of Representatives.
Though Wyoming is the least populated US state, with roughly 600,000 people, it has become home to some crypto firms likely seeking a regulatory-friendly jurisdiction. Custodia Bank, the digital asset bank established by Caitlin Long, is based in Cheyenne. US Senator Cynthia Lummis, who often advocates for crypto-friendly policies, represents Wyoming in the Senate.
Meta, the parent company of Facebook, Instagram, WhatsApp and Messenger, is facing antitrust proceedings that could limit its ability to develop AI amid a field of competitors.
First filed in 2021, the Federal Trade Commission (FTC) alleges that Meta’s strategy of absorbing firms — rather than competing with them — violates antitrust laws. If the court rules against Meta, it could be forced to spin out its various messenger services and social media sites into independent companies.
The loss of its stable of social media companies could harm Facebook’s competitiveness not only in the social media industry but also in its ability to train and develop its proprietary Llama AI models with data from those sites.
The trial could take anywhere from a couple of months to a year, but the outcome will have lasting consequences on Meta’s standing in the AI race.
Meta’s antitrust case and its effect on AI
The FTC first opened its complaint against Meta in 2020 when the firm was still operating as Facebook. The agency’s amended complaint a year later alleges that Meta (then Facebook) used an illegal “buy-or-bury” scheme on more creative competitors after its “failed attempts to develop innovative mobile features for its network.” This resulted in a monopoly of the “friends and family” social media market.
Meta founder and CEO Mark Zuckerberg had the chance to address these allegations on April 14, the first day of the official FTC v. Meta trial. He testified that only 20% of user content on Facebook and some 10% on Instagram was generated by users’ friends. The nature of social media has changed, Zuckerberg claimed.
“People just kept on engaging with more and more stuff that wasn’t what their friends were doing,” he said — meaning that the nature of Meta’s social media holdings was sufficiently diverse.
The FTC alleges that Meta identified potential threat competitors and bought them up. Source: FTC
At the time of the FTC’s initial complaint, Meta called the allegations “revisionist history,” a claim it repeated on April 13 when it stated the agency was “ignoring reality.” The company has argued that the purchases of Instagram and WhatsApp have benefited users and that competition has appeared in the form of YouTube and TikTok.
If the District of Columbia Circuit Court rules against Meta, the global social media giant will be forced to unwind these services into independent firms. Jasmine Enberg, vice president and principal analyst at eMarketer, told the Los Angeles Times that such a ruling could cost Meta its competitive edge in the social media market.
“Instagram really is its biggest growth driver, in the sense that it has been picking up the slack for Facebook for a long time, especially on the user front when it comes to young people,” said Enberg. “Facebook hasn’t been where the cool college kids hang out for a long time.”
The pause came after privacy advocacy group None of Your Business filed complaints in 11 European countries against Meta’s use of public data from its platforms to train its AI models. The Irish Data Protection Commission subsequently ordered a pause on the practice until it could conduct a review.
On April 14, Meta got the go-ahead to use public data — i.e., posts and comments from adult users across all of its platforms — to train the model. If these firms dissolved into separate companies, with their own organizational structures and data protection policies and practices, Meta would be cut off from an ocean of data and human communication with which its AI could be improved.
Andrew Rossow, a cyberspace attorney with Minc Law and CEO of AR Media Consulting, told Cointelegraph that in such an event, “companies would most likely control their own user data, and Meta would be restricted from using it unless new data-sharing agreements were negotiated, which would be subject to regulatory scrutiny and user/consumer privacy laws.”
However, Rossow noted that it wouldn’t be a total loss for Meta. Zuckerberg’s firm would retain the wealth of data from Facebook and Messenger. It could continue to use “opt-in” data from consumers who allow their posts to be used for AI training, and it could also employ synthetic data sets as well as third-party and open data.
Meta, the AI race and data protections
The race to unseat OpenAI and its ChatGPT model from AI dominance has grown more competitive in the last year as DeepSeek joined the fray and Meta launched the fourth iteration of its open-source Llama model.
In addition to training new models, major AI development firms are investing billions in new data centers to accommodate new iterations. In January 2025, Meta announced the construction of a 2-gigawatt data center with more than 1.3 million Nvidia AI graphics processing units.
Zuckerberg wrote in a post on Threads, “This will be a defining year for AI. In 2025, I expect Meta AI will be the leading assistant serving more than 1 billion people […] To power this, Meta is building a 2GW+ datacenter that is so large it would cover a significant part of Manhattan.”
Illustration of the data map coverage. Source: Mark Zuckerberg
His announcement followed the $500-billion Stargate project, which would see massive investment in AI development led by OpenAI and SoftBank, with Microsoft and Oracle as equity partners.
Amid this competition, AI firms are looking for broader and more varied sources of data to train their AI models — and have turned to dubious practices in order to get the data they need. In order to stay competitive with OpenAI when developing its Llama 3 model, Meta harvested thousands of pirated books from the site LibGen. According to court documents in a case pending against Meta, Llama developers harvested data from pirated books because licensing them from sources like Scribd seemed “unreasonably expensive.”
Time was another perceived motivator for using pirated works. “They take like 4+ weeks to deliver data,” one engineer wrote about services through which they could purchase book licenses.
The practice is not limited to Meta. OpenAI has also been accused of mining data from pirated work hosted on LibGen.
Rossow suggested that, “to ensure lasting impact — beyond short-term profit,” Meta would do well to “prioritize investment in advanced data collection, rigorous auditing and the implementation of privacy-preserving and encryption-based technologies.”
By focusing on transparency and responsible practices, “Meta can continue to genuinely advance AI capabilities, rebuild and nurture long-term user trust, and adapt to evolving legal and ethical standards, regardless of changes to its platform portfolio.”
What a ruling for the FTC would mean
Litigation is now hitting tech firms from all sides as they face allegations of privacy violations, copyright law infringement and stifling competition. Major cases like those facing Google, Amazon and Meta that have yet to play out will decide how and whether these firms can proceed as they have, defining the guardrails for AI development as well.
Rossow said that the current antitrust case against Meta could decide how courts interpret antitrust law for tech firms, spanning tech mergers, data usage and market competition. It would also signal that courts are “willing to break up tech conglomerates” when issues of smothering competition are involved, while at the same time, “taking current precedent a step further in harmonizing it with the laws of cyberspace.”