As AI deepfakes cause havoc during other elections, experts warn the UK’s politicians should be prepared.
“Just tell me what you had for breakfast”, says Mike Narouei, of ControlAI, recording on his laptop. I speak for around 15 seconds, about my toast, coffee and journey to their offices.
Within seconds, I hear my own voice, saying something entirely different.
In this case, words I have written: “Deepfakes can be extremely realistic and have the ability to disrupt our politics and damage our trust in the democratic process.”
Image: Tamara Cohen’s voice being turned into a deepfake
We have used free software, it hasn’t taken any advanced technical skills, and the whole thing has taken next to no time at all.
This is an audio deepfake – video ones take more effort to produce – and as well as being deployed by scammers of all kinds, there is deep concern, in a year with some two billion people going to the polls, in the US, India and dozens of other countries including the UK, about their impact on elections.
London mayor Sadiq Khan was also targeted this year, with fake audio of him making inflammatory remarks about Remembrance weekend and calling for pro-Palestine marches going viral at a tense time for communities. He claimed new laws were needed to stop them.
Advertisement
Ciaran Martin, the former director of the UK’s National Cyber Security Centre, told Sky News that expensively made video fakes can be less effective and easier to debunk than audio.
“I’m particularly worried right now about audio, because audio deepfakes are spectacularly easy to make, disturbingly easy”, he said. “And if they’re cleverly deployed, they can have an impact.”
Those which have been most damaging, in his view, are an audio deepfake of President Biden, sent to voters during the New Hampshire primaries in January this year.
A “robocall” with the president’s voice told voters to stay at home and “save” their votes for the presidential election in November. A political consultant later claimed responsibility and has been indicted and fined $6m (£4.7m).
Mr Martin, now a professor at the Blavatnik School of Government at Oxford University, said: “It was a very credible imitation of his voice and anecdotal evidence suggests some people were tricked by that.
“Not least because it wasn’t an email they could forward to someone else to have a look at, or on TV where lots of people were watching. It was a call to their home which they more or less had to judge alone.
“Targeted audio, in particular, is probably the biggest threat right now, and there’s no blanket solution, there’s no button there that you can just press and make this problem go away if you are prepared to pay for it or pass the right laws.
“What you need, and the US did this very well in 2020, is a series of responsible and well-informed eyes and ears throughout different parts of the electoral system to limit and mitigate the damage.”
He says there is a risk to hyping up the threat of deepfakes, when they have not yet caused mass electoral damage.
A Russian-made fake broadcast of Ukrainian TV, he said, featuring a Ukrainian official taking responsibility for a terrorist attack in Moscow, was simply “not believed”, despite being expensively produced.
The UK government has passed a National Security Act with new offences of foreign interference in the country’s democratic processes.
The Online Safety Act requires tech companies to take such content down, and meetings are being regularly held with social media companies during the pre-election period.
Democracy campaigners are concerned that deepfakes could be used not just by hostile foreign actors, or lone individuals who want to disrupt the process – but political parties themselves.
Polly Curtis is chief executive of the thinktank Demos, which has called on the parties to agree to a set of guidelines for the use of AI.
Image: Polly Curtis, the chief executive of Demos
She said: “The risk is that you’ll have foreign actors, you’ll have political parties, you’ll have ordinary people on the street creating content and just stirring the pot of what’s true and what’s not true.
“We want them to come together and agree together how they’re going to use these tools at the election. We want them to agree not to create generative AI or amplify it, and label it when it is used.
“This technology is so new, and there are so many elections going on, there could be a big misinformation event in an election campaign that starts to affect people’s trust in the information they’ve got.”
Deepfakes have already been targeted at major elections.
Last year, within hours before polls closed in the Slovakian presidential election, an audio fake of one of the candidates claiming to have rigged the election went viral. He was heavily defeated and his pro-Russian opponent won.
The UK government established a Joint Election Security Preparations Unit earlier this year – with Whitehall officials working with police and security agencies – to respond to threats as they emerge.
Follow Sky News on WhatsApp
Keep up with all the latest news from the UK and around the world by following Sky News
A UK government spokesperson said: “Security is paramount and we are well-prepared to ensure the integrity of the election with robust systems in place to protect against any potential interference.
“The National Security Act contains tools to tackle deepfake election threats and social media platforms should also proactively take action against state-sponsored content aimed at interfering with the election.”
A Labour spokesperson said: “Our democracy is strong, and we cannot and will not allow any attempts to undermine the integrity of our elections.
“However, the rapid pace of AI technology means that government must now always be one step ahead of malign actors intent on using deepfakes and disinformation to undermine trust in our democratic system.
“Labour will be relentless in countering these threats.”
COVID-19 fraud and error cost the taxpayer nearly £11bn, a government watchdog has found.
Pandemic support programmes such as furlough, bounce-back loans, support grants and Eat Out to Help Out led to £10.9bn in fraud and error, COVID Counter-Fraud Commissioner Tom Hayhoe’s final report has concluded.
Lack of government data to target economic support made it “easy” for fraudsters to claim under more than one scheme and secure dual funding, the report said.
Weak accountability, bad quality data and poor contracting were identified as the primary causes of the loss.
The government has said the sum is enough to fund daily free school meals for the UK’s 2.7 million eligible children for eight years.
An earlier report from Mr Hayhoe for the Treasury in June found that failed personal protective equipment (PPE) contracts during the pandemic cost the British taxpayer £1.4 billion, with £762 million spent on unused protective equipment unlikely ever to be recovered.
Factors behind the lost money had included government over-ordering of PPE, and delays in checking it.
More on Covid-19
Related Topics:
This breaking news story is being updated and more details will be published shortly.
Stablecoin issuer Circle has secured regulatory approval to operate as a financial service provider in the Abu Dhabi International Financial Center, deepening its push into the United Arab Emirates.
In an announcement Tuesday, Circle Internet Group said it received a Financial Services Permission license from the Financial Services Regulatory Authority of the Abu Dhabi Global Market (ADGM), the International Financial Centre of Abu Dhabi. This allows the stablecoin issuer to operate as a Money Services Provider in the IFC.
The USDC (USDC) issuer also appointed Saeeda Jaffar as its managing director for Circle Middle East and Africa. The new executive also serves as a senior vice president and group country manager for the Gulf Operation Council at Visa and will be tasked with developing the stablecoin issuer’s regional strategy and partnerships.
Circle co-founder, chairman and CEO Jeremy Allaire said that the relevant regulatory framework “sets a high bar for transparency, risk management, and consumer protection,” adding that those standards are needed if “trusted stablecoins” are going to support payments and finance at scale.
The newly introduced Federal Decree Law No. 6 of 2025 brings DeFi platforms, related services and infrastructure providers under the scope of regulations if they enable payments, exchange, lending, custody, or investment services, with licenses now required. Local crypto lawyer Irina Heaver said that “DeFi projects can no longer avoid regulation by claiming they are just code.”
Crypto companies seeking a US federal bank charter should be treated no differently than other financial institutions, says Jonathan Gould, the head of the Office of the Comptroller of the Currency (OCC).
Gould told a blockchain conference on Monday that some new charter applicants in the digital or fintech spaces could be seen as offering novel activities for a national trust bank, but noted “custody and safekeeping services have been happening electronically for decades.”
“There is simply no justification for considering digital assets differently,” he added. “Additionally, it is important that we do not confine banks, including current national trust banks, to the technologies or businesses of the past.”
The OCC regulates national banks and has previously seen crypto companies as a risk to the banking system. Only two crypto banks are OCC-licensed: Anchorage Digital, which has held a charter since 2021, and Erebor, which got a preliminary banking charter in October.
Crypto “should have” a way to supervision
Gould said that the banking system has the “capacity to evolve from the telegraph to the blockchain.”
He added that the OCC had received 14 applications to start a new bank so far this year, “including some from entities engaged in novel or digital asset activities,” which was nearly equal to the number of similar applications that the OCC received over the last four years.
Comptroller of the Currency Jonathan Gould giving remarks at the 2025 Blockchain Association Policy Summit. Source: YouTube
“Chartering helps ensure that the banking system continues to keep pace with the evolution of finance and supports our modern economy,” he added. “That is why entities that engage in activities involving digital assets and other novel technologies should have a pathway to become federally supervised banks.”
Gould brushes off banks’ concerns
Gould noted that banks and financial trade groups had raised concerns about crypto companies getting banking charters and the OCC’s ability to oversee them.
“Such concerns risk reversing innovations that would better serve bank customers and support local economies,” he said. “The OCC has also had years of experience supervising a crypto-native national trust bank.”
Gould said the regulator was “hearing from existing national banks, on a near daily basis, about their own initiatives for exciting and innovative products and services.”
“All of this reinforces my confidence in the OCC’s ability to effectively supervise new entrants as well as new activities of existing banks in a fair and even-handed manner,” he added.