University College London: ‘Deepfakes’ ranked as most serious AI crime threat

04 Aug 2020 | Network Updates | Update from University College London
These updates are republished press releases and communications from members of the Science|Business Network

Fake audio or video content has been ranked by experts as the most worrying use of artificial intelligence in terms of its potential applications for crime or terrorism, according to a new UCL report.

The study, published in Crime Science and funded by the Dawes Centre for Future Crime at UCL (and available as a policy briefing), identified 20 ways AI could be used to facilitate crime over the next 15 years. These were ranked in order of concern – based on the harm they could cause, the potential for criminal profit or gain, how easy they would be to carry out and how difficult they would be to stop.

Authors said fake content would be difficult to detect and stop, and that it could have a variety of aims – from discrediting a public figure to extracting funds by impersonating a couple’s son or daughter in a video call. Such content, they said, may lead to a widespread distrust of audio and visual evidence, which itself would be a societal harm.

Aside from fake content, five other AI-enabled crimes were judged to be of high concern. These were using driverless vehicles as weapons, helping to craft more tailored phishing messages (spear phishing), disrupting AI-controlled systems, harvesting online information for the purposes of large-scale blackmail, and AI-authored fake news.

Senior author Professor Lewis Griffin (UCL Computer Science) said: “As the capabilities of AI-based technologies expand, so too has their potential for criminal exploitation. To adequately prepare for possible AI threats, we need to identify what these threats might be, and how they may impact our lives.”

Researchers compiled the 20 AI-enabled crimes from academic papers, news and current affairs reports, and fiction and popular culture. They then gathered 31 people with an expertise in AI for two days of discussions to rank the severity of the potential crimes. The participants were drawn from academia, the private sector, the police, the government and state security agencies.

Crimes that were of medium concern included the sale of items and services fraudulently labelled as “AI”, such as security screening and targeted advertising. These would be easy to achieve, with potentially large profits.

Crimes of low concern included burglar bots – small robots used to gain entry into properties through access points such as letterboxes or cat flaps – which were judged to be easy to defeat, for instance through letterbox cages, and AI-assisted stalking, which, although extremely damaging to individuals, could not operate at scale.

First author Dr Matthew Caldwell (UCL Computer Science) said: “People now conduct large parts of their lives online and their online activity can make and break reputations. Such an online environment, where data is property and information power, is ideally suited for exploitation by AI-based criminal activity.

“Unlike many traditional crimes, crimes in the digital realm can be easily shared, repeated, and even sold, allowing criminal techniques to be marketed and for crime to be provided as a service. This means criminals may be able to outsource the more challenging aspects of their AI-based crime.”

Professor Shane Johnson, Director of the Dawes Centre for Future Crimes at UCL, which funded the study, said: “We live in an ever changing world which creates new opportunities – good and bad. As such, it is imperative that we anticipate future crime threats so that policy makers and other stakeholders with the competency to act can do so before new ‘crime harvests’ occur.  This report is the first in a series that will identify the future crime threats associated with new and emerging technologies and what we might do about them.”

Notes to Editors

For more information or to speak to the researchers involved, please contact Chris Lane, UCL Media Relations. T: +44 (0)20 7679 9222 / +44 (0)7717 728 648, E: [email protected]   

Caldwell M, Andrews JTA, Tanay T, & Griffin LD, ‘AI-enabled Future Crime’ will be published in Crime Science on Tuesday 4 August 2020, 1:00am UK time and is under a strict embargo until this time.

The DOI number for this paper will be 10.1186/s40163-020-00123-8

A policy briefing is available here: https://www.dropbox.com/s/x1hcokletwnlgu3/7973_AI%20crime%20policyFinalWeb_160720.pdf?dl=0

About UCL – London’s Global University

UCL is a diverse community with the freedom to challenge and think differently.

Our community of more than 41,500 students from 150 countries and over 12,500 staff pursues academic excellence, breaks boundaries and makes a positive impact on real world problems.

We are consistently ranked among the top 10 universities in the world and are one of only a handful of institutions rated as having the strongest academic reputation and the broadest research impact.

We have a progressive and integrated approach to our teaching and research – championing innovation, creativity and cross-disciplinary working. We teach our students how to think, not what to think, and see them as partners, collaborators and contributors.

For almost 200 years, we are proud to have opened higher education to students from a wide range of backgrounds and to change the way we create and share knowledge.

We were the first in England to welcome women to university education and that courageous attitude and disruptive spirit is still alive today. We are UCL.

www.ucl.ac.uk| Follow @uclnews on Twitter | Watch our YouTube channel | Listen to UCL podcasts on SoundCloud | Find out what’s on at UCL Minds | #MadeAtUCL

Find out how UCL is helping lead the global fight against COVID-19 www.ucl.ac.uk/covid-19-research

Never miss an update from Science|Business:   Newsletter sign-up