Police use of facial recognition ruled lawful by High Court

Written by Sam Trendall on 4 September 2019 in News

Challenge brought against South Wales force is rejected by judges

Credit: PA

High Court judges have ruled that the use of automated facial-recognition (AFR) technology by the police is lawful.

The ruling concluded a legal challenge against South Wales Police brought by Cardiff man Ed Bridges – whose case was backed by human rights campaign group Liberty. 

Bridges had challenged the legality of the use of the technology on the grounds that it contravened both human rights and data protection legislation. He also contended that the deployment of AFR was not “in accordance with the public sector equality duty contained in the Equality Act 2010”.

His application for a judicial review has been “refused… on all grounds”.

Judges found that, while South Wales Police’s use of facial recognition technology did “engage” the privacy rights of citizens whose images were captured, it did so with “sufficient legal controls… and [was] legally justified”.

The court acknowledged that the use of AFR did involve the “collecting and processing” of citizens’ personal data. But, once again, judges concluded that this data-processing “was lawful and met the conditions set out” in last year’s Data Protection Act.

Related content

Judges were “also satisfied that before commencing the trial of AFR… [South Wales Police] had complied with the requirements of the public sector equality duty”.

Bridges has said he will appeal the decision.

“South Wales Police has been using facial recognition indiscriminately against thousands of innocent people, without our knowledge or consent,” he added. “This sinister technology undermines our privacy and I will continue to fight against its unlawful use to ensure our rights are protected and we are free from disproportionate government surveillance.”

Facial recognition is not yet widely used in law enforcement, but South Wales Police is one a handful of forces to trial the technology “with a view to it being rolled out nationally”, judges said. 

The application brought by Bridges concerned the police’s AFR Locate tool, which processes camera images in real time and compares facial biometric data with that contained on a watchlist of wanted persons. South Wales Police has thus far used AFR Locate on 50 separate occasions.

The Information Commissioner’s Office has, in recent months, voiced its concerns about the growing use of facial recognition – describing it as a “priority area” of work for the regulator.

In response to today’s ruling, an ICO spokesperson said: “We will be reviewing the judgment carefully. We welcome the court’s finding that the police use of live facial recognition (LFR) systems involves the processing of sensitive personal data of members of the public, requiring compliance with the Data Protection Act 2018. This new and intrusive technology has the potential, if used without the right privacy safeguards, to undermine rather than enhance confidence in the police.”

The spokesperson added: “Our investigation into the first police pilots of this technology has recently finished. We will now consider the court’s findings in finalising our recommendations and guidance to police forces about how to plan, authorise and deploy any future LFR systems. In the meantime, any police forces or private organisations using these systems should be aware that existing data protection law and guidance still apply.”


About the author

Sam Trendall is editor PublicTechnology

Share this page




Please login to post a comment or register for a free account.

Related Articles

Government vetting processes hampered by ‘old and unstable’ technology
20 January 2023

NAO report finds ageing IT is a major contributor to the performance issues at UKSV

Cabinet Office migrates data as Covid fraud hotline is wound down
18 January 2023

Dedicated reporting tools for coronavirus-related scams are being shuttered and case information transferred to law-enforcement entity

Government must earn public trust that AI is being used safely and responsibly
5 January 2023

Leaders from two of government’s core digital and data units – the CDDO and CDEI – introduce new guidelines intended to promote transparency in the public sector’s use of algorithms

Extremism increasingly spread via mainstream apps and sites, government research finds
16 December 2022

MoJ-backed study concludes that specialist sites and the dark web are no longer the only means of online radicalisation