Data watchdog warns police forces over use of live facial recognition
The ICO says there needs to be demonstrable evidence that the technology is ‘necessary, proportionate and effective considering the invasiveness’.
The UK data watchdog has warned police forces testing live facial recognition technology that there remains significant privacy and data protection issues which must be addressed.
South Wales Police and the Met Police have been trialling live facial recognition (LFR) technology as a possible way to reduce crime, but the move has been divisive.
Facial recognition technology maps faces in a crowd by measuring the distance between facial features, then compares results with a “watch list” of images, which can include suspects, missing people and persons of interest.
Data protection law applies if an organisation uses software that can recognise a face in a crowd, then scans large databases of people to check for a match. Our latest blog explains more: https://t.co/LLd7HDnpe3 pic.twitter.com/9Y7rynevwP— ICO (@ICOnews) July 9, 2019
The Information Commissioner’s Office (ICO) said it understood the legitimate aims of the controversial system but told police forces that they need to do more to demonstrate their compliance with data protection law, including in how watch lists are compiled and what images are used.
“We understand the purpose is to catch criminals,” Information Commissioner Elizabeth Denham said.
“But these trials also represent the widespread processing of biometric data of thousands of people as they go about their daily lives.
“And that is a potential threat to privacy that should concern us all.
“I believe that there needs to be demonstrable evidence that the technology is necessary, proportionate and effective considering the invasiveness of LFR.”
The watchdog – which is currently investigating the use of live facial recognition – also raised concern about technological bias, which can see more false positive matches from certain ethnic groups.
In May, a court heard from an activist who claimed that the use of facial recognition technology by the police is a breach of human rights law.
Lawyers representing Ed Bridges claim South Wales Police violated his privacy and data protection rights by processing an image taken of him in public, action which he says caused him “distress”.
South Wales Police argue that use of facial recognition technology does not infringe the privacy or data protection rights of Mr Bridges as it is used in the same way as photographing a person’s activities in public, and it does not retain the data of those not on its watch list.
But it does keep CCTV images from the scanning process for up to 31 days.
The ICO said the resulting judgment will form an important part of its investigation, and it will need to consider it before publishing its findings.
“In recent months we have widened our focus to consider the use of LFR in public spaces by private sector organisations, including where they are partnering with police forces,” Ms Denham continued.
“We’ll consider taking regulatory action where we find non-compliance with the law.”
Hannah Couchman, policy and campaigns officer at Liberty, the advocacy group representing Mr Bridges, welcomed the Information Commissioner’s comments, saying: “The Information Commissioner is right to highlight how invasive this technology is.
“Facial recognition is a discriminatory mass surveillance tool which is more likely to misidentify people of colour and women.
“It violates our privacy and forces us to self-censor – eroding or freedom to choose where we go and who we go with. It has no place on our streets.”