Londons Met Police switches on live facial recognition, flying in face of human rights concerns

While EU lawmakers are mulling a temporary ban on the use of facial recognition to safeguard individuals rights, as part of risk-focused plan to regulate AI, Londons Met Police has today forged ahead with deploying the privacy hostile technology flipping the switch on operational use of live facial recognition in the UK capital.
The deployment comes after a multi-year period of trials by the Met and police in South Wales.
The Met says its use of the controversial technology will be targeted to specific locations where intelligence suggests we are most likely to locate serious offenders.
Each deployment will have a bespoke watch list, made up of images of wanted individuals, predominantly those wanted for serious and violent offences, it adds.
It also claims cameras will be clearly signposted, adding that officers will be deployed to the operation will hand out leaflets about the activity.
At a deployment, cameras will be focused on a small, targeted area to scan passers-by, it writes. The technology, which is a standalone system, is not linked to any other imaging system, such as CCTV, body worn video or ANPR.
The biometric system is being provided to the Met by Japanese IT and electronics giant, NEC.
In a press statement, assistant commissioner Nick Ephgrave claimed the force is taking a balanced approach to using the controversial tech.
We all want to live and work in a city which is safe: the public rightly expect us to use widely available technology to stop criminals. Equally I have to be sure that we have the right safeguards and transparency in place to ensure that we protect peoples privacy and human rights. I believe our careful and considered deployment of live facial recognition strikes that balance, he said.
London has seen a rise in violent crime in recent years, with murder rates hitting a ten-year peak last year.
The surge in violent crime has been linked to cuts to policing services although the new Conservative government has pledged to reverse cuts enacted by earlier Tory administrations.
The Met says its hope for the AI-powered tech is will help it tackle serious crime, including serious violence, gun and knife crime, child sexual exploitation and help protect the vulnerable.
However its phrasing is not a little ironic, given that facial recognition systems can be prone to racial bias, for example, owing to factors such as bias in data-sets used to train AI algorithms.
So in fact theres a risk that police-use of facial recognition could further harm vulnerable groups who already face a disproportionate risk of inequality and discrimination.
Yet the Mets PR doesnt mention the risk of the AI tech automating bias.
Instead it makes pains to couch the technology as additional tool to assist its officers.
This is not a case of technology taking over from traditional policing; this is a system which simply gives police officers a prompt, suggesting that person over there may be the person youre looking for, it is always the decision of an officer whether or not to engage with someone, it adds.
While the use of a new tech tool may start with small deployments, as is being touting here, the history of software development underlines how potential to scale is readily baked in.
A targeted small-scale launch also prepares the ground for Londons police force to push for wider public acceptance of a highly controversial and rights-hostile technology via a gradual building out process. Aka surveillance creep.
On the flip side, the text of the draft of an EU proposal for regulating AI which leaked last week floating the idea of a temporary ban on facial recognition in public places noted that a ban would safeguard the rights of individuals. Although its not yet clear whether the Commission will favor such a blanket measure, even temporarily.
UK rights groups have reacted with alarm to the Mets decision to ignore concerns about facial recognition.
Liberty accused the force of ignoring the conclusion of a report it commissioned during an earlier trial of the tech which it says concluded the Met had failed to consider human rights impacts.
It also suggested such use would not meet key legal requirements.
Human rights law requires that any interference with individuals rights be in accordance with the law, pursue a legitimate aim, and be necessary in a democratic society, the report notes, suggesting the Met earlier trials of facial recognition tech would be held unlawful if challenged before the courts.

When the Met trialled #FacialRecognition tech, it commissioned an independent review of its use.
Its conclusions:
?The Met failed to consider the human rights impact of the tech
?Its use was unlikely to pass the key legal test of being "necessary in a democratic society"
Liberty (@libertyhq) January 24, 2020
See also:
Leave a comment
  • Latest
  • Read
  • Commented
Calendar Content
«     2020    »