Facial recognition’s first blanket ban arrived in May, when San Francisco became the only city in the nation to bar police and other government agencies from using the technology.
Now the powerful software, which uses machine learning algorithms to automatically track human faces in digital footage and match them to names, is facing a broader moratorium.
State lawmakers are considering regulation barring all California police officers from running facial recognition programs on body cameras. Berkeley and Oakland are considering following San Francisco’s lead in banning all applications for local police. And federal legislators — from both sides of the aisle — are holding hearings on Capitol Hill to examine how federal agencies are using the technology, and whether it deserves more scrutiny and stricter controls.
Taken together, these efforts, pushed by activists and politicians from the tech industry’s home base in the Bay Area, constitute something not often seen in Silicon Valley: an attempt to impose pre-emptive regulations on a rapidly developing technology.
From social media to smart speakers, technological innovations have upended entire industries and changed the fabric of everyday life, with minimal public debate beforehand and sometimes significant unintended consequences. What makes facial recognition different is an emerging consensus that it poses a unique and alarming threat to basic civil liberties — and once it becomes widespread, it may be too late to stop it.
“People don’t expect to have their identity, their location, and who they associate with logged every time they step outside and walk down the street,” said Matt Cagle, an attorney for the Northern California chapter of the American Civil Liberties Union, which has been a key part of the coalition pushing for stronger regulation. “That’s the kind of world that automated face surveillance would usher in.”
The state measure, Assembly Bill 1215, would ban law enforcement agencies across California from using any “biometric surveillance system” — which includes software that would identify people by tattoo, gait and other individually distinguishable physical characteristics — in real time on police body cameras or on footage collected by those cameras. After passing the Assembly in early May, the bill is set for a key hearing Tuesday in the Senate Public Safety Committee.
Assemblyman Phil Ting, D-San Francisco, the lead author of the bill, sees it as a necessary follow-up to his 2018 legislation requiring law enforcement agencies to release body camera footage within 45 days of incidents in which police kill or seriously injure someone, or any incident in which police fire their guns.
“Body cameras were deployed to build trust with communities, to build more transparency and more openness,” Ting said. “It really was not the intention of body cameras to have roving surveillance cameras on police.”
The bill states biometric surveillance is the “functional equivalent of requiring every person to show a personal photo identification card at all times in violation of recognized constitutional rights,” regardless of consent. It runs the risk of creating massive, unregulated databases about Californians never suspected of committing a crime, and “may chill the exercise of free speech in public places” as the identities of anyone in a crowd could be immediately discerned.
Formal opposition to the bill has come from the California Police Chiefs Association, which said during an Assembly hearing that “prohibiting the use of biometric surveillance systems severely hinders law enforcement’s ability to identify and detain suspects of criminal activity.” Comparing images of suspects against facial recognition databases has led to cold cases being solved years later, and police commonly cite mass shootings or terrorist attacks as potentially useful applications of facial recognition technology applied across a city.
Ting says that he is unaware of any police departments currently using the technology in concert with body camera footage.
In a statement, the Los Angeles Police Department said it does not use facial recognition technology in the department, though it has been used in limited instances in joint investigations with other agencies.
The Los Angeles County Sheriff’s Department has conducted small pilots with body cameras but has not deployed them widely. But the department does rely on facial recognition technology as a way to generate leads in investigations, said Lt. Derek Sabatini, who manages the county biometric identification system.
Comparing suspect images against a database of Los Angeles County mug shots to surface possible persons of interest has proved valuable in solving crimes, said Sabatini, who drew a distinction between how it’s used today and its potential risks as a surveillance tool in real-time deployment.
“Surveillance needs discussion,” Sabatini said. “We should talk about it and understand how it’s used — there’s a lot of trust issues with that, and it’s totally understandable.”
Skeptics say the risks inherent in facial recognition software far outweigh potential benefits.
There’s the problem of false positives. Researchers have shown that the software often turns up incorrect matches, especially when searches are run on images of darker-skinned people and women. An ACLU study found that Amazon’s facial recognition system, Rekognition, incorrectly matched the official photos of 28 sitting members of Congress with mugshots of people who had been arrested for crimes.
The potential real-world effect of relying on unproven algorithms to identify suspects came to life in San Francisco in 2009, when a false positive from an automated license plate reader algorithm led police to believe a woman named Denise Green was driving a stolen Lexus. Green was stopped by police and forced out of her car and onto her knees at gunpoint by six officers, and the city ultimately paid hundreds of thousands of dollars to settle lawsuits linked to her detention.
But even if the software were perfectly accurate, civil libertarians say that allowing police to check the identity of any passersby without consent constitutes an invasion of privacy and undercuts current California laws on the right to anonymity in public.
Brian Hofer, one of the architects of San Francisco’s ban and the chairman of Oakland’s Privacy Advisory Commission, believes the movement for more regulation is still gaining strength.
“People believe that it’s inevitable that there’s going to be more and more surveillance, more and more police state power, and technology is going to keep creeping into our lives,” Hofer said. “But we still have the freedom and ability to say no.”
©2019 the Los Angeles Times Distributed by Tribune Content Agency, LLC.