While San Francisco has long been known to be one of the most tech-friendly and tech-savvy cities in the world, but they are putting their foot down when it comes to the government using facial-recognition technology.
San Francisco supervisors will vote on surveillance oversight legislation Tuesday that includes a ban on the use of facial recognition technology by police and other city departments.
If approved by the full board, San Francisco would become the first U.S. city to outlaw the tech at a time when it is becoming a part of daily life, from international airports to home security cameras. Privacy advocates have squared off with public safety proponents at several heated hearings in San Francisco, a city teeming with tech innovation and the home of Twitter, Airbnb and Uber.
Those who support the ban say the technology is not only flawed, but a serious threat to civil rights, especially in a city that cherishes public protest and privacy. They worry people will one day not be able to go to a mall, the park or a school without being identified and tracked.
“Good policing does not mean living in a police state,” said Supervisor Aaron Peskin, who introduced the measure, at a hearing last week. “Living in a safe and secure community does not mean living in a surveillance state.”
But critics say police need all the help they can get, especially in a city with high-profile events and high rates of property crime. That people expect privacy in public space is unreasonable given the proliferation of cell phones and surveillance cameras, said Meredith Serra, a member of a resident public safety group Stop Crime SF.
“To me, the ordinance seems to be a costly additional layer of bureaucracy that really does nothing to improve the safety of our citizens,” she said at the same hearing.
The legislation would require city departments to get board approval to acquire or continue using surveillance services or technology, specifying how the surveillance would be used and for what purposes. It would ban the use of face ID technology by city departments, but it would not affect business, private or federal government use.
Critics say they worry a politicized board might lag on approving technology policy submitted by the police department.
San Francisco’s police department stopped testing the technology in 2017.
The city of Oakland is considering similar legislation.
Beginning To Fight Back On Tech
San Francisco is on track to become the first U.S. city to ban the use of facial recognition by police and other city agencies, reflecting a growing backlash against a technology that’s creeping into airports, motor vehicle departments, stores, stadiums and home security cameras.
Government agencies around the U.S. have used the technology for more than a decade to scan databases for suspects and prevent identity fraud.
But recent advances in artificial intelligence have created more sophisticated computer vision tools, making it easier for police to pinpoint a missing child or protester in a moving crowd or for retailers to analyze shoppers’ facial expressions as they peruse store shelves.
Efforts to restrict its use are getting pushback from law enforcement groups and the tech industry, though it’s far from a united front. Microsoft, while opposed to an outright ban, has urged lawmakers to set limits on the technology, warning that leaving it unchecked could enable an oppressive dystopia reminiscent of George Orwell’s novel “1984.”
“Face recognition is one of those technologies that people get how creepy it is,” said Alvaro Bedoya, who directs Georgetown University’s Center on Privacy and Technology. “It’s not like cookies on a browser. There’s something about this technology that really sets the hairs on the back of people’s heads up.”
Without regulations barring law enforcement from accessing driver’s license databases, people who have never been arrested could be part of virtual police line-ups without their knowledge, skeptics of the technology say.
They worry people will one day not be able to go to a park, store or school without being identified and tracked.
Already, a handful of big box stores across the U.S. are trying out cameras with facial recognition that can guess their customers’ age, gender or mood as they walk by, with the goal of showing them targeted, real-time ads on in-store video screens.
If San Francisco adopts a ban, other cities, states or even Congress could follow, with lawmakers from both parties looking to curtail government surveillance and others hoping to restrict how businesses analyze the faces, emotions and gaits of an unsuspecting public.
The California Legislature is considering a proposal prohibiting the use of facial ID technology on body cameras. A bipartisan bill in the U.S. Senate would exempt police applications but set limits on businesses analyzing people’s faces without their consent.
Legislation similar to San Francisco’s is pending in Oakland, California, and on Thursday another proposed ban was introduced in Somerville, Massachusetts.
Bedoya said a ban in San Francisco, the “most technologically advanced city in our country,” would send a warning to other police departments thinking of trying out the imperfect technology. But Daniel Castro, vice president of the industry-backed Information Technology and Innovation Foundation, said the ordinance is too extreme to serve as a model.
“It might find success in San Francisco, but I will be surprised if it finds success in a lot of other cities,” he said.
San Francisco is home to tech innovators such as Uber, Airbnb and Twitter, but the city’s relationship with the industry is testy. Some supervisors in City Hall are calling for a tax on stock-based compensation in response to a wave of San Francisco companies going public, including Lyft and Pinterest.
At the same time, San Francisco is big on protecting immigrants, civil liberties and privacy. In November, nearly 60% of voters approved a proposition to strengthen data privacy guidelines.
The city’s proposed face-recognition ban is part of broader legislation aimed at regulating the use of surveillance by city departments. The legislation applies only to San Francisco government and would not affect companies or people who want to use the technology. It also would not affect the use of facial recognition at San Francisco International Airport, where security is mostly overseen by federal agencies.
The Board of Supervisors is scheduled to vote on the bill Tuesday.
San Francisco police say they stopped testing face recognition in 2017. Spokesman David Stevenson said in a statement the department looks forward to “developing legislation that addresses the privacy concerns of technology while balancing the public safety concerns of our growing, international city.”
Supervisor Aaron Peskin acknowledges his legislation, called the “Stop Secret Surveillance Ordinance,” isn’t very tech-friendly. But public oversight is critical given the potential for abuse, he said.
The technology often misfires. Studies have shown error rates in facial-analysis systems built by Amazon, IBM and Microsoft were far higher for darker-skinned women than lighter-skinned men.
Even if facial recognition were perfectly accurate, its use would pose a severe threat to civil rights, especially in a city with a rich history of protest and expression, said Matt Cagle, attorney at the ACLU of Northern California.
“If facial recognition were added to body cameras or public-facing surveillance feeds, it would threaten the ability of people to go to a protest or hang out in Dolores Park without having their identity tracked by the city,” he said, referring to a popular park in San Francisco’s Mission District.
Local critics of San Francisco’s legislation, however, worry about hampering police investigations in a city with a high number of vehicle break-ins and several high-profile annual parades. They want to make sure police can keep using merchants and residents’ video surveillance in investigations without bureaucratic hassles.
Joel Engardio, vice president of grassroots group Stop Crime SF, wants the city to be flexible.
“Our point of view is, rather than a blanket ban forever, why not a moratorium so we’re not using problematic technology, but we open the door for when technology improves?” he said.
Such a moratorium is under consideration in the Massachusetts Legislature, where it has the backing of Republican and Democratic senators.
Often, a government’s facial recognition efforts happen in secret or go unnoticed. In Massachusetts, the motor vehicle registry has used the technology since 2006 to prevent driver’s license fraud, and some police agencies have used it as a tool for detectives.
“It is technology we use,” said Massachusetts State Police Lt. Tom Ryan, adding that “we tend not to get too involved in publicizing” that fact. Ryan and the agency declined to answer further questions about how it’s used.
Massachusetts Sen. Cynthia Creem, a Democrat and sponsor of the moratorium bill, said she worries about a lack of standards protecting the public from inaccurate or biased facial recognition technology. Until better guidelines exist, she said, “it shouldn’t be used” by government.
The California Highway Patrol does not use face recognition technology, spokeswoman Fran Clader said.
California Department of Motor Vehicles spokesman Marty Greenstein says facial recognition technology “is specifically not allowed on DMV photos.” State Justice Department spokeswoman Jennifer Molina said her agency does not use face ID technology, and policy states “DOJ and requesters shall not maintain DMV images for the purpose of creating a database” unless authorized.
Legislators also sought a face recognition moratorium this year in Washington, the home state of Microsoft and Amazon, but it was gutted following industry and police opposition. Microsoft instead backed a lighter-touch proposal as part of a broader data privacy bill, but deliberations stalled before lawmakers adjourned late last month.