Clearview AI keeps lying, while selling incredibly powerful facial recognition that traffics in regular peoples’ sensitive data.
How would you like it if anyone armed with an iPhone could figure out a slew of information about you, pull up any Facebook or Instagram picture you’ve ever been in, and see any other image of you that’s been posted publicly online?
This scenarios appears to be possible thanks to a shady startup called Clearview AI which, as a New York Times investigation earlier this year revealed, has been mining your pictures online to build a vast facial recognition database. At first, Clearview AI maintained that its tool was only meant to be used by law enforcement and a few private companies. But it later became clear that the company has consistently misrepresented both the extent of its work and the breadth of its aspirations.
This is concerning. Facial recognition is an incredibly powerful tool, and Clearview’s tech is trafficking in highly personal information — including, potentially, yours. (If you’re a resident of California or the European Union, feel free to request the data Clearview has on you here, but note you’ll likely need to send them a copy of government-issued photo ID.)
Reporting from Buzzfeed News shows that the company had a client list that went far beyond law enforcement, and included Walmart, Best Buy, and Macy’s. Now some of these companies are distancing themselves from the facial recognition startup.
Clearview exaggerated the role its technology has played in assisting law enforcement in their investigations, while also telling Recode that it had only arranged contracts for its technology with customers in the United States and Canada. Documents identified by Buzzfeed News indicate that the company also provided the technology to users in Saudi Arabia and the United Arab Emirates, including to agencies “related” to the government.
And, just this week, we learned from a New York Times report that Clearview AI’s leaders had even offered up its tech for the rich to try, and that one billionaire used the software to identify a man who his daughter was dating.
The controversy surrounding Clearview has since prompted a probe from US senators. The very existence of facial recognition technology has always been a source of debate. Law enforcement has been using facial recognition for several years now, and companies, schools, and other organizations are increasingly making use of the AI-powered software. But Clearview’s technology represents a frightening step towards an all-powerful system that the world hasn’t seen before. The secretive company says it’s created a database of over 3 billion images that have been scraped from all corners of the internet, including social networks like Facebook, Instagram, and YouTube. From just a snapshot or video still, Clearview claims its app lets someone using the tech to identify a face and match it with publicly available information about the person, within just a few seconds.
Meanwhile, OneZero revealed that the Clearview AI had also sought to supplement its database of web-scrapped photos with mug shots, though it’s not clear whether the company was successful.
Do we want to live in a world where this technology exists? Clearview argues that the tech can help track down dangerous people, and its site points to “child molesters, murderers, suspected terrorists.” And as the Times reported in February, the company’s facial recognition has helped identify child victims in exploitative videos posted to the web. But clear the tech can be used for a lot more than that.
And critics say the facial recognition is way too risky, enabling excessive surveillance and threatening our privacy rights. Another concern is that the technology, broadly, has also been shown to be less accurate on people of color, women, and other minority groups.
Faced with these concerns, the world’s biggest tech companies have stepped up and, sent cease-and-desist letters to Clearview that order the company to stop scraping their sites for our data. But it’s not clear how much good that will do, or how invested they actually are in protecting our personal information. While some lawsuits against Clearview are also popping up, it’s not yet apparent how Clearview could be stopped. That has privacy advocates pointing to the need for a federal law regulating, or even outright banning, facial recognition in the United States.
Facial recognition isn’t new. But this huge database of faces is.
So here’s how Clearview’s tool works. Say you have an image of a person, but you don’t know their name. You could input that photo into Clearview’s app, and it will turn up any image of the person that it had scraped from the internet, as well as links to websites from which those images came. That could be a good amount of information.
Again, Clearview’s database reportedly includes more than 3 billion images taken from around the web. That’s much more than what law enforcement agencies typically have access to. The Times reports that the technology will work with images of faces from many different angles, while older facial recognition tools used by police departments might require the subject to be looking straight ahead, like in a mug shot.
That means images from social media posts on platforms like Instagram could pop up — even images that are no longer, but once were, publicly available. And keep in mind: The tool doesn’t just surface pictures that you’ve taken and posted online. It will also turn up any photos posted of you, even those posted without your consent or knowledge.
“Clearview is a search engine for publicly available images,” Clearview CEO Hoan Ton-That told Recode in an email. “We do not and cannot index any images that are private or protected, such as those in a private Instagram account. A previously public image may or may not be searchable with Clearview depending on the time it was public and that would depend on the individual case.”
Clearview decides who can — and can’t — use this tool
More than 600 law enforcement agencies have used Clearview AI in the past year, as have federal agencies like the Federal Bureau of Investigation (FBI) and the Department of Homeland Security (DHS), according to the Times. The FBI would not confirm to Recode that it had used Clearview’s tool, instead pointing to its testimony last year about its use of facial recognition more broadly. The DHS has not responded to Recode’s request for comment.
Recent reporting suggests that the tool was offered to a wide variety of companies for security, though Ton-That wouldn’t tell Recode which ones.
It’s also unclear if Clearview has any internal limitations regarding who can get access to its technology, though we now know not to expect to company to be upfront with the public. For instance, Ton-That told the Times earlier this year, “If it’s a country where it’s just governed terribly, or whatever, I don’t know if we’d feel comfortable, you know, selling to certain countries.”
But as other reporting indicates, the company does feel comfortable selling this technology to clients in countries with concerning records on civil liberties and human rights, and BuzzFeed recently reported that the technology had offered to users in Saudi Arabia and the UAE.
Meanwhile, Clearview had claimed to Recode previously that it does not have contracts outside the US and Canada (The Royal Canadian Mounted Police, the national police service in Canada, told Recode that it does not comment on specific investigative tools but that it researches emerging technologies).
The use of this tool by law enforcement alone raises questions, however. Imagine, for instance, the ethical problems at play with a police officer using Clearview to identify a protester. Or, say the facial recognition doesn’t work as it should and a false “match” ultimately leads to arresting someone for a crime they didn’t commit.
Initially, there was fear that Clearview’s technology could one day be made available to just about anyone. Such a development could destroy our expectation of being anonymous in public.
It’s not difficult to imagine terrifying uses of this, and we already have a hint, given how associates of Clearview AI’s app have already made use of the tool. But imagine if a nude picture of you was, at some point in time, posted online. With the snap of a phone camera, it’s possible that anyone with Clearview could instantaneously find that image. Or imagine you’re walking down the street, and someone decides they want to know where you live and whether you had kids. All they might need is the Clearview app.
The scope of Clearview’s threat to privacy remains unclear. The Times has reported: “Police officers and Clearview’s investors predict that its app will eventually be available to the public.” Unsurprisingly, that was different from what Ton-That told Recode.
“Our strict commitment at this time and at all times previously is and has been not to make this tool available to the general public,” Ton-That said in an email to Recode. “Our mission is to reduce crime, fraud, and abuse using our powerful new technology. Any abuse of our technology would be in total violation of our mission and values.”
There are many good reasons not to trust this company
Clearview’s past has raised alarm bells. For one thing, CEO Hoan Ton-That’s previous ventures included an app that added Trump’s hair onto photos of people, and he’s also been linked to a phishing site. At one point, the company tried to sell a database for “extreme opposition research” to Paul Nehlen, a white supremacist and anti-Semite, who was running for Congress at the time, according to the Times. One investor in the company is Peter Thiel, who helped found PayPay and Palantir and has vocally supported President Trump.
The company hasn’t been exactly forthright in its advertising claims, either. For instance, BuzzFeed News reporting found two cases in which Clearview claimed a certain law enforcement agency was “using” its product, when in fact the agency had simply received a tip from the company. Clearview also appears to be claiming that hundreds of police departments were working with them when some of those police departments only signed up for a trial, according to BuzzFeed News.
Ton-That told Recode, “Our calculations are based on the number of agencies that are actively using Clearview’s technology to help solve crimes.”
It’s not quite clear how well the tool actually works. According to marketing materials obtained by BuzzFeed News, Clearview claimed to have “accuracy finding a match out of 1 million faces” 98.6 percent of the time. At the same time, Clearview told the New York Times that the tool produces a match up to 75 percent of the time, although we don’t know how many of those are “true” matches. Ton-That also told the Times that one difficulty for Clearview’s algorithms is that photos scraped from the web were generally at eye-level, which is not at the same angle typically captured by surveillance cameras.
The company had also said that it had checked the accuracy of its tool using a methodology from the American Civil Liberties Union (ACLU), an assertion the organization has pushed strongly back against. The claim has since been removed from Clearview’s site, though the company still says that its tool was reviewed and certified by an “independent panel of experts.” (Recode asked Clearview who those experts were, but we never heard back.)
Meanwhile, Craig Watson, the director of the National Institute of Standard and Technology’s Image group, told Recode that Clearview had not volunteered to participate in its facial recognition vendor testing program, and there were no plans to evaluate its algorithms.
Tech companies are fighting back, but that might not be enough
Major internet platforms have responded to Clearview by sending the company cease-and-desist letters. These companies now include LinkedIn, Twitter, Facebook, Venmo, Google, and YouTube. It’s worth noting that Twitter explicitly bans using its platform for the purpose of facial recognition. Meanwhile, Pinterest told Recode that it wasn’t aware of scraping on its site by Clearview, though such web-scraping would violate their community guidelines. When asked about Clearview scraping its images and videos, PornHub VP Blake White said in a statement, “We have no knowledge of this and actively block scrapers as they violate our Terms of Service.”
Apple has since suspended Clearview’s access to its developer app account for violating its terms.
Still, none of the tech companies that responded to Recode’s request for comment have said whether they planned to escalate their demands with lawsuits. Clearview said it has received these letters from companies, and that its “attorneys are responding appropriately.” In an interview with CBS, the company’s CEO insisted that there is a First Amendment right to public information.
As users of these technology platforms, members of the public are now in a curious position. Although we’ve long criticized these platforms for profiting off our data, we’re now potentially reliant on these companies to defend us from a dystopian world of facial recognition. Keep in mind that several of these companies, like Google and Facebook, have worked on facial recognition technology of their own, albeit confronted with varying levels of controversy and ethical concerns. In February, we learned that Facebook had agreed to pay $550 million to settle an Illinois facial recognition lawsuit inspired by its “photo-tagging” suggestion feature. But despite having access to billions of our photos, none of these companies have gone forward with creating such a tool — a boundary Clearview has now crossed.
So Clearview might have your photos. Can you do anything about it?
In a later email, Ton-That said: “We’re processing removal requests for persons in jurisdictions that impose that legal requirement.” In order to make those requests, you would need to confirm your identity with Clearview by sending in — get this — a photo of yourself on a government ID.
But if you (understandably) don’t feel comfortable doing that, you might consider suing Clearview. Several lawsuits have now been filed against Clearview, including in Virginia, Illinois and California. While Texas also has a biometric privacy law, it requires the state’s attorney general to take action. That doesn’t appear to have happened.
Ton-That said that Clearview designed its tool to follow “all relevant laws and regulation.”
Scott Drury, a former Illinois state representative and attorney representing one of those class action lawsuits, told Recode in February they’re not just arguing that Clearview violated the Illinois Biometric Privacy Act. He said they’re also suing on constitutional grounds.
“In this case, you have citizens who clearly have a right to their privacy and they have a right to know how the photographs that they put online are being used,” Drury said. “And Clearview, working with law enforcement, specifically and covertly took these photos and scraped them from the internet and didn’t let anyone know that they were doing that.”
The controversy has prompted some police departments to speak out. Now, the NYPD is denying that it has an “institutional relationship” with the company, according to BuzzFeed News. The Philadelphia Police Department issued a statement saying that, for now, they’re only testing Clearview AI’s tool.
The Chicago Police Department has said that, after a trial, it spent nearly $50,000 to use the tech for two years, according to the Chicago Sun-Times. Meanwhile, the state attorney general of New Jersey has ordered that all police in the state stop using the tool. Others, including the Miami Police Department, have been commenting on cases where Clearview AI has helped to identify suspects.
But a few cities have already banned law enforcement from using facial recognition. The revelations about Clearview AI have also bolstered calls for federal regulation of the technology.
“Congress needs to stop messing around and pass legislation to ban the use of face surveillance technology nationwide,” said Evan Greer, the deputy director for the digital rights group Fight for the Future, in a statement. That mimics what other privacy advocates are saying.
And some lawmakers appear to be listening. When the New York Times story came out, Sen. Cory Booker also tweeted that if “Congress doesn’t act to limit the use of technology in this manner we risk losing our privacy and our freedom.” And at the end of January, Sen. Ed Markey wrote to Ton-That demanding a list of all law enforcement and intelligence agencies in communication with Clearview, along with other questions.
Sen. Booker has since sent another letter directly to the company, calling Clearview’s initial response “unacceptable,” and demanding more information about the company’s potential to collect the images of children and interest in selling its technology to countries with problematic records on human rights.
Still, facial recognition has long been used by law enforcement agencies. So while it’s possible that Clearview AI could finally galvanize enough backlash to get lawmakers to act, only time will tell. In the meantime, now would be a great time to switch your social media accounts to private. You friends might care what you were wearing Saturday night, but maybe that’s something the cops don’t need to know.
Open Sourced is made possible by the Omidyar Network. All Open Sourced content is editorially independent and produced by our journalists.
Author: Rebecca Heilweil