Skip to Content

The walls are closing in on Clearview AI

The controversial face recognition company was just fined $10 million for scraping UK faces from the web. That might not be the end of it.

Hoan Ton-That, CEO of Clearview AI
Hoan Ton-That, CEO of Clearview AI, demonstrates the facial recognition software, using his own face in FebruaryAP Photo/Seth Weni

Controversial facial recognition company Clearview AI has been fined almost $10 million by the UK’s data protection watchdog for collecting the faces of UK citizens from the web and social media. The firm was also ordered to delete all of the data it holds on UK citizens.

The move by the UK’s Information Commissioner’s Office (ICO) is the latest in a string of high-profile fines against the company as data protection authorities around the world eye tougher restrictions on its practices.

The ICO found that Clearview AI had been in breach of data protection laws, collected personal data without people’s consent, and asked for additional information, such as photos, when people asked if they were in the database. It found that this may have “acted as a disincentive” for people who objected to their data being scraped. 

“The company not only enables identification of those people, but effectively monitors their behaviour and offers it as a commercial service. That is unacceptable,” said John Edwards, the UK’s information commissioner, in a statement. 

Clearview AI boasts one of the world’s largest databases of people’s faces, with 20 billion images that it has scraped off the internet from publicly available sources, such as social media, without their consent. Clients such as police departments pay for access to the database to look for matches.

Data protection authorities around the Western world have found this to be a clear violation of privacy and are now beginning to work together to clamp down. Edwards stressed that “international cooperation is essential to protect people’s privacy rights in 2022” and is due to meet with European regulators in Brussels this week. The UK’s investigation into Clearview AI was carried out jointly with the Australian information commissioner.

Earlier this year, Italian data protection authorities fined Clearview AI €20 million ($21 million) for breaching data protection rules. Authorities in Australia, Canada, France, and Germany have reached similar conclusions. 

Even in the US, which does not have a federal data protection law, Clearview AI is facing increasing scrutiny. Earlier this month the ACLU won a major settlement that restricts Clearview from selling its database across the US to most businesses. In the state of Illinois, which has a law on biometric data, Clearview AI cannot sell access to its database to anyone, even the police, for five years. 

According to Silkie Carlo, who is the director of UK-based digital rights group Big Brother Watch, the ICO’s decision “effectively stops Clearview from operating in the UK,” she said on Twitter.  

Carlo added that the decision “should be a nail in the coffin for facial recognition” and called for UK lawmakers to ban facial recognition surveillance. 

Europe is working on an AI law that could ban the use of “real-time” remote biometric identification systems, such as facial recognition, in public places. The current drafting of the text restricts the use of facial recognition by law enforcement unless it is to fight serious crimes, such as terrorism or kidnappings. 

There is a possibility that the EU will go further. The EU’s influential data protection watchdogs have called for the bill to ban not only remote biometric identification in public, but the police use of web-scraped databases, such as Clearview AI’s. 

“Clearview AI is fast becoming so toxic that no credible law enforcement agency or public authority or other company will want to work with them,” says Ella Jakubowska, who works on facial recognition and biometrics for European Digital Rights, a digital rights group. 

Hoan Ton-That, Clearview AI's CEO, said he is disappointed the ICO has “misinterpreted my technology and intentions.” 

“We collect only public data from the open internet and comply with all standards of privacy and law,” he said in a statement that was sent to MIT Technology Review. 

“I would welcome the opportunity to engage in conversation with leaders and lawmakers so the true value of this technology which has proven so essential to law enforcement can continue to make communities safe,” he added.

Deep Dive


What’s next for AI regulation in 2024? 

The coming year is going to see the first sweeping AI laws enter into force, with global efforts to hold tech companies accountable. 

Three technology trends shaping 2024’s elections

The biggest story of this year will be elections in the US and all around the globe

Four lessons from 2023 that tell us where AI regulation is going

What we should expect in the coming 12 months in AI policy

The FTC’s unprecedented move against data brokers, explained

It could signal more aggressive action from policy makers to curb the corrosive effects that data brokers have on personal privacy.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at with a list of newsletters you’d like to receive.