Skip to Content
Artificial intelligence

AI may be used to sift through Trump attorney Michael Cohen’s documents

April 26, 2018

A five-page letter released today by US federal prosecutors recommended that technology-assisted review (TAR) be called on in the case against President Donald Trump’s attorney.

What it does: TAR requires human categorization of a small number of files from the document set to train the software. It is then deployed to sort the remaining documents. Retired magistrate judge Frank Maas, one of the neutral parties recommended by the prosecution to help review the documents, says the process is considered “at least as effective as exhaustive manual review, and far more efficient.”

What it will be used for: The software will scan for records potentially covered by attorney-client privilege among the trove of documents seized earlier this month from Michael Cohen’s home and office. The prosecutors’ letter said the process was chosen to ensure that review happens in a “timely and cost effective manner.”

Why it matters: This may become a high-profile example of a kind of white-collar automation that has been under way for some time: the use of software to review, discover, and categorize legal documents. While for the most part this technology doesn’t threaten the lawyers’ jobs, it does put paralegals at risk.

Want to learn more about the future of work? Sign up for our newest newsletter, Clocking In!

Deep Dive

Artificial intelligence

Geoffrey Hinton tells us why he’s now scared of the tech he helped build

“I have suddenly switched my views on whether these things are going to be more intelligent than us.”

ChatGPT is going to change education, not destroy it

The narrative around cheating students doesn’t tell the whole story. Meet the teachers who think generative AI could actually make learning better.

Deep learning pioneer Geoffrey Hinton has quit Google

Hinton will be speaking at EmTech Digital on Wednesday.

We are hurtling toward a glitchy, spammy, scammy, AI-powered internet

Large language models are full of security vulnerabilities, yet they’re being embedded into tech products on a vast scale.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at with a list of newsletters you’d like to receive.