Skip to Content
Silicon Valley

Twitter picks researchers to help it clean up users’ conversations

July 30, 2018

The site wants these academics to create ways to measure the “health” of conversations on the platforms. It hopes these metrics will lead to kinder tweets (or at least less racism and trolling). 

The details: Twitter said Monday that it selected two groups of academic researchers to lead the conversation health project—the conclusion of a process started in March. The company apparently received more than 230 submissions from institutions eager to work on metrics that engineers can use to help analyze and improve the social network.

The plans: The first group includes researchers from universities in the Netherlands, the US, and in Italy. They will look at the issues of “echo chambers” and impolite conversations, and produce ways to measure how much Twitter users interact with diverse viewpoints on the social network. They’ll also develop algorithms to tell the difference between “incivility” (which could simply be rude chatter) and “intolerant discourse” (such as racist language).

The second group, made up of researchers from universities in the UK and the Netherlands, will look at how showing users others who have different perspectives and backgrounds can reduce prejudice and discrimination.

Isn’t Twitter a US-based company? Well, sure. But while the US has more Twitter users than any other single country, only about a fifth of the total number of Twitter users are based in the US. So it makes a lot of sense that the company would involve researchers who are based outside the country if it wants to get a good sense of what’s going on globally with its user base.

The prognosis: Long-term effects remain to be seen, but when MIT Technology Review  talked to academic researchers inside and outside the US in March, they thought it would be tricky to measure conversational health with just a few metrics.

Twitter seems to be taking the whole process seriously, though, and recently purged tens of millions of fake accounts in another effort to show it’s trying to improve. 

Keep Reading

Most Popular

Large language models can do jaw-dropping things. But nobody knows exactly why.

And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.

The problem with plug-in hybrids? Their drivers.

Plug-in hybrids are often sold as a transition to EVs, but new data from Europe shows we’re still underestimating the emissions they produce.

Google DeepMind’s new generative model makes Super Mario–like games from scratch

Genie learns how to control games by watching hours and hours of video. It could help train next-gen robots too.

How scientists traced a mysterious covid case back to six toilets

When wastewater surveillance turns into a hunt for a single infected individual, the ethics get tricky.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.