Facebook’s latest scientific research, about the way it shapes the political perspectives users are exposed to, has led some academics to call for the company to be more open about what it chooses to study and publish.
This week the company’s data science team published a paper in the prominent journal Science confirming what many had long suspected: that the network’s algorithms filter out some content that might challenge a person’s political leanings. However, the paper also suggested that the effect was fairly small, and less significant than a user’s own filtering behavior (see “Facebook Says You Filter News More Than Its Algorithm Does”).
Several academics have pointed to limitations of the study, such as the fact that the only people involved had indicated their political affiliation on their Facebook page. Critics point out that those users might behave in a different way from everyone else. But beyond that, a few academics have noted a potential tension between Facebook’s desire to explore the scientific value of its data and its own corporate interests.
Zeynep Tufekci, an assistant professor at the University of North Carolina, says the study is fascinating but adds that she would like to see more transparency about the way research is conducted at Facebook. “The study is interesting; I’m thrilled they’re publishing this stuff,” says Tufekci. “But who knows what else they found?”
Tufekci suggests that besides the new paper showing the “filter bubble” phenomenon to be less pronounced than some had thought, several other Facebook papers have painted the network in a positive light.
Facebook has published several important social-science studies in recent years. The enormous amount of data it collects is extremely valuable as an academic resource (see “What Facebook Knows”).
Indeed, many social scientists outside the company are keen to tap into Facebook’s data. Christian Sandvig, a professor at the University of Michigan, has used Facebook’s application programming interface (normally used to develop games or apps that run on Facebook) to study its users. “There’s a huge amount of data, and everybody’s hoping we can benefit from it,” he says.
But Sandvig also thinks greater transparency might help. “If a study is published in Science, and all three authors work for a pharmaceutical company and it says something positive about that company, we have a way to think about that,” he says.
Facebook’s approach to scientific research is evidently evolving. Last July, its data team published a study showing that both positive and negative emotions can spread between users (see “Facebook’s Emotion Study Is Just the Latest Effort to Prod Users”). That study proved controversial because the company had manipulated the information some users got to see in a way that made them feel more depressed (in fact, the changes made to users’ news feeds were minuscule, but the principle still upset many).
In response to the controversy over that study, Facebook’s chief technology officer, Mike Schroepfer, wrote a Facebook post that acknowledged people’s concerns and described new guidelines for its scientific research. “We’ve created a panel including our most senior subject-area researchers, along with people from our engineering, research, legal, privacy and policy teams, that will review projects falling within these guidelines,” he wrote. When asked about how Facebook decides to what to publish, a spokeswoman referred MIT Technology Review to those guidelines.
For some academics outside Facebook, those guidelines could perhaps be refined further still—for example, to guarantee that researchers can publish anything of scientific interest. “I want to empower the data team at Facebook,” Tufekci adds. “My problem is I don’t think they’re given the independence they should have.”
Toronto wants to kill the smart city forever
The city wants to get right what Sidewalk Labs got so wrong.
Saudi Arabia plans to spend $1 billion a year discovering treatments to slow aging
The oil kingdom fears that its population is aging at an accelerated rate and hopes to test drugs to reverse the problem. First up might be the diabetes drug metformin.
Yann LeCun has a bold new vision for the future of AI
One of the godfathers of deep learning pulls together old ideas to sketch out a fresh path for AI, but raises as many questions as he answers.
The dark secret behind those cute AI-generated animal images
Google Brain has revealed its own image-making AI, called Imagen. But don't expect to see anything that isn't wholesome.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.