Neural Network Rates Images for Happiness Levels
Sentiment analysis is revolutionizing the study of communication with a numerous companies now offering it as a service. The idea is to study the patterns of words in messages such as tweets and blogs to determine to what extent they are positive or negative. That allows companies, organizations, and political parties to automatically track opinions about their brands.
But while this technology has been evolving, little research has focused on the sentiment in pictures. Today, that changes thanks to the work of Can Xu at the University of California, San Diego, and a group of researchers from Yahoo Labs in Sunnyvale. These folks have developed a way to automatically assess the sentiment associated with a picture and say that it outperforms other state-of-the-art techniques.
Xu and co do not start from scratch. While sentiment in pictures has been largely ignored, the problem of object recognition in images is a well-developed field that has improved in leaps and bounds in recent years.
So Xu and co begin with a neural network already trained on a data set of images showing objects divided into 1000 classifications. When shown an image, this network gives a distribution showing how likely it is that the image falls into each of these 1,000 classifications.
It is this 1,000-dimension output that Xu and co use in their research. They first take two datasets of images from Tumblr and Twitter that have already been assessed for sentiment on a five-point scale of very negative, negative, neutral, positive, and very positive.
They then train a machine learning algorithm to find a correlation between the 1,000-dimension output and the sentiment. Having trained the machine, they then compare it two other state-of-the-art sentiment analysis techniques, such as one that relies on low level visual features like image color and another called SentiBank, which generates an adjective-noun description of a picture and hence gives a sense of sentiment.
Xu and co say their technique dramatically outperforms the existing approaches. “Experiments demonstrate that our proposed models outperform the state-of-the-art methods on both Twitter and Tumblr datasets,” they say.
That’s a useful start in the incipient field of image sentient analysis. “The results for the first time suggest that Convolutional Neural Networks are highly promising for visual sentiment analysis,” they say.
Nevertheless, there is significant work ahead. One notorious problem with word-based sentiment analysis is that it does not cope with subtle cultural influences, such as sarcasm and irony. And this kind of uniquely human behavior can severely reduce the reliability of sentiment analysis.
Just how important these kinds of idiosyncrasies will be for pictures has yet to be determined but image sentiment is could yet be another area in which human performance will soon be monitored and perhaps even matched by machines.
Ref: arxiv.org/abs/1411.5731 : Visual Sentiment Prediction with Deep Convolutional Neural Networks
Geoffrey Hinton tells us why he’s now scared of the tech he helped build
“I have suddenly switched my views on whether these things are going to be more intelligent than us.”
ChatGPT is going to change education, not destroy it
The narrative around cheating students doesn’t tell the whole story. Meet the teachers who think generative AI could actually make learning better.
Meet the people who use Notion to plan their whole lives
The workplace tool’s appeal extends far beyond organizing work projects. Many users find it’s just as useful for managing their free time.
Learning to code isn’t enough
Historically, learn-to-code efforts have provided opportunities for the few, but new efforts are aiming to be inclusive.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.