Could a computer, at a glance, tell the difference between a joyful image and a depressing one?
Could it distinguish, in a few milliseconds, a romantic comedy from a horror film?
Yes, and so can your brain, according to research published this week by CU Boulder neuroscientists.
“Machine learning technology is getting really good at recognizing the content of images—of deciphering what kind of object it is,” said senior author Tor Wager, who worked on the study while a professor of psychology and neuroscience at CU Boulder. “We wanted to ask: Could it do the same with emotions? The answer is yes.”
Part machine-learning innovation, part human brain-imaging study, the paper, published Wednesday in the journal Science Advances, marks an important step forward in the application of “neural networks”—computer systems modeled after the human brain—to the study of emotion.
It also sheds a new, different light on how and where images are represented in the human brain, suggesting that what we see—even briefly—could have a greater, more swift impact on our emotions than we might assume.
“A lot of people assume that humans evaluate their environment in a certain way and emotions follow from specific, ancestrally older brain systems like the limbic system,” said lead author Philip Kragel, a postdoctoral research associate at the Institute of Cognitive Science. “We found that the visual cortex itself also plays an important role in the processing and perception of emotion.”
The birth of EmoNet
For the study, Kragel started with an existing neural network, called AlexNet, which enables computers to recognize objects. Using prior research that identified stereotypical emotional responses to images, he retooled the network to predict how a person would feel when they see a certain image.
He then “showed” the new network, dubbed EmoNet, 25,000 images ranging from erotic photos to nature scenes and asked it to categorize them into 20 categories such as craving, sexual desire, horror, awe and surprise.
EmoNet could accurately and consistently categorize 11 of the emotion types. But it was better at recognizing some than others. For instance, it identified photos that evoke craving or sexual desire with more than 95 percent accuracy. But it had a harder time with more nuanced emotions like confusion, awe and surprise.
Even a simple color elicited a prediction of an emotion: When EmoNet saw a black screen, it registered anxiety. Red conjured craving. Puppies evoked amusement. If there were two of them, it picked romance. EmoNet was also able to reliably rate the intensity of images, identifying not only the emotion it might illicit but how strong it might be.
When the researchers showed EmoNet brief movie clips and asked it to categorize them as romantic comedies, action films or horror movies, it got it right three-quarters of the time.
What you see is how you feel
To further test and refine EmoNet, the researchers then brought in 18 human subjects.
As a functional magnetic resonance imaging (fMRI) machine measured their brain activity, they were shown 4-second flashes of 112 images. EmoNet saw the same pictures, essentially serving as the 19th subject.
When activity in the neural network was compared to that in the subjects’ brains, the patterns matched up.
“We found a correspondence between patterns of brain activity in the occipital lobe and units in EmoNet that code for specific emotions. This means that EmoNet learned to represent emotions in a way that is biologically plausible, even though we did not explicitly train it to do so,” said Kragel.
The brain imaging itself also yielded some surprising findings. Even a brief, basic image – an object or a face – could ignite emotion-related activity in the visual cortex of the brain. And different kinds of emotions lit up different regions.
“This shows that emotions are not just add-ons that happen later in different areas of the brain,” said Wager, now a professor at Dartmouth College. “Our brains are recognizing them, categorizing them and responding to them very early on.”
Ultimately, the resesarchers say, neural networks like EmoNet could be used in technologies to help people digitally screen out negative images or find positive ones. It could also be applied to improve computer-human interactions and help advance emotion research.
The takeaway for now, says Kragel:
“What you see and what your surroundings are can make a big difference in your emotional life.”
Learn more: A computer system that knows how you feel
The Latest on: Machine learning technology
via Google News
The Latest on: Machine learning technology
- Parasoft wins 2020 VDC Research Embeddy Award for Its Artificial Intelligence (AI) and Machine Learning (ML) Innovationon April 7, 2020 at 8:01 am
The technology research and consulting firm yearly recognizes cutting-edge Software ... Parasoft's latest innovation applies AI/Machine Learning to the process of reviewing static analysis findings.
- Automated Machine Learning (AutoML) Marketon April 6, 2020 at 6:40 pm
Automated Machine Learning (AutoML) Market Research Report: By Offering (Platform, Service), Deployment Type (On-Premises, Cloud), Enterprise Size (Large Enterprise, Small & Medium Enterprise), ...
- Perceive Exits Stealth With Super Efficient Machine Learning Chip For Smarter Deviceson April 6, 2020 at 4:28 pm
Running today’s advanced neural networks has been significantly beyond the reach of many small and battery-powered devices.
- Combining automated microfluidic experimentation with machine learning for efficient polymerization designon April 6, 2020 at 8:10 am
Finding the best ratio of ingredients for polymerization reactions can be time consuming and wasteful. An automated microreactor process with integrated machine learning analysis initiates reactions, ...
- Identifying degradation patterns of lithium ion batteries from impedance spectroscopy using machine learningon April 6, 2020 at 2:07 am
Forecasting the state of health and remaining useful life of batteries is a challenge that limits technologies such as electric vehicles. Here, the authors build an accurate battery performance ...
- A 3-minute guide on Machine-learning-led marketingon April 5, 2020 at 5:21 pm
This whitepaper, A 3-minute Guide on Machine-Learning-led Marketing unpacks practical steps and tips on how to increase ROI through a smarter marketing strategy based on your 1st party data. Learn ...
- Seattle machine learning startup OctoML raises $15M from Amplify and Madronaon April 3, 2020 at 8:04 am
“The ‘secret’ sauce in our technology is to use ML to optimize ML ... “This is a hard time to be positive about the future but we see very good things ahead,” Ceze said. “Machine learning and ...
- Machine Learning as a Service Market Emerging Trends & Global Industry Forecast to 2025on April 3, 2020 at 12:38 am
New York, April 03, 2020: The Machine Learning as a Service Market is expected to exceed more than US$ 7500 Million by ...
- Google is using machine learning to improve the quality of Duo callson April 2, 2020 at 10:04 am
Google says that 99 percent of Duo calls experience packet loss: 20 percent of these lose over 3 percent of their audio, and 10 percent lose over 8 percent. That’s a lot of audio to replace. Every ...
- How To Leverage Artificial Intelligence And Machine Learning During A Pandemicon April 2, 2020 at 7:08 am
After the COVID 19 crisis is over, business success or failure may come down to whether companies have taken advantage of Artificial Intelligence (AI) and Machine Learning (ML) technologies.
via Bing News