EmoNet: new AI can read 11 emotions from your selfie

Credit: DesignNews

EmoNet is an AI developed by researchers from the University of Colorado and Duke University and designed to identify 11 different emotions just from a picture.

Imagine telling your younger self – let’s say 15-20 years ago – that in the future, robots will be able to identify you by your face to unlock a personal, hand-held computer. Sounds like elite-level tech, right? Not only is this widely-accepted technology in 2019, it’s been surpassed by EmoNet.

EmoNet is an AI that can predict not who you are, but how you feel. University of Colorado and Duke University researchers classified 137,482 video frames from 2,187 videos, into 27 distinct emotion categories, including anxiety and sadness. After the training, the researchers validated their results with 25,000 images.

The team asked 18 people to view over 100 images, whilst measuring their brain activity. They showed the same images to EmoNet and compared the rests to train the AI. Whilst the AI struggled to classify confusion and surprise accurately, it was particularly spot on with horror and sexual desire.

Similar facial expressions, such as amusement and joy, were said to be difficult to classify.

How could EmoNet be used?

One of the most obvious and consumer-focused uses for this research could be to apply this AI in a household robot or assistant. IoT devices that could pick up on your emotions would be more likely to provide assistance related to what you require, or even just leave you alone, should your facial expression suggest that’s what you want.

It’s thought, however, that this research could be more valuable in mental health studies. Depression and anxiety sufferers, for example, who are currently dependent on keeping mood diaries could potentially just have to log their facial expression in future and let the AI work out how they’re feeling.


When it comes to measuring emotions, we’re typically still limited only to asking people how they feel. Our work can help move us towards direct measures of emotion-related brain processes.

Tor Wager, Professor of Psychology and Neuroscience at the University of Colorado Boulder

“Moving away from subjective labels such as ‘anxiety’ and ‘depression’ towards brain processes could lead to new targets for therapeutics, treatments, and interventions,” said Philip Kragel, one of the researchers on the study.

There is big potential too for technology that analyses emotion to be used in customer service. If a robot could sense how you’re feeling, it could deliver a more personalised, tactful customer service. By implementing sensors in shops or restaurants, for example, an organisation may be able to gauge the sentiments of its customers and use this data when helping them.

With emotion-reading AI like EmoNet though comes new challenges. The data that can be gleaned from identifying someone’s emotion is sensitive.


The robots are coming: prepare your business in an hour with our AI trends white paper


Wearing your heart on your sleeve is seen as a good attribute in many cultures, however, a robot reading your emotion could be seen as intrusive. It’s essentially one step away from mind-reading. We volunteer most of our data willingly – providing, of course, that we know we’re doing it – but facial expressions are often an unconscious conveyance of feeling.

Intrusion is a difficult enough topic when it comes to data. Some experts believe there just isn’t enough justification to implement tools like EmoNet into the wider world.

What now for emotional recognition?

The Association for Psychological Science conducted a review recently entitled Emotional Expressions Reconsidered: Challenges to Inferring Emotion From Human Facial Movements. Five scientists were asked to assess emotion science and examine the data as to “whether a person’s emotional state be reasonably inferred from that person’s facial movements.”

In two years of data examination, the findings were that emotional recognition is inconclusive. As is often the case for artificial intelligence, scientists on the review felt it didn’t quite factor in the multifaceted nature of human emotion. Facial features are often just one expression of an emotion and an expression – such as frowning – could be used for several emotions.

AI can easily confuse emotions based on expressions. / Credit: Cnet

“When I say to people, ‘Sometimes you shout in anger, sometimes you cry in anger, sometimes you laugh, and sometimes you sit silently and plan the demise of your enemies,’ that convinces them,” Lisa Feldman Barrett, a Professor of Psychology at Northeastern University and one of the review’s five authors, told The Verge. “I say, ‘Listen, what’s the last time someone won an Academy Award for scowling when they’re angry?’ No one considers that great acting.”

This is almost certainly not the endpoint for EmoNet. The technology is still raw and these findings will no doubt improve over time, as AI develops. The challenges of the technology are clear: will they prove to be too great in the development of emotional recognition?

Luke Conrad

Technology & Marketing Enthusiast

How smart labels are transforming supply chains

Sharath Muddaiah • 27th January 2025

As e-commerce continues to rise globally, the impact of just-in-time manufacturing and rising consumer expectations mean the need for real-time visibility has never been greater. Smart labels directly address this demand, offering solutions to long-standing challenges like shipment delays, theft, and the lack of traceability. With the smart label market projected to grow from $14.1...

The rise of loyalty apps

Sue Azari • 17th January 2025

Increased choice and a consumer more price sensitive than ever before, has made customers far more likely to shop around for the best deals. Price is now the number one factor in brand consideration. In an effort to bag a bargain, loyalty programs have become increasingly popular with consumers, with nine out of ten in...

Rocket launch challenges Elon Musk’s space dominance

Professor Sultan Mahmud • 16th January 2025

Amazon founder Jeff Bezos’s space company has blasted its first rocket into orbit in a bid to challenge the dominance of Elon Musk’s SpaceX. The New Glenn rocket launched from Cape Canaveral Space Force Station in Florida at 02:02 local time (07:02 GMT). It firmly pits the world’s two richest men against each other in...

Giesecke+Devrient launches new Smart Label at CES 2025

Giesecke Devrient • 06th January 2025

G+D has today launched the G+D Smart Label, its innovative tracking solution that transforms any package into an IoT device. Ultra-thin and only slightly larger than a credit card, the new Smart Label proposition has been jointly developed by G+D in conjunction with its hardware partner, Sensos to enable cost-effective, accurate location tracking for a...

Choose an AI solution to transform beyond technology

Kit Cox • 09th December 2024

The first step is knowing exactly what your business wants to achieve with AI; think faster, smarter and more efficient. Once you know what you are working towards, you can start looking for a solution that can help you make it a reality. AI integration can feel like a daunting task at the beginning, so...

A Roadmap to Security and Privacy Compliance

John Lynch Director of Kiteworks • 04th December 2024

Only by understanding the current regulatory environment and implementing robust data protection measures, can organisations enhance their security posture, ensure compliance, and build resilience against the latest cyber threats. This article provides a comprehensive roadmap of how to do it.

Data-Sharing Done Right: Finding the Best Business Approach

Bart Koek • 20th November 2024

To ensure data is not only available, but also accessible to those that need it, businesses recognise that it is vital to focus on collecting, sorting and governing all the data in their organisation. But what happens when data also needs to be accessed and shared across the business? That is where organisations discover a...