Scroll Top

In just one second AI knows how you’re feeling from your voice

WHY THIS MATTERS IN BRIEF

In the future as AI figures out human emotions this could be used to create trustful AI’s that manipulate us all.

 

Love the Exponential Future? Join our XPotential Community, future proof yourself with courses from XPotential University, read about exponential tech and trendsconnect, watch a keynote, or browse my blog.

As we see Conversational Artificial Intelligence (AI) emerging increasingly we’re going to be talking to and with the computer systems that are all around us, so having computer systems that understand human emotions is going to be both beneficial and important.

 

See also
Volkswagen to test its quantum computer powered satnav in real traffic

 

Words are important to express ourselves. What we don’t say, however, may be even more instrumental in conveying emotions. Humans can often tell how people around them feel through non-verbal cues embedded in our voice.

Now, researchers in Germany have sought to find out if technical tools, too, can accurately predict emotional undertones in fragments of voice recordings. To do so, they compared three ML models’ accuracy to recognise diverse emotions in audio excepts. Their results were published in Frontiers in Psychology.

“Here we show that machine learning can be used to recognise emotions from audio clips as short as 1.5 seconds,” said the article’s first author Hannes Diemerling, a researcher at the Center for Lifespan Psychology at the Max Planck Institute for Human Development. “Our models achieved an accuracy similar to humans when categorising meaningless sentences with emotional colouring spoken by actors.”

 

See also
Repairing damaged photos becomes a breeze thanks to Nvidia AI

 

The researchers drew non-sensical sentences from two datasets – one Canadian, one German – which allowed them to investigate whether ML models can accurately recognise emotions regardless of language, cultural nuances, and semantic content.

Each clip was shortened to a length of 1.5 seconds, as this is how long humans need to recognise emotion in speech. It is also the shortest possible audio length in which overlapping of emotions can be avoided. The emotions included in the study were joy, anger, sadness, fear, disgust, and neutral.

Based on training data, the researchers generated ML models which worked one of three ways: Deep Neural Networks (DNNs) are like complex filters that analyse sound components like frequency or pitch – for example when a voice is louder because the speaker is angry – to identify underlying emotions.

Convolutional Neural Networks (CNNs) meanwhile scan for patterns in the visual representation of soundtracks, much like identifying emotions from the rhythm and texture of a voice. The hybrid model (C-DNN) merges both techniques, using both audio and its visual spectrogram to predict emotions. The models then were tested for effectiveness on both datasets.

 

See also
Google's new AI image recognition software still has a Cock of the Rock problem

 

“We found that DNNs and C-DNNs achieve a better accuracy than only using spectrograms in CNNs,” Diemerling said. “Regardless of model, emotion classification was correct with a higher probability than can be achieved through guessing and was comparable to the accuracy of humans.”

“We wanted to set our models in a realistic context and used human prediction skills as a benchmark,” Diemerling explained. “Had the models outperformed humans, it could mean that there might be patterns that are not recognizable by us.” The fact that untrained humans and models performed similarly may mean that both rely on resembling recognition patterns, the researchers said.

The present findings also show that it is possible to develop systems that can instantly interpret emotional cues to provide immediate and intuitive feedback in a wide range of situations. This could lead to scalable, cost-efficient applications in various domains where understanding emotional context is crucial, such as therapy and interpersonal communication technology.

 

See also
Researchers used 1980's style ASCII art to jailbreak ChatGPT and Gemini

 

The researchers also pointed to some limitations in their study, for example, that actor-spoken sample sentences may not convey the full spectrum of real, spontaneous emotion. They also said that future work should investigate audio segments that last longer or shorter than 1.5 seconds to find out which duration is optimal for emotion recognition.

Related Posts

Leave a comment

FREE! 2024 TRENDS AND EMERGING TECHNOLOGY CODEXES
+

Awesome! You're now subscribed.

Pin It on Pinterest

Share This