Smartphones are amazing. They tell us where we’re going, let us know if it’s going to rain, and even act like personal assistants. Now, a new research project from the University of Rochester, New York, aims to make your phone capable of sensing your emotions just by measuring how you’re speaking – not based on what you’re saying.
This research, entitled the Bridge Project, focuses on small changes in the human voice. Rather than using the traditional methodology of self-reporting and monitoring body language, this new method is based on automatic passive emotion detection. This technology can always be listening and monitoring a patient’s emotional state without any work on his or her part – providing a bigger picture on the patient’s entire status.
The basics of the system involve measuring twelve different aspects of speech, and then mapping the data onto six different emotions. Wendi Heinzelman, professor of electrical and computer engineering, said that the project analysed completely emotionless phrases of speech, such as saying dates of the month. Impressively, they are able to reach an 81 per cent accuracy rating with this model, while previous attempts were only around 55 per cent. By having actors read scripts with certain performances, the researchers are able to tweak their algorithm to associate certain pitches, volumes, and harmonics to a specific emotional state.
While this is undoubtedly an invaluable tool for psychologists and medical researchers, it also has huge potential for consumers.
Consider Apple’s Siri. It’s designed to appear more human-like by offering humorous answers, apologising, and using more realistic speech, like “Let’s hear some Beatles,” instead of something with less flare, like “Now playing: The Beatles, track one.”
This gives us a better experience because it mimics human interaction. Now, think about this technology integrated into Siri. When you’re getting frustrated, it could offer simple hints on how to interact better. When you’re sad, it could throw in compliments.
In Dr. Oliver Sacks’s book The Man Who Mistook His Wife For a Hat, he tells a story about a group of patients suffering from aphasia – an inability to understand words. In the story, he details how very capable these patients are in detecting emotion through speech. In fact, they are able to use sound cues to effectively communicate with their loved ones and doctors despite not being able to understand the words directly.
He even notes that it is extremely difficult to execute a lie in front of an aphasiac because they are so adept at picking up the hidden emotion. This story truly illustrates how much of our emotional states are expressed verbally, and just how useful this research really is.
Published under license from Ziff Davis, Inc., New York, All rights reserved.
Copyright © 2012-2013 Ziff Davis, Inc