Hume, a startup founded by former Google researcher Alan Cowen, has developed an AI model that can detect and respond to emotional expressions in human speech. Over 1,000 developers and companies, including SoftBank and Lawyer.com, have used Hume’s API to build applications that can interpret a wide range of emotions based on aspects of voice such as tone and rhythm. The startup recently raised $50 million in a funding round and launched a conversational voice API called “Hume EVI” that allows developers to integrate emotion detection into their products.
Hume’s AI aims to generate emotionally attuned responses by adjusting the words and tone of the AI based on the user’s emotional cues. The startup differentiates itself by focusing on identifying underlying expressions instead of just providing generic empathetic responses. Hume’s large language model is trained on data collected from over a million participants across 30 countries to ensure demographic diversity and prevent biases. The AI can interpret emotional tone using external language models and generate responses within milliseconds.
While Hume’s technology appears more advanced than previous emotional detection AI that relied on facial expressions, using AI to interpret human emotions through voice and text is still an imperfect science. Emotional expressions are subjective and influenced by various factors, leading to biased results even with diverse training data. Hume’s AI acknowledges that accurately interpreting tone, intent, and emotional cues in real time is a complex task. The technology is capable of detecting multiple emotional expressions simultaneously, but it may not always be accurate.
Hume’s AI has already been integrated into applications across industries such as health and wellness, customer service, and robotics. In the healthcare sector, the AI is being used to track mental health conditions like depression and borderline personality disorder for patients undergoing experimental treatments. Psychiatrists can use Hume’s models to gain more context on patients’ emotions, especially those that are not easily detectable. The AI has also been integrated into productivity chatbots like Dot, providing users with expanded context on their emotions and offering personalized responses based on their emotional state.
As AI technology continues to advance, the ability to detect and measure human emotions in real time has become increasingly important. Hume’s innovative approach to emotion detection in human speech has potential applications across various industries, from customer service to mental health care. While challenges remain in accurately interpreting emotional cues and avoiding biases, Hume’s AI represents a significant step forward in creating emotionally intelligent conversational interfaces that can enhance user experiences and provide valuable insights into human emotions.