Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

C3 AI Stock Surges 24% On $450 Million Defense Deal

Nvidia Faces $8B Hit as U.S. Halts H20 AI Chip Exports to China

Paper page – Multi-Domain Explainability of Preferences

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Amazon AWS AI
    • Anthropic (Claude)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • Cohere
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Advanced AI News
Home » Emotive voice AI startup Hume launches new EVI 3 model with rapid custom voice creation
VentureBeat AI

Emotive voice AI startup Hume launches new EVI 3 model with rapid custom voice creation

Advanced AI BotBy Advanced AI BotMay 29, 2025No Comments6 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More

New York-based AI startup Hume has unveiled its latest Empathic Voice Interface (EVI) conversational AI model, EVI 3 (pronounced “Evee” Three, like the Pokémon character), targeting everything from powering customer support systems and health coaching to immersive storytelling and virtual companionship.

EVI 3 lets users create their own voices by talking to the model (it’s voice-to-voice/speech-to-speech), and aims to set a new standard for naturalness, expressiveness, and “empathy” according to Hume — that is, how users perceive the model’s understanding of their emotions and its ability to mirror or adjust its own responses, in terms of tone and word choice.

Designed for businesses, developers, and creators, EVI 3 expands on Hume’s previous voice models by offering more sophisticated customization, faster responses, and enhanced emotional understanding.

Individual users can interact with it today through Hume’s live demo on its website and iOS app, but developer access through Hume’s proprietary application programming interface (API) is said to be made available in “the coming weeks,” as a blog post from the company states.

At that point, developers will be able to embed EVI 3 into their own customer service systems, creative projects, or virtual assistants — for a price (see below).

My own usage of the demo allowed me to create a new, custom synthetic voice in seconds based on qualities I described to it — a mix of warm and confident, and a masculine tone. Speaking to it felt more naturalistic and easy than other AI models and certainly the stock voices from legacy tech leaders such Apple with Siri and Amazon with Alexa.

What developers and businesses should know about EVI 3

Hume’s EVI 3 is built for a range of uses—from customer service and in-app interactions to content creation in audiobooks and gaming.

It allows users to specify precise personality traits, vocal qualities, emotional tone, and conversation topics.

This means it can produce anything from a warm, empathetic guide to a quirky, mischievous narrator—down to requests like “a squeaky mouse whispering urgently in a French accent about its scheme to steal cheese from the kitchen.”

EVI 3’s core strength lies in its ability to integrate emotional intelligence directly into voice-based experiences.

Unlike traditional chatbots or voice assistants that rely heavily on scripted or text-based interactions, EVI 3 adapts to how people naturally speak — picking up on pitch, prosody, pauses, and vocal bursts to create more engaging, humanlike conversations.

However, one big feature Hume’s models currently lack — and which is offered by rivals open source and proprietary, such as ElevenLabs — is voice cloning, or the rapid replication of a user’s or other voice, such as a company CEO.

Yet Hume has indicated it will add such a capability to its Octave text-to-speech model, as it is noted as “coming soon” on Hume’s website, and prior reporting by yours truly on the company found it will allow users to replicate voices from as little as five seconds of audio.

Hume has stated it’s prioritizing safeguards and ethical considerations before making this feature broadly available. Currently, this cloning capability is not available in EVI itself, with Hume emphasizing flexible voice customization instead.

Internal benchmarks show users prefer EVI 3 to OpenAI’s GPT-4o voice model

According to Hume’s own tests with 1,720 users, EVI 3 was preferred over OpenAI’s GPT-4o in every category evaluated: naturalness, expressiveness, empathy, interruption handling, response speed, audio quality, voice emotion/style modulation on request, and emotion understanding on request (the “on request” features are covered in “instruction following” seen below).

It also usually bested Google’s Gemini model family and the new open source AI model firm Sesame from former Oculus co-creator Brendan Iribe.

It also boasts lower latency (~300 milliseconds), robust multilingual support (English and Spanish, with more languages coming), and effectively unlimited custom voices. As Hume writes on its website (see screenshot immediately below):

Key capabilities include:

Prosody generation and expressive text-to-speech with modulation.

Interruptibility, enabling dynamic conversational flow.

In-conversation voice customizability, so users can adjust speaking style in real time.

API-ready architecture (coming soon), so developers can integrate EVI 3 directly into apps and services.

Pricing and developer access

Hume offers flexible, usage-based pricing across its EVI, Octave TTS, and Expression Measurement APIs.

While EVI 3’s specific API pricing has not been announced yet (marked as TBA), the pattern suggests it will be usage-based, with enterprise discounts available for large deployments.

For reference, EVI 2 is priced at $0.072 per minute — 30% lower than its predecessor, EVI 1 ($0.102/minute).

For creators and developers working with text-to-speech projects, Hume’s Octave TTS plans range from a free tier (10,000 characters of speech, ~10 minutes of audio) to enterprise-level plans. Here’s the breakdown:

Free: 10,000 characters, unlimited custom voices, $0/month

Starter: 30,000 characters (~30 minutes), 20 projects, $3/month

Creator: 100,000 characters (~100 minutes), 1,000 projects, usage-based overage ($0.20/1,000 characters), $10/month

Pro: 500,000 characters (~500 minutes), 3,000 projects, $0.15/1,000 extra, $50/month

Scale: 2,000,000 characters (~2,000 minutes), 10,000 projects, $0.13/1,000 extra, $150/month

Business: 10,000,000 characters (~10,000 minutes), 20,000 projects, $0.10/1,000 extra, $900/month

Enterprise: Custom pricing and unlimited usage

For developers working on real-time voice interactions or emotional analysis, Hume also offers a Pay as You Go plan with $20 in free credits and no upfront commitment. High-volume enterprise customers can opt for a dedicated Enterprise plan featuring dataset licenses, on-premises solutions, custom integrations, and advanced support.

Hume’s history of emotive AI voice models

Founded in 2021 by Alan Cowen, a former researcher at Google DeepMind, Hume aims to bridge the gap between human emotional nuance and AI interaction.

The company trained its models on an expansive dataset drawn from hundreds of thousands of participants worldwide—capturing not just speech and text, but also vocal bursts and facial expressions.

“Emotional intelligence includes the ability to infer intentions and preferences from behavior. That’s the very core of what AI interfaces are trying to achieve,” Cowen told VentureBeat. Hume’s mission is to make AI interfaces more responsive, humanlike, and ultimately more useful—whether that’s helping a customer navigate an app or narrating a story with just the right blend of drama and humor.

In early 2024, the company launched EVI 2, which offered 40% lower latency and 30% reduced pricing compared to EVI 1, alongside new features like dynamic voice customization and in-conversation style prompts.

February 2025 saw the debut of Octave, a text-to-speech engine for content creators capable of adjusting emotions at the sentence level with text prompts.

With EVI 3 now available for hands-on exploration and full API access just around the corner, Hume hopes to allow developers and creators to reimagine what’s possible with voice AI.

Daily insights on business use cases with VB Daily

If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.

Read our Privacy Policy

Thanks for subscribing. Check out more VB newsletters here.

An error occured.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleCentre Pompidou to Open New Brazil Satellite
Next Article With An Eye Toward More Acquisitions, AI-Powered Grammarly Raises $1B from General Catalyst
Advanced AI Bot
  • Website

Related Posts

Which LLM should you use? Token Monster automatically combines multiple models and tools for you

May 30, 2025

Peer launches Global Simulation as real-time digital Earth with AI agents

May 30, 2025

FLUX.1 Kontext enables in-context image generation for enterprise AI pipelines

May 30, 2025
Leave A Reply Cancel Reply

Latest Posts

Japanese Sculptor Kunimasa Aoki Wins 2025 Loewe Craft Prize

3,000-Year-Old Mayan Complex Uncovered in Guatemala

Wang Chung On ‘Everybody Have Fun Tonight’ Amid New Compilation Set

This Exhibit Shows How Our Relationship With Nature Was Redefined By A Dragonfly

Latest Posts

C3 AI Stock Surges 24% On $450 Million Defense Deal

May 30, 2025

Nvidia Faces $8B Hit as U.S. Halts H20 AI Chip Exports to China

May 30, 2025

Paper page – Multi-Domain Explainability of Preferences

May 30, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.