Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

Meta’s Llama AI Team Suffers Talent Exodus As Top Researchers Join $2B Mistral AI, Backed By Andreessen Horowitz And Salesforce

Reddit Sues Anthropic for Scraping Content to Train Claude AI

Google DeepMind’s CEO Thinks AI Will Make Humans Less Selfish

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Amazon AWS AI
    • Anthropic (Claude)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • Cohere
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Advanced AI News
Home » Text-to-video for AI characters that speak
Video Generation

Text-to-video for AI characters that speak

Advanced AI BotBy Advanced AI BotApril 5, 2025No Comments4 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


ChatGPT’s ability to ignore copyright and common sense while creating images and deepfakes is the talk of the town right now. The image generator model that OpenAI launched last week is so widely used that it’s ruining ChatGPT’s basic functionality and uptime for everyone.

But it’s not just advancements in AI-generated images that we’ve witnessed recently. The Runway Gen-4 video model lets you create incredible clips from a single text prompt and a photo, maintaining character and scene continuity, unlike anything we have seen before.

The videos the company provided should put Hollywood on notice. Anyone can make movie-grade clips with tools like Ruway’s, assuming they work as intended. At the very least, AI can help reduce the cost of special effects for certain movies.

It’s not just Runway’s new AI video tool that’s turning heads. Meta has a MoCha AI product of its own that can be used to create talking AI characters in videos that might be good enough to fool you.

Tech. Entertainment. Science. Your inbox.

Sign up for the most interesting tech & entertainment news out there.

By signing up, I agree to the Terms of Use and have reviewed the Privacy Notice.

MoCha isn’t a type of coffee spelled wrong. It’s short for Movie Character Animator, a research project from Meta and the University of Waterloo. The basic idea of the MoCha AI model is pretty simple. You provide the AI with a text prompt that describes the video and a speech sample. The AI then puts together a video that ensures the characters “speak” the lines in the audio sample almost perfectly.

The researchers provided plenty of samples that show MoCha’s advanced capabilities, and the results are impressive. We have all sorts of clips showing live-action and animated protagonists speaking the lines from the audio sample. Mocha takes into account emotions, and the AI can also support multiple characters in the same scene.

The results are almost perfect, but not quite. There are some visible imperfections in the clips. The eye and face movements are giveaways that we’re looking at AI-generated video. Also, while the lip movement appears to be perfectly synchronized to the audio sample, the movement of the entire mouth is exaggerated compared to real people.

I say that as someone who has seen plenty of similar AI modes from other companies by now, including some incredibly convincing ones.

First, there’s the Runway Gen-4 that we talked about a few days ago. The Gen-4 demo clips are better than MoCha. But that’s a product you can use, MoCha can certainly be improved by the time it becomes a commercial AI model.

Speaking of AI models you can’t use, I always compare new products that can sync AI-generated characters to audio samples to Microsoft’s VASA-1 AI research project, which we saw last April.

VASA-1 lets you turn static photos of real people into videos of speaking characters as long as you provide an audio sample of any kind. Understandably, Microsoft never made the VASA-1 model available to consumers, as such tech opens the door to abuse.

Finally, there’s TikTok’s parent company, ByteDance, which showed a VASA-1-like AI a couple of months ago that does the same thing. It turns a single photo into a fully animated video.

OmniHuman-1 also animates body part movements, something I saw in Meta’s MoCha demo as well. That’s how we got to see Taylor Swift sing the Naruto theme song in Japanese. Yes, it’s a deepfake clip; I’m getting to that.

Products like VASA-1, OmniHuman-1, MoCha, and probably Runway Gen-4 might be used to create deepfakes that can mislead.

Prompt examples for Meta's MoCha AI video generator.
Prompt examples for Meta’s MoCha AI video generator. Image source: Arxiv

Meta researchers working on MoCha and similar projects should address this publicly if and when the model becomes available commercially.

You might spot inconsistencies in the MoCha samples available online, but watch those videos on a smartphone display, and they might not be so evident. Remove your familiarity with AI video generation; you might think some of these MoCha clips were shot with real cameras.

Also important would be the disclosure of the data Meta used to train this AI. The paper said MoCha employed some 500,000 samples, amounting to 300 hours of high-quality speech video samples, without saying where they got that data. Unfortunately, that’s a theme in the industry, not acknowledging the source of the data used to train the AI, and it’s still a concerning one.

You’ll find the full MoCha research paper at this link.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleMIT’s Xstrings facilitates 3D printing parts with embedded actuation | VoxelMatters
Next Article Microsoft has created an AI-generated version of Quake
Advanced AI Bot
  • Website

Related Posts

Samsung Teases Z Fold Ultra, Bing Gets AI Video, and Nothing Sets A Date—Your Gear News of the Week

June 7, 2025

This AI video generator is going viral, and it’s completely free to use

June 7, 2025

A Beginner’s Guide to Automated AI Video Creation

June 7, 2025
Leave A Reply Cancel Reply

Latest Posts

Jiaxing Train Station By Architect Ma Yansong Is A Model Of People-Centric, Green Urban Design

Midwestern Grotto Tradition Celebrated In Sheboygan, WI

Hugh Jackman And Sonia Friedman Boldly Bid To Democratize Theater

Men’s Swimwear Gets Casual At Miami Swim Week 2025

Latest Posts

Meta’s Llama AI Team Suffers Talent Exodus As Top Researchers Join $2B Mistral AI, Backed By Andreessen Horowitz And Salesforce

June 7, 2025

Reddit Sues Anthropic for Scraping Content to Train Claude AI

June 7, 2025

Google DeepMind’s CEO Thinks AI Will Make Humans Less Selfish

June 7, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.