Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

BidMax Launches 0% Commission AI-Powered Real Estate Service to Support South Florida Condo and Homeowner Associations | National Business

Apple’s Lack Of New AI Features At WWDC Is ‘Startling,’ Expert Says

Fine-tune and deploy Meta Llama 3.2 Vision for generative AI-powered web automation using AWS DLCs, Amazon EKS, and Amazon Bedrock

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Industry AI
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
Video Generation

Alibaba unveils Wan2.2, open-source suite for AI video creation

By Advanced AI EditorJuly 29, 2025No Comments3 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Alibaba has announced the release of Wan2.2, an open-source suite of large video generation models based on the Mixture-of-Experts (MoE) architecture.

Model capabilities

The Wan2.2 series includes the text-to-video model Wan2.2-T2V-A14B, the image-to-video model Wan2.2-I2V-A14B, and a hybrid model Wan2.2-TI2V-5B that supports both text-to-video and image-to-video generation in a unified framework. Each model has been designed with a focus on increasing the quality, efficiency and level of user control in generating cinematic-style videos from prompts or images.

Both Wan2.2-T2V-A14B and Wan2.2-I2V-A14B leverage the MoE architecture and use data curated for cinematic aesthetics. These models enable creators to adjust multiple video properties such as lighting, time of day, colour tone, camera angle, frame size, composition, and focal length. According to Alibaba, the models are capable of creating complex movements, including detailed facial expressions and elaborate sports scenes, while following instructions and physical rules more closely than before.

To address computational efficiency in video generation, particularly with long tokens, Wan2.2-T2V-A14B and Wan2.2-I2V-A14B employ a two-expert design throughout the denoising process of diffusion models. One expert focuses on the scene layout under high noise, while the other refines details under low noise conditions. The models operate with a total of 27 billion parameters but only activate 14 billion parameters per step, which the company claims reduces computational consumption by up to half.

Aesthetic tuning

Wan2.2 introduces a cinematic-inspired prompt system that allows users to shape results based on key categories such as lighting, illumination, composition, and colour tone. The company says this approach enables more accurate interpretation and delivery of users’ aesthetic demands throughout the video generation task.

Alibaba has expanded the dataset for Wan2.2, reporting a 65.6% increase in image data and an 83.2% increase in video data compared to the previous version, Wan2.1. This increased dataset is intended to strengthen generalisation and creative diversity, allowing the models to produce more intricate scenes and showcase greater artistic range.

Hybrid model and efficiency

The hybrid model, Wan2.2-TI2V-5B, introduces a dense approach built on a 3D Variational Autoencoder (VAE) architecture, featuring a temporal and spatial compression ratio of 4x16x16. This results in an information compression rate of 64. Alibaba states that the TI2V-5B can generate a five-second, 720P video in several minutes on a single consumer-grade GPU.

“The TI2V-5B can generate a 5-second 720P video in several minutes on a single consumer-grade GPU, enabling efficiency and scalability to developers and content creators.”

Open-source and community engagement

All Wan2.2 models are available for download on Hugging Face, GitHub, and Alibaba Cloud’s open-source platform, ModelScope. Alibaba reports that since open-sourcing four Wan2.1 models in February 2025 and Wan 2.1-VACE in May 2025, its models have collectively achieved over 5.4 million downloads on Hugging Face and ModelScope.

“A major contributor to the global open source community, Alibaba open sourced four Wan2.1 models in February 2025 and Wan 2.1-VACE (Video All-in-one Creation and Editing) in May 2025. To date, the models have attracted over 5.4 million downloads on Hugging Face and ModelScope.”

Alibaba’s release of Wan2.2 underscores its continued activity within the open-source ecosystem and the ongoing development of video generation models aimed at supporting creators and developers globally.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleNo more links, no more scrolling—The browser is becoming an AI Agent
Next Article Mayo Clinic deploys NVIDIA AI to transform medicine | Health
Advanced AI Editor
  • Website

Related Posts

Elon Musk Confirms Grok AI Video Generator Launching in October

July 29, 2025

Reallusion’s AI Render Brings 3D Precision to AI Image and Video Generation in ComfyUI

July 29, 2025

Alibaba Releases Wan2.2 AI Video Generation Models, Combining Mixture-of-Experts and Video Diffusion

July 29, 2025

Comments are closed.

Latest Posts

John Roberts Prevented Firing of National Portrait Gallery Director

At Comic-Con, George Lucas Previews Forthcoming Lucas Museum

Betye Saar Assembles an All-Star Group to Steward Her Legacy

Picasso’s ‘Demoiselles’ May Not Have Been Inspired by African Art

Latest Posts

BidMax Launches 0% Commission AI-Powered Real Estate Service to Support South Florida Condo and Homeowner Associations | National Business

July 29, 2025

Apple’s Lack Of New AI Features At WWDC Is ‘Startling,’ Expert Says

July 29, 2025

Fine-tune and deploy Meta Llama 3.2 Vision for generative AI-powered web automation using AWS DLCs, Amazon EKS, and Amazon Bedrock

July 29, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • BidMax Launches 0% Commission AI-Powered Real Estate Service to Support South Florida Condo and Homeowner Associations | National Business
  • Apple’s Lack Of New AI Features At WWDC Is ‘Startling,’ Expert Says
  • Fine-tune and deploy Meta Llama 3.2 Vision for generative AI-powered web automation using AWS DLCs, Amazon EKS, and Amazon Bedrock
  • NTT DATA and Mistral AI Form Partnership
  • Partner POV | The Economic Benefits of HPE Private Cloud AI with NVIDIA AI Computing by HPE

Recent Comments

  1. binance on OpenAI updates its new Responses API rapidly with MCP support, GPT-4o native image gen, and more enterprise features
  2. binance kód on Anthropic closes $2.5 billion credit facility as Wall Street continues plunging money into AI boom – NBC Los Angeles
  3. 🖨 🔵 Incoming Message: 1.95 Bitcoin from exchange. Claim transfer => https://graph.org/ACTIVATE-BTC-TRANSFER-07-23?hs=40f06aae45d2dc14b01045540f836756& 🖨 on SFC Dialogue丨Jeffrey Sachs says he uses DeepSeek every hour_to_facts_its
  4. 📪 ✉️ Unread Notification: 1.65 BTC from user. Claim transfer >> https://graph.org/ACTIVATE-BTC-TRANSFER-07-23?hs=63f0a8159ef8316c31f5a9a8aca50f39& 📪 on Sean Carroll: Arrow of Time
  5. 🔋 📬 Unread Alert - 1.65 BTC from exchange. Accept funds > https://graph.org/ACTIVATE-BTC-TRANSFER-07-23?hs=db3ef91843302da628b83636ef7db949& 🔋 on Rohit Prasad: Amazon Alexa and Conversational AI | Lex Fridman Podcast #57

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.