Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

UniMMVSR: A Unified Multi-Modal Framework for Cascaded Video Super-Resolution – Takara TLDR

Training-Free Group Relative Policy Optimization – Takara TLDR

Singapore company allegedly helped China smuggle $2 billion worth of Nvidia AI processors, report claims — Nvidia denies that the accused has any China ties, but a U.S. investigation is underway

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Business AI
    • Advanced AI News Features
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
Google Gemma

Google announces full launch of Gemma 3n, its mobile-focused AI model

By Advanced AI EditorJune 27, 2025No Comments3 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Google has announced the full release of Gemma 3n, its latest on-device AI model, delivering multimodal capabilities directly to smartphones and other edge devices. The AI model was first previewed last month.

“Building on this incredible momentum, we’re excited to announce the full release of Gemma 3n. While last month’s preview offered a glimpse, today unlocks the full power of this mobile-first architecture. Gemma 3n is designed for the developer community that helped shape Gemma. It’s supported by your favorite tools including Hugging Face Transformers, llama.cpp, Google AI Edge, Ollama, MLX, and many others, enabling you to fine-tune and deploy for your specific on-device applications with ease. This post is the developer deep dive: we’ll explore some of the innovations behind Gemma 3n, share new benchmark results, and show you how to start building today,” the company announced in a blog post.

The Gemma 3n comes with a new architectural design termed MatFormer, short for Matryoshka Transformer. Google explains this structure by likening it to Russian nesting dolls: the model contains smaller, fully functional sub-models nested within larger ones. This design grants developers the flexibility to scale performance dynamically based on the available hardware. Gemma 3n is currently available in two primary versions: E2B, which operates efficiently with as little as 2GB of memory, and E4B, requiring approximately 3GB.

Despite their raw parameter counts of 5 billion and 8 billion respectively, these models exhibit resource consumption comparable to much smaller models. This efficiency is further boosted by “Per-Layer Embeddings (PLE),” which can offload certain computational workloads from a device’s graphics processor to its central processor, thereby liberating valuable memory on the accelerator. Additionally, KV Cache Sharing is introduced to accelerate the processing of extended audio and video inputs, a feature Google claims can improve response times by up to two times.

Gemma 3n’s multimodal prowess is a key highlight. For speech-based applications, the model integrates a built-in audio encoder, adapted from Google’s Universal Speech Model. This allows it to perform tasks like speech-to-text conversion and language translation entirely on-device, without an internet connection. Initial evaluations have demonstrated particularly strong performance in translations between English and major European languages, including Spanish, French, Italian, and Portuguese. The audio encoder can process audio in 160-millisecond chunks, enabling detailed analysis of sound context.

The model’s visual understanding is powered by MobileNet-V5, Google’s latest lightweight vision encoder. This system is capable of processing video streams at up to 60 frames per second on devices such as the Google Pixel, enabling smooth, real-time video analysis directly on the device. Despite its optimized size and speed, MobileNet-V5 is reported to surpass earlier vision models in both performance and accuracy. Gemma 3n also supports over 140 languages for text processing and understanding content across 35 languages, setting a new benchmark for accessible on-device AI globally.

Developers can readily access and integrate Gemma 3n using a range of popular tools and frameworks, including Hugging Face Transformers, Ollama, MLX, and llama.cpp. To further stimulate innovation, Google has launched the “Gemma 3n Impact Challenge,” inviting developers to create applications that leverage the model’s offline and multimodal capabilities, with a prize pool of $150,000 for winning entries. This opens up possibilities for AI-powered apps in remote areas where internet connectivity is unreliable or nonexistent, as well as in privacy-sensitive scenarios where transmitting data to cloud-based models is not viable.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleDeepSeek R2 launch delayed: Nvidia chips to blame?
Next Article Paper page – Where to find Grokking in LLM Pretraining? Monitor Memorization-to-Generalization without Test
Advanced AI Editor
  • Website

Related Posts

Google Gemma 3 Outperforms Larger AI Models Like DeepSeek V3

October 9, 2025

New Google Gemma 3: Advanced AI Models for Text and AI Vision

September 7, 2025

Google’s Gemma 3 270M is a compact yet powerful AI model that can run on your toaster

August 24, 2025
Leave A Reply

Latest Posts

The Rubin Names 2025 Art Prize, Research and Art Projects Grants

Kochi-Muziris Biennial Announces 66 Artists for December Exhibition

Instagram Launches ‘Rings’ Awards for Creators—With KAWS as a Judge

Museums Prepare to Close Their Doors as Government Shutdown Continues

Latest Posts

UniMMVSR: A Unified Multi-Modal Framework for Cascaded Video Super-Resolution – Takara TLDR

October 12, 2025

Training-Free Group Relative Policy Optimization – Takara TLDR

October 12, 2025

Singapore company allegedly helped China smuggle $2 billion worth of Nvidia AI processors, report claims — Nvidia denies that the accused has any China ties, but a U.S. investigation is underway

October 12, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • UniMMVSR: A Unified Multi-Modal Framework for Cascaded Video Super-Resolution – Takara TLDR
  • Training-Free Group Relative Policy Optimization – Takara TLDR
  • Singapore company allegedly helped China smuggle $2 billion worth of Nvidia AI processors, report claims — Nvidia denies that the accused has any China ties, but a U.S. investigation is underway
  • Memory Retrieval and Consolidation in Large Language Models through Function Tokens – Takara TLDR
  • When You Tell AI Models to Act Like Women, Most Become More Risk-Averse: Study

Recent Comments

  1. online wettbüro on Tesla threatened in France with claims of ‘deceptive’ practices
  2. Gaylord Hertzler on Class Dismissed? Representative Claims in Getty v. Stability AI | Cooley LLP
  3. Wettstrategien Sportwetten on AI Competitive Self-Play | Two Minute Papers #205
  4. Wettbüro berlin on Annoyed ChatGPT users complain about bot’s relentlessly positive tone
  5. Kala on A Bitter AI Lesson – Compute Reigns Supreme!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.