Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

MIT 6.S094: Recurrent Neural Networks for Steering Through Time

Data Shows What Candidates Want

Microsoft’s New AI Lab Powers Wisconsin Manufacturing – Microsoft (NASDAQ:MSFT)

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Amazon (Titan)
    • Anthropic (Claude 3)
    • Cohere (Command R)
    • Google DeepMind (Gemini)
    • IBM (Watsonx)
    • Inflection AI (Pi)
    • Meta (LLaMA)
    • OpenAI (GPT-4 / GPT-4o)
    • Reka AI
    • xAI (Grok)
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Facebook X (Twitter) Instagram
Advanced AI News
Google Gemma

Google announces full launch of Gemma 3n, its mobile-focused AI model

Advanced AI EditorBy Advanced AI EditorJune 27, 2025No Comments3 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


Google has announced the full release of Gemma 3n, its latest on-device AI model, delivering multimodal capabilities directly to smartphones and other edge devices. The AI model was first previewed last month.

“Building on this incredible momentum, we’re excited to announce the full release of Gemma 3n. While last month’s preview offered a glimpse, today unlocks the full power of this mobile-first architecture. Gemma 3n is designed for the developer community that helped shape Gemma. It’s supported by your favorite tools including Hugging Face Transformers, llama.cpp, Google AI Edge, Ollama, MLX, and many others, enabling you to fine-tune and deploy for your specific on-device applications with ease. This post is the developer deep dive: we’ll explore some of the innovations behind Gemma 3n, share new benchmark results, and show you how to start building today,” the company announced in a blog post.

The Gemma 3n comes with a new architectural design termed MatFormer, short for Matryoshka Transformer. Google explains this structure by likening it to Russian nesting dolls: the model contains smaller, fully functional sub-models nested within larger ones. This design grants developers the flexibility to scale performance dynamically based on the available hardware. Gemma 3n is currently available in two primary versions: E2B, which operates efficiently with as little as 2GB of memory, and E4B, requiring approximately 3GB.

Despite their raw parameter counts of 5 billion and 8 billion respectively, these models exhibit resource consumption comparable to much smaller models. This efficiency is further boosted by “Per-Layer Embeddings (PLE),” which can offload certain computational workloads from a device’s graphics processor to its central processor, thereby liberating valuable memory on the accelerator. Additionally, KV Cache Sharing is introduced to accelerate the processing of extended audio and video inputs, a feature Google claims can improve response times by up to two times.

Gemma 3n’s multimodal prowess is a key highlight. For speech-based applications, the model integrates a built-in audio encoder, adapted from Google’s Universal Speech Model. This allows it to perform tasks like speech-to-text conversion and language translation entirely on-device, without an internet connection. Initial evaluations have demonstrated particularly strong performance in translations between English and major European languages, including Spanish, French, Italian, and Portuguese. The audio encoder can process audio in 160-millisecond chunks, enabling detailed analysis of sound context.

The model’s visual understanding is powered by MobileNet-V5, Google’s latest lightweight vision encoder. This system is capable of processing video streams at up to 60 frames per second on devices such as the Google Pixel, enabling smooth, real-time video analysis directly on the device. Despite its optimized size and speed, MobileNet-V5 is reported to surpass earlier vision models in both performance and accuracy. Gemma 3n also supports over 140 languages for text processing and understanding content across 35 languages, setting a new benchmark for accessible on-device AI globally.

Developers can readily access and integrate Gemma 3n using a range of popular tools and frameworks, including Hugging Face Transformers, Ollama, MLX, and llama.cpp. To further stimulate innovation, Google has launched the “Gemma 3n Impact Challenge,” inviting developers to create applications that leverage the model’s offline and multimodal capabilities, with a prize pool of $150,000 for winning entries. This opens up possibilities for AI-powered apps in remote areas where internet connectivity is unreliable or nonexistent, as well as in privacy-sensitive scenarios where transmitting data to cloud-based models is not viable.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleDeepSeek R2 launch delayed: Nvidia chips to blame?
Next Article Paper page – Where to find Grokking in LLM Pretraining? Monitor Memorization-to-Generalization without Test
Advanced AI Editor
  • Website

Related Posts

Google Releases Gemma 3n Open-Source AI Model That Can Run Locally on 2GB RAM

June 27, 2025

Google Gemma 3 is a new open-source AI that can run on a single GPU

June 26, 2025

Google Gemma 3n is What Apple Intelligence Wants to Be

June 22, 2025
Leave A Reply Cancel Reply

Latest Posts

At Proper Hotels, Come For Vacation, Stay For The Live Music

New EU Law Aimed at Art Trafficking Goes Into Effect on June 28

Peek Inside ‘Leading Hotels Of The World’ With Luxe Travel Book ‘Culture’

Marcia Resnick, Photographer of Downtown Manhattan Scene, Dies at 74

Latest Posts

MIT 6.S094: Recurrent Neural Networks for Steering Through Time

June 27, 2025

Data Shows What Candidates Want

June 27, 2025

Microsoft’s New AI Lab Powers Wisconsin Manufacturing – Microsoft (NASDAQ:MSFT)

June 27, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • MIT 6.S094: Recurrent Neural Networks for Steering Through Time
  • Data Shows What Candidates Want
  • Microsoft’s New AI Lab Powers Wisconsin Manufacturing – Microsoft (NASDAQ:MSFT)
  • Creating Clear AI Priorities and Skills Between Organizations And Employees
  • AWS costs estimation using Amazon Q CLI and AWS Cost Analysis MCP

Recent Comments

No comments to show.

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.