Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

The Future Of Sound Is Not Recorded. It is Computed.

Parameter Count 1T, Alibaba Officially Introduces ‘Qwen3-Max-Preview’, the Strongest Language Model in the Tongyi Qianwen Series_model_night

Your last chance to exhibit at Disrupt 2025 is today

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • OpenAI (GPT-4 / GPT-4o)
    • Anthropic (Claude 3)
    • Google DeepMind (Gemini)
    • Meta (LLaMA)
    • Cohere (Command R)
    • Amazon (Titan)
    • IBM (Watsonx)
    • Inflection AI (Pi)
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • AI Experts
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • The TechLead
    • Matt Wolfe AI
    • Andrew Ng
    • OpenAI
    • Expert Blogs
      • François Chollet
      • Gary Marcus
      • IBM
      • Jack Clark
      • Jeremy Howard
      • Melanie Mitchell
      • Andrew Ng
      • Andrej Karpathy
      • Sebastian Ruder
      • Rachel Thomas
      • IBM
  • AI Tools
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
  • AI Policy
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
  • Business AI
    • Advanced AI News Features
    • Finance AI
    • Healthcare AI
    • Education AI
    • Energy AI
    • Legal AI
LinkedIn Instagram YouTube Threads X (Twitter)
Advanced AI News
Alibaba Cloud (Qwen)

Has the Open Source Route Been Successful?_Points_model_Qwen’s

By Advanced AI EditorSeptember 6, 2025No Comments9 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


1. Alibaba has released its largest model to date, surpassing one trillion parameters, with programming capabilities now exceeding Claude, proving that Scaling Law still works. 2. Alibaba’s “Model + Cloud” strategy has created the shortest path from technological research and development to commercial implementation, which is one of the key factors in Qwen’s success as a latecomer. 3. The core challenge of Alibaba’s open-source model lies in balancing openness and profitability. In the future, Qwen not only needs to continue technological breakthroughs but also must prove its value in business models and organizational capabilities.

Yesterday, Anthropic made an extreme move, and Alibaba retaliated late at night with a major announcement—launching its largest model ever, Qwen3-Max-Preview, with over one trillion parameters!

Using the official title from the “Tongyi Model” WeChat public account to describe its capabilities, it is “stronger than just a little.”

From the benchmark test results, Qwen3-Max-Preview has already surpassed its previous model, Qwen3-235B-A22B-2507.

Moreover, the official team has released comparison results with Kimi K2, Claude Opus 4 (Non-thinking), and DeepSeek-V3.1.

From the table below, it is evident that Qwen3-Max-Preview has outperformed other competitors in benchmark tests such as SuperGPQA, AIME2025, LiveCodeBench V6, Arena-Hard V2, and LiveBench.

Especially in programming capabilities, which were previously dominated by Claude, Qwen3-Max-Preview achieved a remarkable lead this time, leaving many netizens astonished.

Qwen’s post on X seems to reveal the “secret”: Scaling works.

01 Real-World Testing of Alibaba’s Largest Model

Currently, Qwen3-Max-Preview is available for experience; you just need to select it from the model dropdown menu:

Experience link: https://chat.qwen.ai

The official API service has also been opened: https://bailian.console.aliyun.com/?tab=model#/model-market (search for Qwen3-Max-Preview).

In terms of actual performance, many netizens from both domestic and international communities have started real-world testing. For example, well-known blogger AK on X integrated Qwen3-Max-Preview into the project AnyCoder on HuggingFace, entering the following prompt:

Design and create a very creative, elaborate, and detailed voxel art scene of a pagoda in a beautiful garden with trees, including some cherry blossoms. Make the scene impressive and varied and use colorful voxels. Use whatever libraries to get this done.

Then, according to AK’s deion, Qwen3-Max-Preview achieved the following effect in one go:

When we entered the following prompt on the official website:

Create a beautiful celebratory landing page for the launch of Qwen3 Max.

In just a few seconds, Qwen3-Max-Preview generated a complete celebration page, with a very fast response time:

Next, we increased the difficulty by posing a classic programming problem—simulating a bouncing ball collision.

We started with a simple one:

Write a java code that shows a ball bouncing inside a spinning hexagon. The ball should be affected by gravity and friction, and it must bounce off the rotating walls realistically, implement it in java and html.

As you can see, the ball moves based on physical laws within the hexagon; and when we use the “up arrow key” to apply force to the ball, it responds immediately.

When we set the number of balls to 10, the generated effect was also quite natural:

Finally, we asked Qwen3-Max-Preview to generate a mini-game:

Make a mini-game like “Angry Birds”.

Although it was successfully generated in one go, perhaps due to the simplicity of the prompt, the mini-game still has some minor flaws, such as the monster’s position not being very accurate; interested friends can try it multiple times.

02 Why Qwen Succeeded as a Latecomer?

From Alibaba’s leap from a hundred billion parameter model to a trillion scale (nearly a fourfold increase) and achieving first place in many assessments, Qwen has firmly established itself in the top tier globally.

However, to be fair, looking at the entire AI large model competition, Alibaba was not the first company in China to launch a product comparable to ChatGPT, but it is undoubtedly one of the latecomers who has surpassed expectations.

Compared to domestic pioneers (such as Baidu), Alibaba’s large model was relatively low-key in its early stages, but its path has been exceptionally clear—building an ecosystem through model open-sourcing while exploring technological frontiers with self-developed closed-source models.

For instance, in the area of open-source, starting from 2023, Qwen has been rapidly releasing multiple versions of models to global developers. From the 7 billion parameter Qwen-7B to 14 billion, 72 billion parameters, and then to visual, audio, and other multimodal models, it almost covers all mainstream sizes and application scenarios. More crucially, Alibaba not only open-sourced the model weights but also opened up commercial licensing, greatly stimulating enthusiasm among small and medium-sized enterprises and individual developers.

This series of actions has allowed it to quickly establish a broad influence in top global open-source communities like Hugging Face, attracting a large number of developers to innovate and develop around the Qwen ecosystem, forming a strong community driving force. This strategy of accumulating resources has earned Tongyi Qwen valuable developer mindshare and application scenario data, which is an advantage that closed-source models find hard to achieve.

But beyond open-sourcing, Alibaba has never ceased exploring the limits of model capabilities internally. As the official statement during the release of the trillion-parameter model said, Scaling works. Behind this is a strong belief in Scaling Law—an exponential increase in model parameters, data volume, and computing power will lead to a qualitative leap in model capabilities.

Training a trillion-parameter model like Qwen3 Max Preview is not merely about resource accumulation; it requires ultimate precision in every detail of stability in ultra-large-scale computing clusters, efficiency of distributed training algorithms, data processing finesse, and engineering optimization.

This is backed by Alibaba’s significant investments over the years in computing power infrastructure and deep accumulation in AI engineering. It is this saturation-level investment that has allowed Qwen to surpass top models like Claude Opus in core capabilities like programming and reasoning.

In addition to the open-source model and capability exploration, Alibaba Cloud is also a key factor in Qwen’s success as a latecomer.

After all, training and inferring large models is a well-known resource-intensive endeavor; Alibaba Cloud provides stable and efficient computing infrastructure for Qwen’s research and development, integrating the full-link tools from data labeling, model development, distributed training to deployment and inference, significantly reducing the engineering burden on R&D teams, allowing them to focus on algorithm and model innovation.

Moreover, regarding model application and popularization, due to Alibaba Cloud’s MaaS strategy, Qwen can quickly penetrate various industries; for instance, enterprise clients can call the Qwen API directly on Alibaba Cloud without training a model from scratch or fine-tune the open-source Qwen model using platform tools to quickly build AI applications.

This “Model + Cloud” strategy has created the shortest path from technological research and development to commercial implementation.

03 But It’s Not Perfect

Although Alibaba’s strategic choices in the development of large models have contributed to its advantages as a latecomer, this does not mean that the current state of Qwen is without risks.

Because Alibaba’s choice of open-source model to attract users and monetize cloud services can be seen as a path filled with both opportunities and challenges; it is similar to Meta’s Llama series, aiming to quickly capture market share and developer mindshare through an open ecosystem, ultimately directing commercial value to its own infrastructure.

This stands in stark contrast to the closed-source + API elite routes of OpenAI, Anthropic, and others, which have the advantage of better protecting core technology, maintaining technological generational differences, and directly generating high profits through high-value API services.

While Alibaba’s open-source strategy can quickly popularize technology, it also means that its most advanced models may struggle to create a significant gap from competitors, and the business model becomes more indirect, requiring clients to first recognize the value of its cloud platform.

The core challenge of the open-source model lies in balancing openness and profitability. When companies can obtain and privately deploy sufficiently good open-source models for free, how strong will their willingness be to pay for official cloud services?

In other words, Alibaba Cloud not only needs to provide simple model hosting but must also offer performance optimization, security guarantees, as well as a robust toolchain and enterprise-level services that far exceed the open-source versions, to build a sufficiently deep moat. How to effectively convert a large open-source user base into high-value paying cloud customers will be the most critical commercial leap on this path.

Aside from commercialization challenges, in today’s fiercely competitive environment for top AI talent, any loss of core personnel could have profound impacts on the team.

In recent years, some key technical talents, including Jia Yangqing in the AI framework and infrastructure field, have left Alibaba to join the entrepreneurial wave or other giants. Although for a company of Alibaba’s size, the departure of a few individuals may not shake its foundation, the negative impact still exists.

After all, the departure of core leaders may affect team morale, send negative signals externally, and increase the difficulty of attracting top talent in the future; changes in leadership in key technical directions may also introduce uncertainties into the long-term strategic continuity of projects.

Just like the ongoing talent war in Silicon Valley, where Meta is involved, the departing talents often become new competitors, fully aware of the strengths and weaknesses of the original system, potentially posing a more precise threat in niche areas.

Therefore, how Alibaba continues to maintain its appeal to top global AI talents under intense competition and establishes a stable and sustainable talent pipeline is a serious issue it must face in its future development.

Conclusion

Overall, Alibaba’s Tongyi Qwen is undoubtedly a top force in the field of large models in China and even globally. With its clear strategy of “parallel open-source and self-research,” backed by Alibaba Cloud’s powerful ecosystem and deep technical talent accumulation, it has successfully secured a leading position in intense competition. The release of the trillion-parameter model further demonstrates its determination and strength in Scaling Law.

However, the path to success is also accompanied by clear challenges. The business model of exchanging open-source for ecosystem still needs to be tested by the market for its sustainability; the pursuit of technological generational differences with closed-source giants like OpenAI will be a long process; and the retention and attraction of top talent is the lifeline for maintaining innovative vitality.

In the future, Qwen not only needs to continue breaking through technically but also must prove its unique value in business models and organizational capabilities. Whether it can transform today’s technological advantage into an unshakeable market dominance tomorrow will be a focal point of interest for the entire industry, including the capital market. This is also the key to whether Alibaba’s future market value can reach new heights.

1. https://x.com/Alibaba_Qwen/status/1963991502440562976

2. https://chat.qwen.ai/

3. https://x.com/_akhaliq/status/1964001592710975971返回搜狐,查看更多



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleShadow AI enters workforce, employees embrace AI adoption: IBM
Next Article Drivel-ology: Challenging LLMs with Interpreting Nonsense with Depth – Takara TLDR
Advanced AI Editor
  • Website

Related Posts

Parameter Count 1T, Alibaba Officially Introduces ‘Qwen3-Max-Preview’, the Strongest Language Model in the Tongyi Qianwen Series_model_night

September 6, 2025

Alibaba Tongyi Qwen-3-Max-Preview Model Launched, Significant Improvements in Inference and Various Performance Aspects_model_has_The

September 6, 2025

Qwen3-Max-Preview Launched, Officially Claimed to be the Most Powerful Language Model in the Tongyi Qianwen Series_model_and_that

September 6, 2025

Comments are closed.

Latest Posts

Tony Shafrazi and the Art of the Comeback

Basquiats Linked to 1MDB Scandal Auctioned by US Government

US Ambassador to UK Fills Residence with Impressionist Masters

New Code of Ethics Implores UK Museums to End Fossil Fuel Sponsorships

Latest Posts

The Future Of Sound Is Not Recorded. It is Computed.

September 6, 2025

Parameter Count 1T, Alibaba Officially Introduces ‘Qwen3-Max-Preview’, the Strongest Language Model in the Tongyi Qianwen Series_model_night

September 6, 2025

Your last chance to exhibit at Disrupt 2025 is today

September 6, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Recent Posts

  • The Future Of Sound Is Not Recorded. It is Computed.
  • Parameter Count 1T, Alibaba Officially Introduces ‘Qwen3-Max-Preview’, the Strongest Language Model in the Tongyi Qianwen Series_model_night
  • Your last chance to exhibit at Disrupt 2025 is today
  • Tencent Hunyuan Launches a New 3D World Model, Dominating the WorldScore Rankings with Its Strength._scenes_image_its
  • Revisiting ServiceNow (NOW) Valuation as U.S. Government Adoption Accelerates

Recent Comments

  1. wildpickle3Nalay on Nvidia takes $4.5bn hit due to export restrictions
  2. RichardDisee on 1-800-CHAT-GPT—12 Days of OpenAI: Day 10
  3. wildpickle3Nalay on New MIT CSAIL study suggests that AI won’t steal as many jobs as expected
  4. wildpickle3Nalay on Artist Stuart Semple Loses Trademark Lawsuit From Yves Klein Estate
  5. zanycactus2Nalay on OpenAI countersues Elon Musk, calls for enjoinment from ‘further unlawful and unfair action’

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

LinkedIn Instagram YouTube Threads X (Twitter)
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.