Close Menu
  • Home
  • AI Models
    • DeepSeek
    • xAI
    • OpenAI
    • Meta AI Llama
    • Google DeepMind
    • Amazon AWS AI
    • Microsoft AI
    • Anthropic (Claude)
    • NVIDIA AI
    • IBM WatsonX Granite 3.1
    • Adobe Sensi
    • Hugging Face
    • Alibaba Cloud (Qwen)
    • Baidu (ERNIE)
    • C3 AI
    • DataRobot
    • Mistral AI
    • Moonshot AI (Kimi)
    • Google Gemma
    • xAI
    • Stability AI
    • H20.ai
  • AI Research
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Microsoft Research
    • Meta AI Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Matt Wolfe AI
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Manufacturing AI
    • Media & Entertainment
    • Transportation AI
    • Education AI
    • Retail AI
    • Agriculture AI
    • Energy AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
What's Hot

AI makes us impotent

Stanford HAI’s 2025 AI Index Reveals Record Growth in AI Capabilities, Investment, and Regulation

New MIT CSAIL study suggests that AI won’t steal as many jobs as expected

Facebook X (Twitter) Instagram
Advanced AI News
  • Home
  • AI Models
    • Adobe Sensi
    • Aleph Alpha
    • Alibaba Cloud (Qwen)
    • Amazon AWS AI
    • Anthropic (Claude)
    • Apple Core ML
    • Baidu (ERNIE)
    • ByteDance Doubao
    • C3 AI
    • Cohere
    • DataRobot
    • DeepSeek
  • AI Research & Breakthroughs
    • Allen Institue for AI
    • arXiv AI
    • Berkeley AI Research
    • CMU AI
    • Google Research
    • Meta AI Research
    • Microsoft Research
    • OpenAI Research
    • Stanford HAI
    • MIT CSAIL
    • Harvard AI
  • AI Funding & Startups
    • AI Funding Database
    • CBInsights AI
    • Crunchbase AI
    • Data Robot Blog
    • TechCrunch AI
    • VentureBeat AI
    • The Information AI
    • Sifted AI
    • WIRED AI
    • Fortune AI
    • PitchBook
    • TechRepublic
    • SiliconANGLE – Big Data
    • MIT News
    • Data Robot Blog
  • Expert Insights & Videos
    • Google DeepMind
    • Lex Fridman
    • Meta AI Llama
    • Yannic Kilcher
    • Two Minute Papers
    • AI Explained
    • TheAIEdge
    • Matt Wolfe AI
    • The TechLead
    • Andrew Ng
    • OpenAI
  • Expert Blogs
    • François Chollet
    • Gary Marcus
    • IBM
    • Jack Clark
    • Jeremy Howard
    • Melanie Mitchell
    • Andrew Ng
    • Andrej Karpathy
    • Sebastian Ruder
    • Rachel Thomas
    • IBM
  • AI Policy & Ethics
    • ACLU AI
    • AI Now Institute
    • Center for AI Safety
    • EFF AI
    • European Commission AI
    • Partnership on AI
    • Stanford HAI Policy
    • Mozilla Foundation AI
    • Future of Life Institute
    • Center for AI Safety
    • World Economic Forum AI
  • AI Tools & Product Releases
    • AI Assistants
    • AI for Recruitment
    • AI Search
    • Coding Assistants
    • Customer Service AI
    • Image Generation
    • Video Generation
    • Writing Tools
    • AI for Recruitment
    • Voice/Audio Generation
  • Industry Applications
    • Education AI
    • Energy AI
    • Finance AI
    • Healthcare AI
    • Legal AI
    • Media & Entertainment
    • Transportation AI
    • Manufacturing AI
    • Retail AI
    • Agriculture AI
  • AI Art & Entertainment
    • AI Art News Blog
    • Artvy Blog » AI Art Blog
    • Weird Wonderful AI Art Blog
    • The Chainsaw » AI Art
    • Artvy Blog » AI Art Blog
Advanced AI News
Home » OpenAI’s new GPT-4.1 models can process a million tokens and solve coding problems better than ever
OpenAI

OpenAI’s new GPT-4.1 models can process a million tokens and solve coding problems better than ever

Advanced AI BotBy Advanced AI BotApril 15, 2025No Comments6 Mins Read
Share Facebook Twitter Pinterest Copy Link Telegram LinkedIn Tumblr Email
Share
Facebook Twitter LinkedIn Pinterest Email


OpenAI launched a new family of AI models this morning that significantly improve coding abilities while cutting costs, responding directly to growing competition in the enterprise AI market.

The San Francisco-based AI company introduced three models — GPT-4.1, GPT-4.1 mini, and GPT-4.1 nano — all available immediately through its API. The new lineup performs better at software engineering tasks, follows instructions more precisely, and can process up to one million tokens of context, equivalent to about 750,000 words.

“GPT-4.1 offers exceptional performance at a lower cost,” said Kevin Weil, chief product officer at OpenAI, during Monday’s announcement. “These models are better than GPT-4o on just about every dimension.”

Perhaps most significant for enterprise customers is the pricing: GPT-4.1 will cost 26% less than its predecessor, while the lightweight nano version becomes OpenAI’s most affordable offering at just 12 cents per million tokens.

How GPT-4.1’s improvements target enterprise developers’ biggest pain points

In a candid interview with VentureBeat, Michelle Pokrass, post training research lead at OpenAI, emphasized that practical business applications drove the development process.

“GPT-4.1 was trained with one goal: being useful for developers,” Pokrass told VentureBeat. “We’ve found GPT-4.1 is much better at following the kinds of instructions that enterprises use in practice, which makes it much easier to deploy production-ready applications.”

This focus on real-world utility is reflected in benchmark results. On SWE-bench Verified, which measures software engineering capabilities, GPT-4.1 scored 54.6% — a substantial 21.4 percentage point improvement over GPT-4o.

For businesses developing AI agents that work independently on complex tasks, the improvements in instruction following are particularly valuable. On Scale’s MultiChallenge benchmark, GPT-4.1 scored 38.3%, outperforming GPT-4o by 10.5 percentage points.

Why OpenAI’s three-tiered model strategy challenges competitors like Google and Anthropic

The introduction of three distinct models at different price points addresses the diversifying AI marketplace. The flagship GPT-4.1 targets complex enterprise applications, while mini and nano versions address use cases where speed and cost efficiency are priorities.

“Not all tasks need the most intelligence or top capabilities,” Pokrass told VentureBeat. “Nano is going to be a workhorse model for use cases like autocomplete, classification, data extraction, or anything else where speed is the top concern.”

Simultaneously, OpenAI announced plans to deprecate GPT-4.5 Preview — its largest and most expensive model released just two months ago — from its API by July 14. The company positioned GPT-4.1 as a more cost-effective replacement that delivers “improved or similar performance on many key capabilities at much lower cost and latency.”

This move allows OpenAI to reclaim computing resources while providing developers a more efficient alternative to its costliest offering, which had been priced at $75 per million input tokens and $150 per million output tokens.

Real-world results: How Thomson Reuters, Carlyle and Windsurf are leveraging GPT-4.1

Several enterprise customers who tested the models prior to launch reported substantial improvements in their specific domains.

Thomson Reuters saw a 17% improvement in multi-document review accuracy when using GPT-4.1 with its legal AI assistant, CoCounsel. This enhancement is particularly valuable for complex legal workflows involving lengthy documents with nuanced relationships between clauses.

Financial firm Carlyle reported 50% better performance on extracting granular financial data from dense documents — a critical capability for investment analysis and decision-making.

Varun Mohan, CEO of coding tool provider Windsurf (formerly Codeium), shared detailed performance metrics during the announcement.

“We found that GPT-4.1 reduces the number of times that it needs to read unnecessary files by 40% compared to other leading models, and also modifies unnecessary files 70% less,” Mohan said. “The model is also surprisingly less verbose… GPT-4.1 is 50% less verbose than other leading models.”

Million-token context: What businesses can do with 8x more processing capacity

All three models feature a context window of one million tokens — eight times larger than GPT-4o’s 128,000 token limit. This expanded capacity allows the models to process multiple lengthy documents or entire codebases at once.

In a demonstration, OpenAI showed GPT-4.1 analyzing a 450,000-token NASA server log file from 1995, identifying an anomalous entry hiding deep within the data. This capability is particularly valuable for tasks involving large datasets, such as code repositories or corporate document collections.

However, OpenAI acknowledges performance degradation with extremely large inputs. On its internal OpenAI-MRCR test, accuracy dropped from around 84% with 8,000 tokens to 50% with one million tokens.

How the enterprise AI landscape is shifting as Google, Anthropic and OpenAI compete for developers

The release comes as competition in the enterprise AI space heats up. Google recently launched Gemini 2.5 Pro with a comparable one-million-token context window, while Anthropic’s Claude 3.7 Sonnet has gained traction with businesses seeking alternatives to OpenAI’s offerings.

Chinese AI startup DeepSeek also recently upgraded its models, putting additional pressure on OpenAI to maintain its leadership position.

“It’s been really cool to see how improvements in long context understanding have translated into better performance on specific verticals like legal analysis and extracting financial data,” Pokrass said. “We’ve found it’s critical to test our models beyond the academic benchmarks and make sure they perform well with enterprises and developers.”

By releasing these models specifically through its API rather than ChatGPT, OpenAI signals its commitment to developers and enterprise customers. The company plans to gradually incorporate features from GPT-4.1 into ChatGPT over time, but the primary focus remains on providing robust tools for businesses building specialized applications.

To encourage further research in long-context processing, OpenAI is releasing two evaluation datasets: OpenAI-MRCR for testing multi-round coreference abilities and Graphwalks for evaluating complex reasoning across lengthy documents.

For enterprise decision-makers, the GPT-4.1 family offers a more practical, cost-effective approach to AI implementation. As organizations continue integrating AI into their operations, these improvements in reliability, specificity, and efficiency could accelerate adoption across industries still weighing implementation costs against potential benefits.

While competitors chase larger, costlier models, OpenAI’s strategic pivot with GPT-4.1 suggests the future of AI may not belong to the biggest models, but to the most efficient ones. The real breakthrough may not be in the benchmarks, but in bringing enterprise-grade AI within reach of more businesses than ever before.

Daily insights on business use cases with VB Daily

If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.

Read our Privacy Policy

Thanks for subscribing. Check out more VB newsletters here.

An error occured.



Source link

Follow on Google News Follow on Flipboard
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Copy Link
Previous ArticleHDC: Hierarchical Distillation for Multi-level Noisy Consistency in Semi-Supervised Fetal Ultrasound Segmentation
Next Article ‘AI will be applicable to almost every field’
Advanced AI Bot
  • Website

Related Posts

What Is ChatGPT? Everything You Need to Know About OpenAI’s Popular Chatbot

June 7, 2025

What Is ChatGPT? Everything You Need to Know About OpenAI’s Popular Chatbot

June 7, 2025

Sam Altman’s Brief Ouster at OpenAI Is Getting the Movie Treatment

June 7, 2025
Leave A Reply Cancel Reply

Latest Posts

Jiaxing Train Station By Architect Ma Yansong Is A Model Of People-Centric, Green Urban Design

Hugh Jackman And Sonia Friedman Boldly Bid To Democratize Theater

Men’s Swimwear Gets Casual At Miami Swim Week 2025

Original Prototype for Jane Birkin’s Hermes Bag Consigned to Sotheby’s

Latest Posts

AI makes us impotent

June 7, 2025

Stanford HAI’s 2025 AI Index Reveals Record Growth in AI Capabilities, Investment, and Regulation

June 7, 2025

New MIT CSAIL study suggests that AI won’t steal as many jobs as expected

June 7, 2025

Subscribe to News

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

Welcome to Advanced AI News—your ultimate destination for the latest advancements, insights, and breakthroughs in artificial intelligence.

At Advanced AI News, we are passionate about keeping you informed on the cutting edge of AI technology, from groundbreaking research to emerging startups, expert insights, and real-world applications. Our mission is to deliver high-quality, up-to-date, and insightful content that empowers AI enthusiasts, professionals, and businesses to stay ahead in this fast-evolving field.

Subscribe to Updates

Subscribe to our newsletter and never miss our latest news

Subscribe my Newsletter for New Posts & tips Let's stay updated!

YouTube LinkedIn
  • Home
  • About Us
  • Advertise With Us
  • Contact Us
  • DMCA
  • Privacy Policy
  • Terms & Conditions
© 2025 advancedainews. Designed by advancedainews.

Type above and press Enter to search. Press Esc to cancel.