OpenAI launches GPT-4.1 models with better coding
OpenAI has launched GPT-4.1, mini, and nano models with major improvements in coding, instruction following, and long context memory. These API-only models outperform GPT-4.5 and are cheaper to run, with up to 1 million token support and updated knowledge till June 2024.
OpenAI has introduced a new lineup of modelsâGPT-4.1, GPT-4.1 mini, and GPT-4.1 nanoâmarking a big leap forward in how artificial intelligence handles coding, instruction following, and long context memory.
These models are now live on OpenAIâs API and, interestingly, they even outperform the companyâs highly touted GPT-4o in some key areas.
GPT-4.5 to be phased out:
As part of the rollout, OpenAI confirmed it will be turning off GPT-4.5 preview in July, ending the brief chapter of a model that many saw as a bridge between GPT-4 and GPT-4o.
The company didnât reveal pricing details publicly, but said the GPT-4.1 family operates at a much lower cost, making it attractive for businesses looking to scale up AI usage without breaking the bank.
What this means going forward:
If the original GPT-4 wowed users with its leap over GPT-3.5, this new set of models is more like a tune-up: smarter, more efficient, and easier to work with in large-scale production environments.
Long context windows make GPT-4.1 a better fit for technical documentation, research papers, and customer support tools that need to retain long histories.
And with mini and nano options on the table, even budget-conscious developers now have access to smarter LLMs that donât compromise on quality.
As of April 15, 2025, the new GPT-4.1 family is officially live on the API. If youâre building anything AI-related, this might be the right time to test what the new models can actually do.
Why developers are excited:
In a post on X, OpenAI CEO Sam Altman said, âBenchmarks are strong, but we focused on real-world utility, and developers seem very happy.â Thatâs a strong signal that the team prioritised practical value over just topping charts.
The models also come with updated knowledge up to June 2024, a welcome refresh for anyone building apps that need to respond with up-to-date context. And since theyâre only available via API for now, this is aimed squarely at the developer crowd rather than general ChatGPT usersâat least for now.
Smaller models, same DNA:
The announcement isnât just about one model. OpenAI is also rolling out GPT-4.1 mini and GPT-4.1 nano, scaled-down versions that maintain the same core capabilities, but are optimised for lower latency and cost-sensitive use cases. Think mobile apps, embedded systems, or use cases where response speed matters more than raw horsepower.
Theyâre part of OpenAIâs ongoing effort to offer models tailored to different performance and budget needsâa move that will likely appeal to startups and developers experimenting with AI at scale. GPT-4.1 is faster, smarter, and remembers more
At the heart of the update is the flagship GPT-4.1 model. According to OpenAI, the model has shown a 21% improvement over GPT-4o and 27% over GPT-4.5 in pure coding performance, based on internal benchmarks. But itâs not just about better code.
These new models can now process up to 1 million tokensâa major upgrade for handling long documents, in-depth chat histories, or complex instruction chains. Tokens, in simple terms, are chunks of data the AI reads. The more tokens it can handle, the more it can ârememberâ in one go.Â
This is a game-changer for AI agents and applications that need to sift through massive chunks of information without losing context. Announced late Monday, the launch signals a quiet but clear shift in focus:
 improving real-world usefulness for developers while keeping costs down. OpenAI said it will sunset the GPT-4.5 preview in July, as GPT-4.1 offers âimproved or similar performanceâ at a much lower price.
Read More: