OpenAI debuts GPT-4.1 family offering 1M token context window

· 2 min read
ChatGPT

OpenAI has introduced three new models available through its API: GPT-4.1, GPT-4.1 mini, and GPT-4.1 nano. These models succeed the GPT-4o series, showing advances in coding capabilities, adherence to instructions, and processing extended context lengths up to 1 million tokens. They also feature updated knowledge through June 2024. Aimed at developers using the API, these models are available now. The company highlights that this release results from collaboration with the developer community to tune the models for practical tasks.

ChatGPT

The GPT-4.1 family demonstrates superior results on various benchmarks compared to GPT-4o and GPT-4o mini. GPT-4.1 shows a significant jump on the SWE-bench Verified coding test (54.6%) and Scale’s MultiChallenge for instruction following (38.3%). All three models handle 1 million token context windows, enabling work with large codebases or documents. GPT-4.1 mini offers substantial performance, sometimes surpassing GPT-4o, but with reduced latency and cost. GPT-4.1 nano is presented as the quickest and most affordable option, suitable for tasks like classification, featuring the 1 million token context despite its small size. These capabilities also make the models suited for building agent systems.

While available via API, OpenAI notes that many underlying improvements are being integrated into the current GPT-4o version within ChatGPT. Concurrent with this launch, OpenAI announced the deprecation of the GPT-4.5 Preview model in the API, set for July 14, 2025, citing GPT-4.1's comparable or better utility at lower cost. Pricing for the new models is lower than GPT-4o, with GPT-4.1 being 26% less expensive for median queries and increased prompt caching discounts. Alpha testers like Qodo reported GPT-4.1 produced better code review suggestions in 55% of cases compared to other models, while Carlyle found it 50% better at data retrieval from large documents.

Source