OpenAI has officially introduced a new family of GPT-4.1 models, including GPT-4.1, GPT-4.1 mini, and GPT-4.1 nano. These models are designed specifically for developers and are available exclusively via the OpenAI API. They feature improvements in coding, instruction following, and handling large volumes of data, surpassing the previous GPT-4o and GPT-4o mini models.
One of the main features of GPT-4.1 is the ability to process up to one million tokens at once, enabling the model to work with large amounts of text. This significantly expands opportunities for developers, who can use these models to create more efficient AI agents capable of performing tasks without detailed step-by-step instructions.
OpenAI notes that the new models have reduced latency, making them faster and more cost-effective. For example, GPT-4.1 is twenty-six percent cheaper than GPT-4o for average queries, and GPT-4.1 mini reduces costs by eighty-three percent. This is especially important for developers looking to optimize their AI usage expenses.
Although the new family of models is intended for developers, ChatGPT users can still benefit from improvements that have been integrated into the latest version of GPT-4o in chat. This allows a wide range of users to access new capabilities without switching to the new models.