Azure AI Foundry and AOAI (Azure OpenAI Providers) retains on getting higher on a regular basis! The most recent addition in Azure AI Foundry (as of April 14, 2025, yesterday) is the GPT-4.1 mannequin, that has a beneficiant 1M context window and cutoff day for June 2024! 1 million tokens understanding “reminiscence”, which interprets to roughly 1500 pages in a doc! I ponder if the lately printed new summarizing characteristic in Phrase makes use of this mannequin with some trick, or simply has a extremely intelligent trick on summarizing because it ought to assist summarizing paperwork as much as 3 000 pages..
What’s New: GPT-4.1 Overview
GPT-4.1 is the most recent iteration of the GPT-4o mannequin, skilled to excel at coding and instruction-following duties. This mannequin will enhance the standard of agentic workflows and speed up the productiveness of builders throughout all eventualities.
Announcing the GPT-4.1 model series for Azure AI Foundry and GitHub developers
The GPT-4.1 context window of 1 million tokens may be very beneficiant and superior. Because the mannequin helps varied options, it is vitally usable mannequin for brokers, coding and evaluation eventualities. No, it isn’t a reasoning mannequin, if you’re on the lookout for evaluation focus however this one has a really massive context window that does assist to take massive variety of information into consideration. And this helps particularly with coding. If you wish to strive it out, GPT-4.1 is already out there for Copilot in GitHub for public preview.
OpenAI GPT-4.1 is rolling out for all Copilot Plans, together with Copilot Free. You’ll be able to entry it by the mannequin picker in Visible Studio Code and on github.com chat. To speed up your workflow, whether or not you’re debugging, refactoring, modernizing, testing, or simply getting began, choose “GPT-4.1 (Preview)” to start utilizing it.
OpenAI GPT-4.1 now available in public preview for GitHub Copilot and GitHub Models
What’s 1M token context good for, are brokers. As we’re shifting an increasing number of in direction of the world the place AI has a reminiscence, and Responses/Assistants API has been implementing that already, the bigger context counts. We are able to proceed the dialog for an extended interval with extra data included.
Key Options of GPT-4.1
What else is included within the checklist? With 16K token output assist there are
- Textual content, picture processing
- JSON Mode
- parallel perform calling
- Enhanced accuracy and responsiveness
- Parity with English textual content and coding duties in comparison with GPT-4 Turbo with Imaginative and prescient
- Superior efficiency in non-English languages and in imaginative and prescient duties
- Help for enhancements
- Help for advanced structured outputs.
I’m more than happy to see superior efficiency listed for non-English languages and naturally advanced structured outputs with JSON mode will assist huge time with brokers.
From model descriptions page, particulars are

- Textual content & picture enter
- Textual content output
- Chat completions API
- Responses API
- Streaming
- Perform calling
- Structured outputs (chat completions)
What’s odd, is that in Be taught Max Output Tokens are 32k, and within the mannequin description (when deploying) it’s 16K.

Now that I received the mannequin deployed, it is going to be quickly the time to start out testing it.
Pricing and testing

Simply hold within the thoughts, that the billing mannequin is completely different for as much as 128K tokens and to massive one (as much as 1M tokens) context inputs. For the time being, once I was scripting this publish, there wasn’t any details about the pricing in Azure OpenAI Providers Pricing web page. Additionally, I can’t see GPT-4.1-mini nor GPT-4.1-nano fashions but within the catalogue. Along with the usual mannequin, you’re going to get 1M context size with mini and nano as nicely, however with a decrease price (on the expense of effectivity).

For more information, check out Microsoft’s announcement blog article about GPT-4.1. From the article some extra highlights:
- Supervised fine-tuning for GPT-4.1 and 4.1-mini is coming quickly (this week)
- Improved instruction following: The mannequin excels at following detailed directions, particularly brokers containing a number of requests. It’s extra intuitive and collaborative, making it simpler to work with for varied functions.
- Enhanced coding and instruction following: The mannequin is optimized for higher dealing with of advanced technical and coding issues. It generates cleaner, less complicated front-end code, precisely identifies obligatory adjustments in present code, and constantly produces outputs that compile and run efficiently.
Only for a fast check, I connected 4.1 to an information supply and requested about “create an in depth testing plan, that solutions to all recognized dangers within the mission”. There are simply two paperwork: one a couple of threat administration and one with recognized dangers.

One other check I did, was to ask about enhancements to this weblog draft, primarily based the plain textual content model and with a number of connected footage.

And as this weblog publish was written with a stream, I didn’t use AI to generate the primary draft. Asking for higher grammar gave me loads of recommendation.

I mounted a few of these, and that hopefully helped within the readability.
These are simply easy checks, however in time I’ll use this for extra superior eventualities.
Revealed by
I work, weblog and discuss Future Work : AI, Microsoft 365, Copilot, Loop, Azure, and different companies & platforms within the cloud connecting digital and bodily and other people collectively.
I’ve 30 years of expertise in IT enterprise on a number of industries, domains, and roles.
View all posts by Vesa Nopanen