OpenAI has officially brought its AI models to AWS for the first time in the cloud AI landscape. Announced on August 5.
This partnership means AWS customers can now have the power of OpenAI models directly within their existing cloud environment.
What makes it even more exciting is that these GPT-OSS models come with open weights under the Apache 2.0 license. That gives developers full freedom to fine-tune on laptops or private servers, boosting performance and maintaining data privacy.
The larger model, gpt-oss-120B, rivals the performance of OpenAI’s o4-mini and can run efficiently on a single GPU. Meanwhile, the smaller gpt-oss-20B is optimized to run on devices with just 16GB VRAM, making advanced AI more accessible.
AWS claims these models are up to three times more cost-effective than competitors like Google’s Gemini and OpenAI’s API models.
For developers, this isn’t just another model drop; it’s a new era.
This launch is a strategic move by AWS, which until now relied primarily on Anthropic’s Claude models for generative AI. This marks a major turning point not just for AWS, but for the cloud AI industry as a whole.
For more information, follow The Verbo.
FAQs
1. What’s the big news about OpenAI and AWS?
OpenAI models are now available on AWS for the first time, announced on August 5, 2025.
2. Which OpenAI models are included?
The gpt-oss-120B and gpt-oss-20B models are both open-weight under the Apache 2.0 license.
3. Can OpenAI models run locally?
Yes, they’re optimized to run on laptops or private servers, offering more control and privacy.
4. Why do OpenAI models matter for developers?
It’s cost-effective, powerful, and gives developers full flexibility to fine-tune models within AWS or on-premises.