your123 - stock.adobe.com

Belatedly, OpenAI releases two new open weight models

This is a highly anticipated release, as the vendor has received criticism for straying from its open background after GPT-2.

For the first time since 2019, OpenAI has released an open weight model.

On Tuesday, OpenAI launched gpt-oss-120b and gpt-oss-20b, open weight language models available under the Apache 2.0 license.

According to OpenAI, the models demonstrate strong tool use capabilities and are optimized for deployment on consumer hardware. The AI vendor said the models were trained using a mix of reinforcement learning techniques informed by OpenAI's other models, such as o3.

Compared to core reasoning models, gpt-oss-120b is similar to OpenAI’s o4-mini, while the gpt-oss-20b model is similar to o3-mini. The 20b model can also run on edge devices with 16GB of memory, OpenAI said.

OpenAI trained the models on Nvidia H100 GPUs, optimized across the Nvidia stack and can run inference on Nvidia Blackwell and RTX GPUs, Nvidia said.

Anticipated release

OpenAI's open weight models are highly anticipated. They come after heavy criticism from the AI community that the pioneering generative AI (GenAI) vendor has veered far from its original path when it used to produce open models. Other GenAI providers have also capitalized on the growing popularity of open source following the big splash of Chinese vendor’s DeepSeek R1 model earlier this year.

Open weight models are models for which the training parameters are openly available. With open source models, coding sources, training data and parameters are available.

"This is a response to a lot of the open source models out there," said Lian Jye Su, an analyst at Omdia, a division of Informa TechTarget. Many vendors in the U.S and China have released open models, including Google, Alibaba, IBM, Ai2 and others, he noted.

"It's about time," Su said. "It's good to see them contributing to the open source community after being absent for a long time."

Su said the models themselves are typical compared to other models in the market in that OpenAI is using known techniques, such as a mixture of experts.

Some challenges and strategy

However, the models seem limited in terms of the languages they were trained on. They were mostly trained in English. "They may have some problems working in other languages," Rowan Curran, an analyst at Gartner, said.

The models’ release is also significant after Meta CEO Mark Zuckerberg revealed on July 30 that the social media giant won't make its anticipated superintelligence AI models open. This could mean that Meta might no longer release new Meta Llama models. Llama is one of the most widely used open weight model families.

"It's a very nice moment to see this today," Curran said. He added that it's also useful that OpenAI released a larger version for data centers and a miniature version for the edge.

David Nicholson, an analyst at Futurum Group, said the release is also a strategy for OpenAI to target companies that are fine-tuning models. Because OpenAI released gpt-oss-120b and gpt-oss-20b under an Apache license, one could argue that the licensing is less restrictive and more open to commercial activity, Nicholson added.

"It's one thing to say that your model is open weight ... for academic purposes, but as soon as you're going to start making money, that's a whole different story," Nicholson said. "Apache is well known. Enterprises will be more comfortable with how these models are released and the licensing wrapper around them."

He added that the release also speaks to OpenAI's strategy in the tight GenAI race.

"This, from OpenAI, is a very strategic move," Nicholson continued. "It undermines Meta's attempt to creep into the commercial space."

He said that it will be interesting to see the kind of market penetration OpenAI's open-weight models achieve compared to Meta Llama models by the time Meta holds its next LLamacon conference September 17-18.

A call to do more

Despite the excitement about OpenAI's open weight models and how the vendor will contribute to the open source community, there is still a push for ChatGPT creator to do more.

"The quality of their model seems good," said Hannaneh Hajishirzi, senior director of AI at Ai2 and a professor at the University of Washington.

She added that while the models will enable the open source community to build better models, OpenAI should release the model's data, full recipe and process of what worked and didn't work at various stages of training. That would, in essence, make the models fully open source.

“I understand why they might not want to release all these details because almost all the secrets are always in the details," she said. "This is an added benefit of what we mean by fully open."

As part of the release, Microsoft is bringing GPU-optimized versions of the gpt-oss-20b model to Windows. Developers who want to try the model can do so on OpenAI's open model playground and download it on Hugging Face.

The models are also available on Together AI, and gpt-oss-120B is now live on Cerebras.

Esther Shittu is an Informa TechTarget news writer and podcast host covering AI software and systems.

Dig Deeper on AI technologies