OpenAI announces two “gpt-oss” open AI models, and you can download them today

Date:

Share:


OpenAI is releasing new generative AI models today, and no, GPT-5 is not one of them. Depending on how you feel about generative AI, these new models may be even more interesting, though. The company is rolling out gpt-oss-120b and gpt-oss-20b, its first open-weight models since the release of GPT-2 in 2019. You can download and run these models on your own hardware, with support for simulated reasoning, tool use, and deep customization.

When you access the company’s proprietary models in the cloud, they’re running on powerful server infrastructure that cannot be replicated easily, even in enterprise. The new OpenAI models come in two variants (120b and 20b) to run on less powerful hardware configurations. Both are transformers with a configurable chain of thought (CoT), supporting low, medium, and high settings. The lower settings are faster and use fewer compute resources, but the outputs are better with the highest setting. You can set the CoT level with a single line in the system prompt.

The smaller gpt-oss-20b has a total of 21 billion parameters, utilizing mixture-of-experts (MoE) to reduce that to 3.6 billion parameters per token. As for gpt-oss-120b, its 117 billion parameters come down to 5.1 billion per token with MoE. The company says the smaller model can run on a consumer-level machine with 16GB or more of memory. To run gpt-oss-120b, you need 80GB of memory, which is more than you’re likely to find in the average consumer machine. It should fit on a single AI accelerator GPU like the Nvidia H100, though. Both models have a context window of 128,000 tokens.

The team says users of gpt-oss can expect robust performance similar to its leading cloud-based models. The larger one benchmarks between the o3 and o4-mini proprietary models in most tests, with the smaller version running just a little behind. It gets closest in math and coding tasks. In the knowledge-based Humanity’s Last Exam, o3 is far out in front with 24.9 percent (with tools), while gpt-oss-120b only manages 19 percent. For comparison, Google’s leading Gemini Deep Think hits 34.8 percent in that test.



Source link

━ more like this

NASA explains how it keeps the Curiosity rover running, 13 years later

Thirteen years ago, the Curiosity rover landed on Mars, inside Gale crater in particular. It was originally sent to the red planet for...

Dell Premium 14 review: New name, same great laptop

Every now and then companies make truly boneheaded decisions, which is exactly and replaced it with something generic. It's like if Ford...

Navigating Croydon’s traffic: Tips for efficient moves – London Business News | Londonlovesbusiness.com

It’s never simple to move, and dealing with Croydon’s traffic may make you feel like the main character in your own comedy of...

What are the common mistakes foreigners make when applying for personal loans? – London Business News | Londonlovesbusiness.com

Foreigners look for personal loans to manage their finances or cover sudden expenses in a new country. The process, however, often feels complicated...

Oil market rebounds on geopolitical tensions as demand risks remain – London Business News | Londonlovesbusiness.com

Crude oil futures rebounded early on Wednesday, underpinned by escalating geopolitical tensions as the United States renewed tariff threats against India over its...
spot_img