If you like the premise of AI doing, well something, in your rig, but don't much fancy feeding your information back [[link]] to a data set for future use, a local LLM is likely the answer to your prayers. With OpenAI's latest model, you can do just that, assuming you have the hardware to power it.
Announced in collaboration with , gpt-oss-10b and gpt-oss-120b are both live and available to download via (or via ). You can access a cloud-based demo of it with toggleable reasoning levels via .
Model name | Transformer blocks | Total parameters | Active params per token | # of experts | Active experts per token | Input context length |
gpt-oss-20b | 24 | 20B | 3.6B | 32 | 4 | 128K |
gpt-oss-120b | 36 | 117B | 5.1B | 128 | 4 | 128k |
Both new OpenAI models (gpt-oss-20b and gpt-oss-120b) are open-weight and they are both reasoning models, which effectively 'think' before giving an answer. This is the same sort of model that's said to be behind agentic AI, essentially breaking down broader questions and tasks into a smaller chain. It's worth noting, these models aren't intended to replace GPT-5, OpenAI's upcoming advanced cloud-based version.
Gpt-oss-120b can reportedly run on an 80 GB GPU, and OpenAI reports it offers similar performance to its o4-mini model. This means can run it, but the likelihood of you getting one on your home rig is unlikely.
Gpt-oss-20b, however, can run on a 16 GB GPU, and OpenAI claims it offers similar performance to the o3-mini "on common benchmarks."
You won't be left behind if you're all Team Red either, as AMD CEO and stated, "AMD is proud to be a Day 0 partner enabling these models to run everywhere - across cloud, edge and clients. The power of open models is clear… and this is a big step forward."
Congrats @sama @OpenAI on today’s launch of gpt-oss! @AMD is proud to be a Day 0 partner enabling these models to run everywhere - across cloud, edge and clients. The power of open models is clear… and this is a big step forward. https://t.co/2XUS5ceP8w
The , or any AMD AI CPU with 32 GB [[link]] of memory, can also run the latest 20b model. If you've ever wanted to run a local AI and have power to spare, these new models may be worth playing around with. The AI Max+ 395, with a 128 GB RAM configuration, can run the full-fat 120b model, and we've been playing with it on a new 128 GB desktop machine sporting that chip, and it's certainly impressive.
The release of these open-weight models comes at a particularly interesting time for AI in general, too. Recently, we saw by setting up data centers in tents (data tenters, if you will) to catch up with its competition. If you don't particularly care about the AI machinations of these tech giants (I get it), recently, with Microsoft's going into beta. As well as this, opening up around the world.
AI is worming its [[link]] way into most facets of digital life, and should you want it to make its way into your PC even when you're offline, you now have that option.

👉👈
1. Best gaming laptop:
2. Best gaming PC:
3. Best handheld gaming PC:
4. Best mini PC:
5. Best VR headset: