Download Latest Version ollama-linux-arm64-jetpack6.tgz (362.3 MB)
Email in envelope

Get an email when there's a new version of Ollama

Home / v0.11.0
Name Modified Size InfoDownloads / Week
Parent folder
sha256sum.txt 2025-08-05 1.1 kB
OllamaSetup.exe 2025-08-05 735.9 MB
Ollama.dmg 2025-08-05 46.4 MB
ollama-windows-arm64.zip 2025-08-05 21.6 MB
ollama-windows-amd64.zip 2025-08-05 1.3 GB
ollama-windows-amd64-rocm.zip 2025-08-05 257.5 MB
ollama-linux-arm64.tgz 2025-08-05 1.1 GB
ollama-linux-arm64-jetpack6.tgz 2025-08-05 362.3 MB
ollama-linux-arm64-jetpack5.tgz 2025-08-05 455.5 MB
ollama-linux-amd64.tgz 2025-08-05 1.3 GB
ollama-linux-amd64-rocm.tgz 2025-08-05 1.1 GB
Ollama-darwin.zip 2025-08-05 46.3 MB
ollama-darwin.tgz 2025-08-05 23.7 MB
README.md 2025-08-05 2.7 kB
v0.11.0 source code.tar.gz 2025-08-05 10.5 MB
v0.11.0 source code.zip 2025-08-05 10.8 MB
Totals: 16 Items   6.9 GB 13

ollama OpenAI gpt-oss

Welcome OpenAI's gpt-oss models

Ollama partners with OpenAI to bring its latest state-of-the-art open weight models to Ollama. The two models, 20B and 120B, bring a whole new local chat experience, and are designed for powerful reasoning, agentic tasks, and versatile developer use cases.

Feature highlights

  • Agentic capabilities: Use the models’ native capabilities for function calling, web browsing (Ollama is providing a built-in web search that can be optionally enabled to augment the model with the latest information), python tool calls, and structured outputs.
  • Full chain-of-thought: Gain complete access to the model's reasoning process, facilitating easier debugging and increased trust in outputs.
  • Configurable reasoning effort: Easily adjust the reasoning effort (low, medium, high) based on your specific use case and latency needs.
  • Fine-tunable: Fully customize models to your specific use case through parameter fine-tuning.
  • Permissive Apache 2.0 license: Build freely without copyleft restrictions or patent risk—ideal for experimentation, customization, and commercial deployment.

Quantization - MXFP4 format

OpenAI utilizes quantization to reduce the memory footprint of the gpt-oss models. The models are post-trained with quantization of the mixture-of-experts (MoE) weights to MXFP4 format, where the weights are quantized to 4.25 bits per parameter. The MoE weights are responsible for 90+% of the total parameter count, and quantizing these to MXFP4 enables the smaller model to run on systems with as little as 16GB memory, and the larger model to fit on a single 80GB GPU.

Ollama is supporting the MXFP4 format natively without additional quantizations or conversions. New kernels are developed for Ollama’s new engine to support the MXFP4 format.

Ollama collaborated with OpenAI to benchmark against their reference implementations to ensure Ollama’s implementations have the same quality.

Get started

You can get started by downloading the latest Ollama version (v0.11)

The model can be downloaded directly in Ollama’s new app or via the terminal:

:::ollama run gpt-oss:20b```

:::ollama run gpt-oss:120b```

What's Changed

Full Changelog: https://github.com/ollama/ollama/compare/v0.10.1...v0.11.0

Source: README.md, updated 2025-08-05