Name | Modified | Size | Downloads / Week |
---|---|---|---|
Parent folder | |||
Ollama.dmg | 2025-09-19 | 47.9 MB | |
OllamaSetup.exe | 2025-09-19 | 1.2 GB | |
ollama-linux-arm64.tgz | 2025-09-19 | 1.9 GB | |
ollama-linux-arm64-jetpack6.tgz | 2025-09-19 | 359.6 MB | |
ollama-linux-arm64-jetpack5.tgz | 2025-09-19 | 450.2 MB | |
ollama-linux-amd64.tgz | 2025-09-19 | 1.9 GB | |
ollama-linux-amd64-rocm.tgz | 2025-09-19 | 1.1 GB | |
ollama-darwin.tgz | 2025-09-19 | 25.1 MB | |
ollama-windows-arm64.zip | 2025-09-19 | 22.0 MB | |
Ollama-darwin.zip | 2025-09-19 | 47.8 MB | |
ollama-windows-amd64-rocm.zip | 2025-09-19 | 257.5 MB | |
ollama-windows-amd64.zip | 2025-09-19 | 1.9 GB | |
sha256sum.txt | 2025-09-19 | 1.1 kB | |
README.md | 2025-09-18 | 1.2 kB | |
v0.12.0 source code.tar.gz | 2025-09-18 | 11.0 MB | |
v0.12.0 source code.zip | 2025-09-18 | 11.4 MB | |
Totals: 16 Items | 9.3 GB | 63 |
Cloud models
Cloud models are now available in preview, allowing you to run a group of larger models with fast, datacenter-grade hardware.
To run a cloud model, use:
ollama run qwen3-coder:480b-cloud
What's Changed
- Models with the Bert architecture now run on Ollama's engine
- Models with the Qwen 3 architecture now run on Ollama's engine
- Fix issue where older NVIDIA GPUs would not be detected if newer drivers were installed
- Fixed issue where models would not be imported correctly with
ollama create
- Ollama will skip parsing the initial
<think>
if provided in the prompt for /api/generate by @rick-github
New Contributors
- @egyptianbman made their first contribution in https://github.com/ollama/ollama/pull/12300
- @russcoss made their first contribution in https://github.com/ollama/ollama/pull/12280
Full Changelog: https://github.com/ollama/ollama/compare/v0.11.11...v0.12.0