Download Latest Version ollama-linux-arm64-jetpack6.tgz (365.4 MB)
Email in envelope

Get an email when there's a new version of Ollama

Home / v0.12.6
Name Modified Size InfoDownloads / Week
Parent folder
Ollama.dmg 2025-10-16 48.3 MB
OllamaSetup.exe 2025-10-16 1.2 GB
ollama-linux-arm64.tgz 2025-10-16 2.0 GB
ollama-linux-arm64-jetpack6.tgz 2025-10-16 365.4 MB
ollama-linux-arm64-jetpack5.tgz 2025-10-16 461.5 MB
ollama-linux-amd64.tgz 2025-10-16 1.9 GB
ollama-linux-amd64-rocm.tgz 2025-10-16 1.3 GB
ollama-darwin.tgz 2025-10-16 25.5 MB
ollama-windows-arm64.zip 2025-10-16 22.1 MB
ollama-windows-amd64.zip 2025-10-16 1.9 GB
Ollama-darwin.zip 2025-10-16 48.2 MB
ollama-windows-amd64-rocm.zip 2025-10-16 356.9 MB
sha256sum.txt 2025-10-16 1.1 kB
README.md 2025-10-16 1.5 kB
v0.12.6 source code.tar.gz 2025-10-16 11.4 MB
v0.12.6 source code.zip 2025-10-16 11.8 MB
Totals: 16 Items   9.7 GB 102

What's Changed

  • Flash attention is now enabled by default for Gemma 3, improving performance and memory utilization
  • Fixed issue where Ollama would hang while generating responses
  • Fixed issue where qwen3-coder would act in raw mode when using /api/generate or ollama run qwen3-coder <prompt>
  • Fixed qwen3-embedding providing invalid results
  • Ollama will now evict models correctly when num_gpu is set
  • Fixed issue where tool_index with a value of 0 would not be sent to the model

Experimental Vulkan Support

Experimental support for Vulkan is now available when you build locally from source. This will enable additional GPUs from AMD, and Intel which are not currently supported by Ollama. To build locally, install the Vulkan SDK and set VULKAN_SDK in your environment, then follow the developer instructions. In a future release, Vulkan support will be included in the binary release as well. Please file issues if you run into any problems.

New Contributors

Full Changelog: https://github.com/ollama/ollama/compare/v0.12.5...v0.12.6

Source: README.md, updated 2025-10-16