Download Latest Version ollama-linux-arm64-jetpack6.tgz (365.4 MB)
Email in envelope

Get an email when there's a new version of Ollama

Home / v0.12.7
Name Modified Size InfoDownloads / Week
Parent folder
Ollama.dmg 2025-10-29 48.4 MB
OllamaSetup.exe 2025-10-29 1.2 GB
ollama-linux-arm64.tgz 2025-10-29 2.0 GB
ollama-darwin.tgz 2025-10-29 25.6 MB
Ollama-darwin.zip 2025-10-29 48.3 MB
ollama-linux-amd64-rocm.tgz 2025-10-29 1.3 GB
ollama-linux-amd64.tgz 2025-10-29 1.9 GB
ollama-linux-arm64-jetpack5.tgz 2025-10-29 461.6 MB
ollama-linux-arm64-jetpack6.tgz 2025-10-29 365.4 MB
ollama-windows-amd64-rocm.zip 2025-10-29 356.9 MB
ollama-windows-amd64.zip 2025-10-29 1.9 GB
ollama-windows-arm64.zip 2025-10-29 22.2 MB
sha256sum.txt 2025-10-29 1.1 kB
README.md 2025-10-29 2.3 kB
v0.12.7 source code.tar.gz 2025-10-29 20.3 MB
v0.12.7 source code.zip 2025-10-29 20.8 MB
Totals: 16 Items   9.6 GB 20

Ollama screenshot 2025-10-29 at 13 56 55@2x

New models

  • Qwen3-VL: Qwen3-VL is now available in all parameter sizes ranging from 2B to 235B
  • MiniMax-M2: a 230 Billion parameter model built for coding & agentic workflows available on Ollama's cloud

Add files and adjust thinking levels in Ollama's new app

Ollama's new app now includes a way to add one or many files when prompting the model:

Screenshot 2025-10-29 at 2 16 55 PM

For better responses, thinking levels can now be adjusted for the gpt-oss models:

Screenshot 2025-10-29 at 2 12 33 PM

New API documentation

New API documentation is available for Ollama's API: https://docs.ollama.com/api

Screenshot 2025-10-29 at 4 02 53 PM

What's Changed

  • Model load failures now include more information on Windows
  • Fixed embedding results being incorrect when running embeddinggemma
  • Fixed gemma3n on Vulkan backend
  • Increased time allocated for ROCm to discover devices
  • Fixed truncation error when generating embeddings
  • Fixed request status code when running cloud models
  • The OpenAI-compatible /v1/embeddings endpoint now supports encoding_format parameter
  • Ollama will now parse tool calls that don't conform to {"name": name, "arguments": args} (thanks @rick-github!)
  • Fixed prompt processing reporting in the llama runner
  • Increase speed when scheduling models
  • Fixed issue where FROM <model> would not inherit RENDERER or PARSER commands

New Contributors

Full Changelog: https://github.com/ollama/ollama/compare/v0.12.6...v0.12.7-rc0

Source: README.md, updated 2025-10-29