Name | Modified | Size | Downloads / Week |
---|---|---|---|
Parent folder | |||
textgen-portable-3.9-windows-cuda12.4.zip | 2025-08-06 | 833.6 MB | |
textgen-portable-3.9-windows-cuda11.7.zip | 2025-08-06 | 722.1 MB | |
textgen-portable-3.9-windows-vulkan.zip | 2025-08-06 | 202.5 MB | |
textgen-portable-3.9-windows-cpu.zip | 2025-08-06 | 193.5 MB | |
textgen-portable-3.9-linux-cuda11.7.zip | 2025-08-06 | 766.6 MB | |
textgen-portable-3.9-linux-cuda12.4.zip | 2025-08-06 | 838.5 MB | |
textgen-portable-3.9-macos-x86_64.zip | 2025-08-06 | 164.8 MB | |
textgen-portable-3.9-linux-vulkan.zip | 2025-08-06 | 240.6 MB | |
textgen-portable-3.9-linux-cpu.zip | 2025-08-06 | 231.5 MB | |
textgen-portable-3.9-macos-arm64.zip | 2025-08-06 | 176.5 MB | |
README.md | 2025-08-06 | 1.8 kB | |
v3.9 source code.tar.gz | 2025-08-06 | 24.9 MB | |
v3.9 source code.zip | 2025-08-06 | 25.0 MB | |
Totals: 13 Items | 4.4 GB | 2 |
Experimental GPT-OSS support!
I have obtained some success with the GGUF models under
https://huggingface.co/ggml-org/gpt-oss-20b-GGUF/tree/main https://huggingface.co/ggml-org/gpt-oss-120b-GGUF/tree/main
It may be necessary to re-download those models in the next days if bugs are found, so make sure to recheck those pages.
Changes
- Add a new Reasoning effort UI element in the chat tab, with
low
,medium
, andhigh
options for GPT-OSS - Support standalone .jinja chat templates -- makes it possible to load GPT-OSS through Transformers
- Make web search functional with thinking models
Bug fixes
- Fix an edge case in chat history loading that caused a crash (closes [#7155])
- Handle both int and str types in grammar char processing (fixes a rare crash when using grammar)
Backend updates
- Update llama.cpp to https://github.com/ggml-org/llama.cpp/tree/fd1234cb468935ea087d6929b2487926c3afff4b
- Update Transformers to 4.55 (adds GPT-OSS support)
Portable builds
Below you can find portable builds: self-contained packages that work with GGUF models (llama.cpp) and require no installation! Just download the right version for your system, unzip, and run.
Which version to download:
- Windows/Linux:
- NVIDIA GPU: Use
cuda12.4
for newer GPUs orcuda11.7
for older GPUs and systems with older drivers. - AMD/Intel GPU: Use
vulkan
builds. -
CPU only: Use
cpu
builds. -
Mac:
- Apple Silicon: Use
macos-arm64
. - Intel CPU: Use
macos-x86_64
.
Updating a portable install:
- Download and unzip the latest version.
- Replace the
user_data
folder with the one in your existing install. All your settings and models will be moved.