Name | Modified | Size | Downloads / Week |
---|---|---|---|
Parent folder | |||
koboldcpp-linux-x64-oldpc | 2025-06-30 | 468.6 MB | |
koboldcpp-nocuda.exe | 2025-06-30 | 76.0 MB | |
koboldcpp.exe | 2025-06-30 | 575.2 MB | |
koboldcpp-mac-arm64 | 2025-06-30 | 27.3 MB | |
koboldcpp-linux-x64-nocuda | 2025-06-30 | 85.2 MB | |
koboldcpp-linux-x64 | 2025-06-30 | 569.1 MB | |
koboldcpp-oldpc.exe | 2025-06-30 | 403.3 MB | |
koboldcpp-1.95 source code.tar.gz | 2025-06-29 | 31.0 MB | |
koboldcpp-1.95 source code.zip | 2025-06-29 | 31.4 MB | |
README.md | 2025-06-29 | 2.9 kB | |
Totals: 10 Items | 2.3 GB | 5 |
koboldcpp-1.95
- NEW: Added support for Flux Kontext: This is a powerful image editing model based on Flux that can edit images using natural language. Easily replace backgrounds, edit text, or add extra items into your images. You can download a ready-to-use kcppt template here, simply load it into KoboldCpp and all necessary model files will be downloaded on launch. Then open StableUI at http://localhost:5001/sdui, add your prompt, reference images and generate. Thanks to @stduhpf for the sd.cpp implementation!
- Photomaker now supports uploading multiple reference images, same as Kontext. Up to 4 reference images are accepted.
- Merged upstream support and added AutoGuess template for Gemma3n (text only) and ERNIE.
- Further grammar sampling speedups from caching by @Reithan
- Fixed a bug when combining save states with draft models.
- Fixed an issue where prompt processing encountered errors after the KV refactor
- Fixed support for python 3.13 (thanks @tsite)
- Updated Kobold Lite, multiple fixes and improvements
- Fixed Push-to-Talk on mobile, added Toggle to Talk (voice input) option.
- Improved some error handling for aborted streaming
- Fixed some linebreaks in corpo chat mode
- Fixed a bug in thinking regex
- Merged new model support, fixes and improvements from upstream
Download and run the koboldcpp.exe (Windows) or koboldcpp-linux-x64 (Linux), which is a one-file pyinstaller for NVIDIA GPU users.
If you have an older CPU or older NVIDIA GPU and koboldcpp does not work, try oldpc version instead (Cuda11 + AVX1).
If you don't have an NVIDIA GPU, or do not need CUDA, you can use the nocuda version which is smaller.
If you're using AMD, we recommend trying the Vulkan option in the nocuda build first, for best support. Alternatively, you can try koboldcpp_rocm at YellowRoseCx's fork here if you are a Windows user or download our rolling ROCm binary here if you use Linux.
If you're on a modern MacOS (M-Series) you can use the koboldcpp-mac-arm64 MacOS binary.
Click here for .gguf conversion and quantization tools
Run it from the command line with the desired launch parameters (see --help
), or manually select the model in the GUI.
and then once loaded, you can connect like this (or use the full koboldai client):
http://localhost:5001
For more information, be sure to run the program from command line with the --help
flag. You can also refer to the readme and the wiki.