Name | Modified | Size | Downloads / Week |
---|---|---|---|
Parent folder | |||
README.md | 2025-05-30 | 2.6 kB | |
v1.6.1 source code.tar.gz | 2025-05-30 | 32.6 MB | |
v1.6.1 source code.zip | 2025-05-30 | 33.3 MB | |
Totals: 3 Items | 65.9 MB | 1 |
What's new in 1.6.1 (2025-05-30)
These are the changes in inference v1.6.1.
New features
- FEAT: llama.cpp backend support multimodal by @codingl2k1 in https://github.com/xorbitsai/inference/pull/3442
- FEAT: Auto ngl for llama.cpp backend by @codingl2k1 in https://github.com/xorbitsai/inference/pull/3518
- FEAT: [UI] add hint for common parameters with support for custom input. by @yiboyasss in https://github.com/xorbitsai/inference/pull/3521
- FEAT: add some other paraformer series models by @leslie2046 in https://github.com/xorbitsai/inference/pull/3536
- FEAT: support Deepseek-R1-0528 by @Jun-Howie in https://github.com/xorbitsai/inference/pull/3539
- FEAT: support deepseek-r1-0528-qwen3 by @Jun-Howie in https://github.com/xorbitsai/inference/pull/3552
Enhancements
- ENH: [rerank] add instruction for minicpm-reranker by @llyycchhee in https://github.com/xorbitsai/inference/pull/3453
- ENH: pass extra arguments for speech2text API. by @leslie2046 in https://github.com/xorbitsai/inference/pull/3516
- ENH: add modelscope support for kolors by @qinxuye in https://github.com/xorbitsai/inference/pull/3534
- ENH: remove check when specified GPU index for vllm by @kota-iizuka in https://github.com/xorbitsai/inference/pull/3527
- ENH: Supports
HybridCache
intransformers
lib, mainly forgemma3
chat model by @ChengjieLi28 in https://github.com/xorbitsai/inference/pull/3538 - ENH: support virtualenv for chattts by @qinxuye in https://github.com/xorbitsai/inference/pull/3541
- BLD: fix setup.cfg by @qinxuye in https://github.com/xorbitsai/inference/pull/3467
- BLD: update flashinfer version by @amumu96 in https://github.com/xorbitsai/inference/pull/3549
- REF: Refactor for multimodal llm models by @ChengjieLi28 in https://github.com/xorbitsai/inference/pull/3462
Bug fixes
- BUG: fix input for jina clip by @llyycchhee in https://github.com/xorbitsai/inference/pull/3440
- BUG: [ui] delete cache file white screen bug. by @yiboyasss in https://github.com/xorbitsai/inference/pull/3482
- BUG: fix import_submodules, ignore test files by @Gmgge in https://github.com/xorbitsai/inference/pull/3545
Documentation
- DOC: remove llama-cpp-python related doc & refine model_ability parts by @qinxuye in https://github.com/xorbitsai/inference/pull/3519
- DOC: Update doc about cosyvoice-2.0 stream and auto NGL by @codingl2k1 in https://github.com/xorbitsai/inference/pull/3547
New Contributors
- @kota-iizuka made their first contribution in https://github.com/xorbitsai/inference/pull/3527
Full Changelog: https://github.com/xorbitsai/inference/compare/v1.6.0...v1.6.1