Tencent-Hunyuan-Large is the flagship open-source large language model family from Tencent Hunyuan, offering both pre-trained and instruct (fine-tuned) variants. It is designed with long-context capabilities, quantization support, and high performance on benchmarks across general reasoning, mathematics, language understanding, and Chinese / multilingual tasks. It aims to provide competitive capability with efficient deployment and inference. FP8 quantization support to reduce memory usage (~50%) while maintaining precision. High benchmarking performance on tasks like MMLU, MATH, CMMLU, C-Eval, etc.
Features
- Long context window support: up to 256K tokens in pretrain; 128K tokens for instruct models
- FP8 quantization support to reduce memory usage (~50%) while maintaining precision
- Expert-specific learning rate scaling in training (for mixture or expert architectures)
- High benchmarking performance on tasks like MMLU, MATH, CMMLU, C-Eval, etc.
- Hugging Face format compatibility for fine-tuning / inference using frameworks like hf-deepspeed, plus support for flash attention, efficient operators (TRT-LLM)
- Throughput and efficiency improvements: TRT-LLM backend surpasses vLLM by ~30 %, quantized/inference optimizations included
Follow Tencent-Hunyuan-Large
Other Useful Business Software
Earn up to 16% annual interest with Nexo.
Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform.
Geographic restrictions, eligibility, and terms apply.
Rate This Project
Login To Rate This Project
User Reviews
Be the first to post a review of Tencent-Hunyuan-Large!