Tencent-Hunyuan-Large is the flagship open-source large language model family from Tencent Hunyuan, offering both pre-trained and instruct (fine-tuned) variants. It is designed with long-context capabilities, quantization support, and high performance on benchmarks across general reasoning, mathematics, language understanding, and Chinese / multilingual tasks. It aims to provide competitive capability with efficient deployment and inference. FP8 quantization support to reduce memory usage (~50%) while maintaining precision. High benchmarking performance on tasks like MMLU, MATH, CMMLU, C-Eval, etc.

Features

  • Long context window support: up to 256K tokens in pretrain; 128K tokens for instruct models
  • FP8 quantization support to reduce memory usage (~50%) while maintaining precision
  • Expert-specific learning rate scaling in training (for mixture or expert architectures)
  • High benchmarking performance on tasks like MMLU, MATH, CMMLU, C-Eval, etc.
  • Hugging Face format compatibility for fine-tuning / inference using frameworks like hf-deepspeed, plus support for flash attention, efficient operators (TRT-LLM)
  • Throughput and efficiency improvements: TRT-LLM backend surpasses vLLM by ~30 %, quantized/inference optimizations included

Project Samples

Project Activity

See All Activity >

Follow Tencent-Hunyuan-Large

Tencent-Hunyuan-Large Web Site

Other Useful Business Software
Earn up to 16% annual interest with Nexo. Icon
Earn up to 16% annual interest with Nexo.

Let your crypto work for you

Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
Get started with Nexo.
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of Tencent-Hunyuan-Large!

Additional Project Details

Programming Language

Python

Related Categories

Python Large Language Models (LLM), Python AI Models

Registered

2025-09-23