Ling is a MoE LLM provided and open-sourced by InclusionAI
A Customizable Image-to-Video Model based on HunyuanVideo
Official code base for LeWorldModel: Stable End-to-End Joint-Embedding
Lets make video diffusion practical
GLM-4.5V and GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning
Audio foundation model excelling in audio understanding
GLM-4.5: Open-source LLM for intelligent agents by Z.ai
Open-source multi-speaker long-form text-to-speech model
Pokee Deep Research Model Open Source Repo
Block Diffusion for Ultra-Fast Speculative Decoding
GLM-4 series: Open Multilingual Multimodal Chat LMs
Hunyuan Translation Model Version 1.5
CogView4, CogView3-Plus and CogView3(ECCV 2024)
Mixture-of-Experts Vision-Language Models for Advanced Multimodal
GPT4V-level open-source multi-modal model based on Llama3-8B
A series of math-specific large language models of our Qwen2 series
Advancing Open-source World Models
High-Resolution Image Synthesis with Latent Diffusion Models
Qwen2.5-VL is the multimodal large language model series
Tongyi Deep Research, the Leading Open-source Deep Research Agent
Bidirectional token-classification model for identifiable info
CodeGeeX2: A More Powerful Multilingual Code Generation Model
Open Source Speech Language Model
Long-form streaming TTS system for multi-speaker dialogue generation
Open-source industrial-grade ASR models