Capable of understanding text, audio, vision, video
Repo of Qwen2-Audio chat & pretrained large audio language model
Chat & pretrained large audio language model proposed by Alibaba Cloud
A series of math-specific large language models of our Qwen2 series
Wan2.1: Open and Advanced Large-Scale Video Generative Model
Generating Immersive, Explorable, and Interactive 3D Worlds
High-Resolution 3D Assets Generation with Large Scale Diffusion Models
Qwen3 is the large language model series developed by Qwen team
Qwen3-omni is a natively end-to-end, omni-modal LLM
A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming
Inference framework for 1-bit LLMs
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training
Chinese LLaMA-2 & Alpaca-2 Large Model Phase II Project
Wan2.2: Open and Advanced Large-Scale Video Generative Model
Tongyi Deep Research, the Leading Open-source Deep Research Agent
Qwen3-Coder is the code version of Qwen3
Qwen2.5-VL is the multimodal large language model series
Qwen-Image is a powerful image generation foundation model
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
PyTorch implementation of VALL-E (Zero-Shot Text-To-Speech)
Text-to-3D & Image-to-3D & Mesh Exportation with NeRF + Diffusion
LLaMA: Open and Efficient Foundation Language Models
GLM-4.5: Open-source LLM for intelligent agents by Z.ai
High-Resolution Image Synthesis with Latent Diffusion Models
Implementation of model parallel autoregressive transformers on GPUs