An experimental version of DeepSeek model
Large Multimodal Models for Video Understanding and Editing
CogView4, CogView3-Plus and CogView3(ECCV 2024)
Multimodal model achieving SOTA performance
Ling is a MoE LLM provided and open-sourced by InclusionAI
State-of-the-art (SoTA) text-to-video pre-trained model
MiniMax-M2, a model built for Max coding & agentic workflows
Release for Improved Denoising Diffusion Probabilistic Models
Locally run an Instruction-Tuned Chat-Style LLM
Code release for ConvNeXt V2 model
Code release for "Masked-attention Mask Transformer
Code for the paper "Improved Techniques for Training GANs"
Tencent’s 36-language state-of-the-art translation model
685B model with improved agents and consistency
Qwen3-Next: 80B instruct LLM with ultra-long context up to 1M tokens
Jan-v1-edge: efficient 1.7B reasoning model optimized for edge devices