GLM-4.5: Open-source LLM for intelligent agents by Z.ai
Personalize Any Characters with a Scalable Diffusion Transformer
Industrial-level controllable zero-shot text-to-speech system
From Images to High-Fidelity 3D Assets
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI
Tencent Hunyuan Multimodal diffusion transformer (MM-DiT) model
Open-source large language model family from Tencent Hunyuan
High-Resolution 3D Assets Generation with Large Scale Diffusion Models
Diversity-driven optimization and large-model reasoning ability
Large Multimodal Models for Video Understanding and Editing
Designed for text embedding and ranking tasks
GLM-4-Voice | End-to-End Chinese-English Conversational Model
CogView4, CogView3-Plus and CogView3(ECCV 2024)
A Customizable Image-to-Video Model based on HunyuanVideo
Models for object and human mesh reconstruction
Pokee Deep Research Model Open Source Repo
GLM-4.5V and GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning
Reference PyTorch implementation and models for DINOv3
Generating Immersive, Explorable, and Interactive 3D Worlds
A Powerful Native Multimodal Model for Image Generation
A series of math-specific large language models of our Qwen2 series
Capable of understanding text, audio, vision, video
Reproduction of Poetiq's record-breaking submission to the ARC-AGI-1
Ling is a MoE LLM provided and open-sourced by InclusionAI
A Unified Framework for Text-to-3D and Image-to-3D Generation