A Powerful Native Multimodal Model for Image Generation
A series of math-specific large language models of our Qwen2 series
Inference framework for 1-bit LLMs
Lets make video diffusion practical
Towards Real-World Vision-Language Understanding
Tool for exploring and debugging transformer model behaviors
A state-of-the-art open visual language model
Open-weight, large-scale hybrid-attention reasoning model
Qwen3-omni is a natively end-to-end, omni-modal LLM
PyTorch code and models for the DINOv2 self-supervised learning
Towards Ultimate Expert Specialization in Mixture-of-Experts Language
Official implementation of DreamCraft3D
State-of-the-art (SoTA) text-to-video pre-trained model
ChatGPT interface with better UI
Stable Diffusion WebUI Forge is a platform on top of Stable Diffusion
VGGSfM: Visual Geometry Grounded Deep Structure From Motion
An AI-powered security review GitHub Action using Claude
Implementation of the Surya Foundation Model for Heliophysics
Inference code for scalable emulation of protein equilibrium ensembles
Programmatic access to the AlphaGenome model
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training
Chat & pretrained large vision language model
Tongyi Deep Research, the Leading Open-source Deep Research Agent
Qwen2.5-VL is the multimodal large language model series
Implementation of "MobileCLIP" CVPR 2024