Qwen3.6 is the large language model series developed by Qwen team
Image generation model with single-stream diffusion transformer
MiniMax M2.1, a SOTA model for real-world dev & agents.
Tooling for the Common Objects In 3D dataset
FlashMLA: Efficient Multi-head Latent Attention Kernels
Qwen3-Coder is the code version of Qwen3
Extension index for stable-diffusion-webui
Uncommon Objects in 3D dataset
Scaling Reinforcement Learning with LLMs
Open-weight, large-scale hybrid-attention reasoning model
Qwen2.5-Coder is the code version of Qwen2.5, the large language model
Dataset of GPT-2 outputs for research in detection, biases, and more
Software that can generate photos from paintings
ChatGPT integration with Unity Editor
An implementation of model parallel GPT-2 and GPT-3-style models
Code for reproducing key results in the paper
Model that fuses instruct, reasoning and agentic skills
Lightweight 24B agentic coding model with vision and long context
Text-to-image model optimized for artistic quality and safe generation
OpenAI’s open-weight 120B model optimized for reasoning and tooling
Agentic 123B coding model optimized for large-scale engineering
Russian ASR model fine-tuned on Common Voice and CSS10 datasets
Multimodal agent model for coding, orchestration, and autonomy