GLM-4.5: Open-source LLM for intelligent agents by Z.ai
Audio foundation model excelling in audio understanding
Project Lyra: Open Generative 3D World Models
GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning
GLM-4.5V and GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning
gpt-oss-120b and gpt-oss-20b are two open-weight language models
Sharp Monocular Metric Depth in Less Than a Second
State-of-the-art Image & Video CLIP, Multimodal Large Language Models
GLM-4.6V/4.5V/4.1V-Thinking, towards versatile multimodal reasoning
Open-source framework for intelligent speech interaction
Block Diffusion for Ultra-Fast Speculative Decoding
Multimodal embedding and reranking models built on Qwen3-VL
Ling is a MoE LLM provided and open-sourced by InclusionAI
High-Fidelity and Controllable Generation of Textured 3D Assets
Multi-modal large language model designed for audio understanding
OCR expert VLM powered by Hunyuan's native multimodal architecture
High-Resolution Image Synthesis with Latent Diffusion Models
A Conversational Speech Generation Model
Example Discord bot written in Python that uses the completions API
PyTorch implementation of VALL-E (Zero-Shot Text-To-Speech)
A library for Multilingual Unsupervised or Supervised word Embeddings
LL model providing reasoning and conversational capabilities
Model that fuses instruct, reasoning and agentic skills
JetBrains’ 4B parameter code model for completions