Large-language-model & vision-language-model based on Linear Attention
Ling is a MoE LLM provided and open-sourced by InclusionAI
Revolutionizing Database Interactions with Private LLM Technology
A series of math-specific large language models of our Qwen2 series
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI
Qwen2.5-VL is the multimodal large language model series
Chat & pretrained large audio language model proposed by Alibaba Cloud
Chat & pretrained large vision language model
GLM-4.5V and GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning
Official inference repo for FLUX.2 models
Capable of understanding text, audio, vision, video
Diversity-driven optimization and large-model reasoning ability
CLIP, Predict the most relevant text snippet given an image
Qwen-Image is a powerful image generation foundation model
OCR expert VLM powered by Hunyuan's native multimodal architecture
Tool for exploring and debugging transformer model behaviors
GLM-4-Voice | End-to-End Chinese-English Conversational Model
Inference framework for 1-bit LLMs
GPT4V-level open-source multi-modal model based on Llama3-8B
Renderer for the harmony response format to be used with gpt-oss
Tongyi Deep Research, the Leading Open-source Deep Research Agent
A SOTA open-source image editing model
Generate Any 3D Scene in Seconds
An AI-powered security review GitHub Action using Claude
Chinese LLaMA-2 & Alpaca-2 Large Model Phase II Project