GPT4V-level open-source multi-modal model based on Llama3-8B
Capable of understanding text, audio, vision, video
Agentic, Reasoning, and Coding (ARC) foundation models
Qwen3-omni is a natively end-to-end, omni-modal LLM
Open-source, high-performance AI model with advanced reasoning
Qwen2.5-VL is the multimodal large language model series
The official repo of Qwen chat & pretrained large language model
Qwen3-Coder is the code version of Qwen3
Qwen-Image is a powerful image generation foundation model
GLM-4.5V and GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning
⚡ Building applications with LLMs through composability ⚡
Chinese Llama-3 LLMs) developed from Meta Llama 3
CogView4, CogView3-Plus and CogView3(ECCV 2024)
Train a 26M-parameter GPT from scratch in just 2h
GLM-4-Voice | End-to-End Chinese-English Conversational Model
Open source libraries and APIs to build custom preprocessing pipelines
Code for Language models can explain neurons in language models paper
OpenDAN is an open source Personal AI OS
Advanced techniques for RAG systems
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
Guiding Instruction-based Image Editing via Multimodal Large Language
Data Lake for Deep Learning. Build, manage, and query datasets
Open-weight, large-scale hybrid-attention reasoning model
Framework that is dedicated to making neural data processing
Open-source, high-performance Mixture-of-Experts large language model