Get started w/ building Fullstack Agents using Gemini 2.5 & LangGraph
Large-scale Self-supervised Pre-training Across Tasks, Languages, etc.
Ongoing research training transformer models at scale
Python bindings for llama.cpp
Open-source, high-performance AI model with advanced reasoning
Tongyi Deep Research, the Leading Open-source Deep Research Agent
Research code artifacts for Code World Model (CWM)
The official repo of Qwen chat & pretrained large language model
CogView4, CogView3-Plus and CogView3(ECCV 2024)
Diversity-driven optimization and large-model reasoning ability
GLM-4 series: Open Multilingual Multimodal Chat LMs
Repo of Qwen2-Audio chat & pretrained large audio language model
Set of tools to assess and improve LLM security
LLM training code for MosaicML foundation models
Infinite Craft but in Pyside6 and Python with local LLM
Open-source, high-performance Mixture-of-Experts large language model
Inference code for Llama models
A central, open resource for data and tools
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
AI R&D Efficiency Improvement Research: Do-It-Yourself Training LoRA
Implementation of model parallel autoregressive transformers on GPUs
Open-source pre-training implementation of Google's LaMDA in PyTorch