Structured outputs for llms
Python bindings for llama.cpp
Run Local LLMs on Any Device. Open-source
Low-code app builder for RAG and multi-agent AI applications
Agentic, Reasoning, and Coding (ARC) foundation models
Port of Facebook's LLaMA model in C/C++
GLM-4.5: Open-source LLM for intelligent agents by Z.ai
A high-throughput and memory-efficient inference and serving engine
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
Powerful AI language model (MoE) optimized for efficiency/performance
Interact with your documents using the power of GPT
Diversity-driven optimization and large-model reasoning ability
Open-source, high-performance AI model with advanced reasoning
PandasAI is a Python library that integrates generative AI
lightweight package to simplify LLM API calls
ChatGLM2-6B: An Open Bilingual Chat LLM
Inference code for CodeLlama models
CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
An AI personal assistant for your digital brain
Building applications with LLMs through composability
ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat
The official repo of Qwen chat & pretrained large language model
Access large language models from the command-line
Easy-to-use LLM fine-tuning framework (LLaMA-2, BLOOM, Falcon