Python bindings for llama.cpp
Run Local LLMs on Any Device. Open-source
Interface for OuteTTS models
Personal AI, On Personal Devices
Your Personal AI Assistant; easy to install, deploy on local or coud
Inference Llama 2 in one file of pure C
Run a full local LLM stack with one command using Docker
An easy-to-understand framework for LLM samplers
Qwen3 is the large language model series developed by Qwen team
Oobabooga - The definitive Web UI for local AI, with powerful features
Performance-optimized AI inference on your GPUs
Towards Human-Sounding Speech
GLM-4 series: Open Multilingual Multimodal Chat LMs
Powerful Android AI agent with tools, automation, and Linux shell
Run GGUF models easily with a UI or API. One File. Zero Install.
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
JetBrains’ 4B parameter code model for completions