Port of Facebook's LLaMA model in C/C++
Python bindings for llama.cpp
Run Local LLMs on Any Device. Open-source
Interface for OuteTTS models
Your Personal AI Assistant; easy to install, deploy on local or coud
Inference Llama 2 in one file of pure C
An easy-to-understand framework for LLM samplers
Qwen3 is the large language model series developed by Qwen team
Oobabooga - The definitive Web UI for local AI, with powerful features
Performance-optimized AI inference on your GPUs
Towards Human-Sounding Speech
Run GGUF models easily with a UI or API. One File. Zero Install.
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
Chinese LLaMA & Alpaca large language model + local CPU/GPU training
JetBrains’ 4B parameter code model for completions