Power CLI and Workflow manager for LLMs (core package)
Make your agents learn from experience
A Simple and Universal Swarm Intelligence Engine
Universal LLM Deployment Engine with ML Compilation
A high-performance inference engine for AI models
Jlama is a modern LLM inference engine for Java
A high-throughput and memory-efficient inference and serving engine
Alibaba's high-performance LLM inference engine for diverse apps
SQL-Driven RAG Engine
A tension reasoning engine over 131 S-class problems
A Next-Generation Training Engine Built for Ultra-Large MoE Models
AI-Powered Data Processing: Use LOTUS to process all of your datasets
A lightweight vLLM implementation built from scratch
950 line, minimal, extensible LLM inference engine built from scratch
AI search engine - self-host with local or cloud LLMs
A modular graph-based Retrieval-Augmented Generation (RAG) system
Fast Multimodal LLM on Mobile Devices
Query anything (GitHub, Notion, +40 more) with SQL and let LLMs
Mooncake is the serving platform for Kimi
Emscripten: An LLVM-to-WebAssembly Compiler
A @ClickHouse fork that supports high-performance vector search
local-first semantic code search engine
High-performance inference framework for large language models
Fast, flexible LLM inference
Claude + Obsidian knowledge companion