Make your agents learn from experience
A Simple and Universal Swarm Intelligence Engine
Power CLI and Workflow manager for LLMs (core package)
Alibaba's high-performance LLM inference engine for diverse apps
Query anything (GitHub, Notion, +40 more) with SQL and let LLMs
Universal LLM Deployment Engine with ML Compilation
AI-Powered Data Processing: Use LOTUS to process all of your datasets
Jlama is a modern LLM inference engine for Java
A Next-Generation Training Engine Built for Ultra-Large MoE Models
A high-throughput and memory-efficient inference and serving engine
Mooncake is the serving platform for Kimi
950 line, minimal, extensible LLM inference engine built from scratch
A lightweight vLLM implementation built from scratch
High-performance inference framework for large language models
SQL-Driven RAG Engine
A high-performance inference engine for AI models
A tension reasoning engine over 131 S-class problems
A modular graph-based Retrieval-Augmented Generation (RAG) system
local-first semantic code search engine
AI search engine - self-host with local or cloud LLMs
Fast Multimodal LLM on Mobile Devices
Fast, flexible LLM inference
Emscripten: An LLVM-to-WebAssembly Compiler
A @ClickHouse fork that supports high-performance vector search
Run AI models locally on your machine with node.js bindings for llama