Make your agents learn from experience
Power CLI and Workflow manager for LLMs (core package)
A Simple and Universal Swarm Intelligence Engine
Alibaba's high-performance LLM inference engine for diverse apps
Universal LLM Deployment Engine with ML Compilation
A high-performance inference engine for AI models
A Next-Generation Training Engine Built for Ultra-Large MoE Models
AI-Powered Data Processing: Use LOTUS to process all of your datasets
SQL-Driven RAG Engine
A tension reasoning engine over 131 S-class problems
A lightweight vLLM implementation built from scratch
Mooncake is the serving platform for Kimi
AI search engine - self-host with local or cloud LLMs
Fast Multimodal LLM on Mobile Devices
High-performance inference framework for large language models
A @ClickHouse fork that supports high-performance vector search
Claude + Obsidian knowledge companion
WebAssembly binding for llama.cpp - Enabling on-browser LLM inference
Masks sensitive data and secrets before they reach AI
Fully private LLM chatbot that runs entirely with a browser
Request recommended movies, TV shows and anime to Jellyseer/Overseer
Build multimodal language agents for fast prototype and production
Based on the LangChain/LangGraph framework
Parallax is a distributed model serving framework
Run a 1-billion parameter LLM on a $10 board with 256MB RAM