Open-source, developer-first LLMOps platform
Debug, evaluate, and monitor your LLMapps, RAG systems, and agentic AI
The open-source data curation platform for LLMs
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
NVIDIA plugin for secure installation of OpenClaw
Library for efficiently connecting and optimizing teams of AI agents
Python binding to the Apache Tika™ REST services
ChatMCP is an AI chat client implementing the Model Context Protocol
Kubernetes Controller for building, testing and deploying MCP servers
Easiest and laziest way for building multi-agent LLMs applications
Plugin for JADX to integrate MCP server
Lemonade helps users run local LLMs with the highest performance
Personal AI, On Personal Devices
gpt-oss-120b and gpt-oss-20b are two open-weight language models
Open platform connecting AI agents to tools via unified MCP server
Universal MCP-Server for your Databases optimized for LLMs
Chrome DevTools for coding agents
High-performance inference server for text embeddings models API layer
A unified hub for centralized management and dynamic organization
Dockerized FastAPI wrapper for Kokoro-82M text-to-speech model
A pure Javascript Multilingual OCR
A Model Context Protocol (MCP) Gateway & Registry
Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML
Supercharge Your LLM with the Fastest KV Cache Layer
MLOps tools for managing & orchestrating the ML LifeCycle