Showing 159 open source projects for "apache server monitor"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 1
    Pezzo

    Pezzo

    Open-source, developer-first LLMOps platform

    Pezzo enables you to build, test, monitor and instantly ship AI all in one platform, while constantly optimizing for cost and performance. Packed with powerful features to streamline your workflow, so you can focus on what matters. Pezzo is a fully cloud-native and open-source LLMOps platform. Seamlessly observe and monitor your AI operations, troubleshoot issues, save up to 90% on costs and latency, collaborate and manage your prompts in one place, and instantly deliver AI changes.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    Opik

    Opik

    Debug, evaluate, and monitor your LLMapps, RAG systems, and agentic AI

    Confidently evaluate, test, and monitor LLM applications. Opik is an open-source platform for evaluating, testing, and monitoring LLM applications. Built by Comet. Record, sort, search, and understand each step your LLM app takes to generate a response. Manually annotate, view, and compare LLM responses in a user-friendly table. Log traces during development and in production. Run experiments with different prompts and evaluate against a test set. Choose and run pre-configured evaluation...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 3
    Argilla

    Argilla

    The open-source data curation platform for LLMs

    ...Most annotation tools treat data collection as a one-off activity at the beginning of each project. In real-world projects, data collection is a key activity of the iterative process of ML model development. Once a model goes into production, you want to monitor and analyze its predictions, and collect more data to improve your model over time. Argilla is designed to close this gap, enabling you to iterate as much as you need.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 4
    LoRAX

    LoRAX

    Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs

    Lorax is a multi-LoRA (Low-Rank Adaptation) inference server that scales to thousands of fine-tuned Large Language Models (LLMs). It enables efficient deployment and management of numerous fine-tuned models, facilitating scalable AI applications. Lorax is designed to handle high concurrency and provides a robust infrastructure for serving multiple LLMs simultaneously.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    NemoClaw

    NemoClaw

    NVIDIA plugin for secure installation of OpenClaw

    NVIDIA NemoClaw is an open-source tool designed to simplify the deployment and management of always-on AI assistants using the OpenClaw ecosystem. It installs and configures the NVIDIA OpenShell runtime, which provides a secure environment for running autonomous AI agents. NemoClaw enables users to launch sandboxed agent environments that control network access, file permissions, and inference requests through policy-based security. The platform integrates with AI models such as NVIDIA...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 6
    NVIDIA NeMo Agent Toolkit

    NVIDIA NeMo Agent Toolkit

    Library for efficiently connecting and optimizing teams of AI agents

    NVIDIA NeMo Agent Toolkit is an open-source framework designed to build, optimize, and manage AI agents across different development ecosystems. It provides enterprise-grade tools for improving agent performance, reliability, and observability throughout the development lifecycle. The toolkit integrates with popular agent frameworks such as LangChain, LlamaIndex, CrewAI, Microsoft Semantic Kernel, and Google ADK. Developers can monitor agent execution, trace workflows, and analyze...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 7
    tika-python

    tika-python

    Python binding to the Apache Tika™ REST services

    A Python port of the Apache Tika library that makes Tika available using the Tika REST Server. This makes Apache Tika available as a Python library, installable via Setuptools, Pip and easy to install. To use this library, you need to have Java 7+ installed on your system as tika-python starts up the Tika REST server in the background. To get this working in a disconnected environment, download a tika server file (both tika-server.jar and tika-server.jar.md5, which can be found here) and set the TIKA_SERVER_JAR environment variable to TIKA_SERVER_JAR="file:////tika-server.jar" which successfully tells python-tika to "download" this file and move it to /tmp/tika-server.jar and run as a background process. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    ChatMCP

    ChatMCP

    ChatMCP is an AI chat client implementing the Model Context Protocol

    ChatMCP is a cross‑platform AI chat client that implements the Model Context Protocol (MCP) to provide unified chat experiences across environments—including desktop, mobile, and web—with synchronization and protocol support tailored for MCP. You can install MCP Server from MCP Server Market, MCP Server Market is a collection of MCP Server, you can use it to chat with different data. Tested on major distributions: Ubuntu, Fedora, Arch Linux, openSUSE. Improved Experience: Latest versions...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 9
    kMCP

    kMCP

    Kubernetes Controller for building, testing and deploying MCP servers

    KMCP is a companion toolchain for building, testing, and deploying MCP servers with a workflow that spans local development through Kubernetes production deployments. It includes a CLI for day-to-day development tasks like scaffolding new MCP projects, managing tools, building container images, and running an MCP server locally for validation. For cluster operations, it includes a Kubernetes controller that manages MCP server lifecycles using a dedicated Custom Resource Definition (CRD),...
    Downloads: 3 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 11
    JADX-AI-MCP

    JADX-AI-MCP

    Plugin for JADX to integrate MCP server

    JADX-AI-MCP is an open-source plugin that integrates large language models into the JADX Android decompiler to assist with reverse engineering and code analysis tasks. The project connects JADX with AI assistants through the Model Context Protocol, enabling language models to interact directly with decompiled Android application code. Through this integration, AI systems can inspect classes, analyze methods, retrieve application manifests, and examine other elements of Android packages in...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 12
    Lemonade

    Lemonade

    Lemonade helps users run local LLMs with the highest performance

    Lemonade is a local LLM runtime that aims to deliver the highest possible performance on your own hardware by auto-configuring state-of-the-art inference engines for both NPUs and GPUs. The project positions itself as a “local LLM server” you can run on laptops and workstations, abstracting away backend differences while giving you a single place to serve and manage models. Its README emphasizes real-world adoption across startups, research groups, and large companies, signaling a focus on...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 13
    OpenJarvis

    OpenJarvis

    Personal AI, On Personal Devices

    OpenJarvis is an open-source framework designed to build personal AI agents that run primarily on local devices rather than relying on cloud infrastructure. Developed as part of the Intelligence Per Watt research initiative, it focuses on improving the efficiency and practicality of on-device AI systems. The framework provides shared primitives for building local-first agents, along with evaluation tools that measure performance using metrics such as energy consumption, latency, cost, and...
    Downloads: 185 This Week
    Last Update:
    See Project
  • 14
    gpt-oss

    gpt-oss

    gpt-oss-120b and gpt-oss-20b are two open-weight language models

    gpt-oss is OpenAI’s open-weight family of large language models designed for powerful reasoning, agentic workflows, and versatile developer use cases. The series includes two main models: gpt-oss-120b, a 117-billion parameter model optimized for general-purpose, high-reasoning tasks that can run on a single H100 GPU, and gpt-oss-20b, a lighter 21-billion parameter model ideal for low-latency or specialized applications on smaller hardware. Both models use a native MXFP4 quantization for...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 15
    ACI.dev

    ACI.dev

    Open platform connecting AI agents to tools via unified MCP server

    ACI is an open source platform designed to enable AI agents to interact with external tools through a unified and structured interface. It focuses on simplifying tool integration by connecting hundreds of pre-built services into agentic environments, allowing developers to avoid building custom API clients and authentication flows for each service. ACI provides intent-aware tool access, meaning agents can dynamically discover and use tools based on context rather than rigid configurations....
    Downloads: 6 This Week
    Last Update:
    See Project
  • 16
    Gateway MCP

    Gateway MCP

    Universal MCP-Server for your Databases optimized for LLMs

    Gateway is an MCP server that connects to structured databases like PostgreSQL, automatically analyzing schemas and data samples to generate optimized API structures. It leverages large language models (LLMs) during the discovery stage to produce API configurations, ensuring secure and efficient interactions between AI agents and databases. ​
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Chrome DevTools MCP

    Chrome DevTools MCP

    Chrome DevTools for coding agents

    chrome-devtools-mcp is an MCP server that connects AI agents to the Chrome DevTools Protocol so they can inspect pages, record traces, read console/network data, and modify the live browser state under user control. It makes a running Chrome instance visible to MCP clients, enabling agents to debug websites end-to-end—launching Chrome, navigating, profiling, and collecting artifacts in a structured way. The repository spells out environment requirements and cautions that exposing a live...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 18
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems. It provides an API interface that allows developers to...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    MCPHub

    MCPHub

    A unified hub for centralized management and dynamic organization

    MCPHub is a unified hub that organizes many MCP servers behind clean Streamable HTTP (SSE) endpoints so clients can connect to “all tools,” a specific server, or logical groups. It’s designed to simplify multi-server operations: one process can multiplex several named STDIO servers and re-expose them over HTTP for IDEs and services. The README ships in multiple languages and includes deployment and routing guidance, signaling an intention to reach a broad user base. The project publishes to...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    FastKoko

    FastKoko

    Dockerized FastAPI wrapper for Kokoro-82M text-to-speech model

    FastKoko is a self-hosted text-to-speech server built around the Kokoro-82M model and exposed through a FastAPI backend. It is designed to be easy to deploy via Docker, with separate CPU and GPU images so that users can choose between pure CPU inference and NVIDIA GPU acceleration. The project exposes an OpenAI-compatible speech endpoint, which means existing code that talks to the OpenAI audio API can often be pointed at a Kokoro-FastAPI instance with minimal changes. It supports multiple...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 21
    Tesseract.js

    Tesseract.js

    A pure Javascript Multilingual OCR

    Tesseract.js is a pure Javascript port of the popular Tesseract OCR engine. Tesseract.js' library supports more than 100 languages, automatic text orientation and script detection, a simple interface for reading paragraph, word, and character bounding boxes. Tesseract.js can run either in a browser and on a server with NodeJS. Tesseract.js is a javascript library that gets words in almost any spoken language out of images. The main Tesseract.js functions (ex. recognize, detect) take an image...
    Downloads: 28 This Week
    Last Update:
    See Project
  • 22
    ContextForge MCP Gateway

    ContextForge MCP Gateway

    A Model Context Protocol (MCP) Gateway & Registry

    MCP Context Forge is a feature-rich gateway and registry that federates Model Context Protocol (MCP) servers and traditional REST services behind a single, governed endpoint. It exposes an MCP-compliant interface to clients while handling discovery, authentication, rate limiting, retries, and observability on the server side. The gateway scales horizontally, supports multi-cluster deployments on Kubernetes, and uses Redis for federation and caching across instances. Operators can define...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 23
    Nexa SDK

    Nexa SDK

    Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML

    Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML models. It supports text generation, image generation, vision-language models (VLM), and speech-to-text (ASR), and text-to-speech (TTS) capabilities. Additionally, it offers an OpenAI-compatible API server with JSON schema mode for function calling and streaming support, and a user-friendly Streamlit UI. Users can run Nexa SDK in any device with Python environment, and GPU acceleration is supported, including CUDA, Metal, and...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 24
    LMCache

    LMCache

    Supercharge Your LLM with the Fastest KV Cache Layer

    LMCache is an extension layer for LLM serving engines that accelerates inference, especially with long contexts, by storing and reusing key-value (KV) attention caches across requests. Instead of rebuilding KV states for repeated or shared text segments, LMCache persists and retrieves them from multiple tiers—GPU memory, CPU DRAM, and local disk—then injects them into subsequent requests to reduce TTFT and increase throughput. Its design supports reuse beyond strict prefix matching and...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 25
    Polyaxon

    Polyaxon

    MLOps tools for managing & orchestrating the ML LifeCycle

    Polyaxon is an open-source machine learning operations (MLOps) platform built to help individuals, teams, and organizations develop, train, orchestrate, and monitor machine learning and deep learning workflows at scale with reproducibility and automation as core principles. It provides a unified solution for tracking experiments, managing datasets, scheduling jobs, and comparing results across runs, which greatly improves productivity and collaboration in data science teams. Polyaxon...
    Downloads: 2 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB