Showing 47 open source projects for "local host server"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • Add Two Lines of Code. Get Full APM. Icon
    Add Two Lines of Code. Get Full APM.

    AppSignal installs in minutes and auto-configures dashboards, alerts, and error tracking.

    Works out of the box for Rails, Django, Express, Phoenix, and more. Monitoring exceptions and performance in no time.
    Start Free
  • 1
    mcp-server-chatsum

    mcp-server-chatsum

    Query and Summarize your chat messages

    mcp-server-chatsum is an MCP server that indexes your chat history and provides tools to query and produce focused summaries on demand. It offers a simple flow: point the server at a local chat database, run the companion chatbot to ingest messages, and then use the MCP tool to retrieve scoped threads and generate concise syntheses. The tool design lets agents filter by participants, time ranges, or keywords before summarizing, which keeps outputs relevant and reduces hallucinated context. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    MCP Server Chart

    MCP Server Chart

    A visualization mcp contains 25+ visual charts

    ...The README documents private deployment, record generation, and tool filtering, giving teams a path from local experimentation to managed usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Open Responses

    Open Responses

    Specification for multi-provider, interoperable LLM interfaces

    Open Responses is an open-source implementation of an API compatible with the OpenAI Responses API that lets developers self-host a drop-in alternative endpoint for AI interactions while preserving compatibility with existing Agents SDKs and model workflows. It enables you to run a local or private server that speaks the standard Responses API, so tools, applications, and agents built against that API can operate without contacting OpenAI’s cloud and can instead route calls to any large language model provider you choose, such as Claude, Qwen, Ollama, or others. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Farfalle

    Farfalle

    AI search engine - self-host with local or cloud LLMs

    ...The project integrates large language models with multiple search APIs so that the system can gather information from external sources and synthesize responses into concise answers. It can run either with local language models or with cloud-based providers, allowing developers to deploy it privately or integrate with hosted AI services. The architecture separates the frontend and backend, using modern web technologies such as Next.js and FastAPI to deliver an interactive interface and scalable server logic. Farfalle also includes an agent-based search workflow that plans queries and executes multiple search steps to produce more accurate results than traditional keyword searches. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    Perplexica

    Perplexica

    Perplexica is an AI-powered answering engine.

    Perplexica is a privacy-focused AI answering engine like Perplexity that you can self-host on your own hardware for private, source-cited web research. It combines live internet search results with AI models, letting you use local LLMs via Ollama or connect to providers like OpenAI, Claude, Gemini, and Groq. Powered by SearxNG, it aggregates results from multiple search engines while keeping your identity and queries private.
    Downloads: 15 This Week
    Last Update:
    See Project
  • 6
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    AingDesk is an open-source desktop and server-based AI assistant platform designed to provide a user-friendly environment for interacting with language models and building AI-powered tools. The software enables users to run local AI models or connect to external model APIs through a unified interface. One of its primary goals is to simplify the process of building knowledge-based assistants by allowing users to create local knowledge bases that the AI can search and analyze. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    MCPB

    MCPB

    One-click local MCP server installation in desktop apps

    MCPB (MCP Bundles) defines a packaging format and toolchain for one-click installation of local Model Context Protocol (MCP) servers in desktop apps like Claude for macOS and Windows. An .mcpb file is a zip archive containing your server and a manifest.json that declares capabilities, entry points, permissions, and configuration inputs, much like how .crx packages Chrome extensions or .vsix packages VS Code extensions. The goal is to make local tool servers easy for end users to install, update, and configure, while giving app developers a consistent way to discover and load them safely. ...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 8
    Jan.ai

    Jan.ai

    Open source alternative to ChatGPT that runs 100% offline

    Jan.ai is an open-source, privacy-focused AI assistant that serves as an alternative to ChatGPT, running completely locally on your device. It allows you to download and run LLMs (local language models) offline while also offering optional integration with cloud-based model providers—giving you full control over your data and AI interactions. Download and run LLMs (Llama, Gemma, Qwen, GPT-oss etc.) from HuggingFace. Connect to GPT models via OpenAI, Claude models via Anthropic, Mistral,...
    Downloads: 58 This Week
    Last Update:
    See Project
  • 9
    Shadcn UI v4 MCP Server

    Shadcn UI v4 MCP Server

    A mcp server to allow LLMS gain context about shadcn ui component

    ...The server supports multiple frontend frameworks including React, Svelte, Vue, and React Native, making it highly versatile for cross-platform development. It includes smart caching and efficient GitHub API usage to optimize performance and handle rate limits during component retrieval. The system also supports multiple transport modes such as standard input/output and Server-Sent Events, enabling both local and distributed deployments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    Dive

    Dive

    Dive is an open-source MCP Host Desktop Application

    ...One-click access to managed MCP servers via OAPHub.ai - eliminates complex local deployments. Modern Tauri version alongside traditional Electron version for optimal performance.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    Chatbot UI

    Chatbot UI

    AI chat for any model

    Chatbot UI is an open-source, full-featured chat interface for AI models that aims to lower the barrier for anyone wanting to run their own AI-powered chatbot — whether using remote LLM APIs or local/self-hosted models. It bundles a modern frontend (Next.js + TypeScript + Tailwind) with a backend (Supabase by default) for persistent storage of chats, history, and user settings, replacing earlier local-storage-only approaches. The project supports multiple models/providers (e.g. OpenAI, local models via Ollama), and by switching env-vars you can self-host the whole stack or deploy in the cloud (e.g. via Vercel + Supabase) for personal or shared use. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    OpenClaw Studio

    OpenClaw Studio

    A clean web dashboard for OpenClaw

    ...The platform provides real-time chat capabilities, approval management, and job configuration tools for agent operations. Built with a control-plane architecture, it handles communication between the browser and the gateway through server-managed connections. OpenClaw Studio supports both local and cloud deployments, making it flexible for different development and production environments. It is ideal for developers and teams looking to streamline agent management and improve visibility into autonomous AI workflows.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 13
    5ire

    5ire

    5ire is a cross-platform desktop AI assistant, MCP client

    5ire is a sleek, cross‑platform desktop AI assistant and MCP client that connects to major service providers, supports a local knowledge base and tool integration via MCP servers, enabling robust RAG and assistant features. These components are required as they constitute the runtime environment for the MCP Server. If you don't anticipate using the tools feature immediately, you may choose to skip this installation step and complete it later when the need arises.
    Downloads: 16 This Week
    Last Update:
    See Project
  • 14
    XcodeBuildMCP

    XcodeBuildMCP

    A Model Context Protocol (MCP) server

    ...The server aims to be “agent-ready,” surfacing capabilities (build, clean, test, archive, install, run, log collection) with explicit schemas instead of brittle prompt instructions. It supports MCP transports suitable for local IDEs and service deployments, and pairs with a public website that positions it as a bridge between Xcode and autonomous assistants.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    pyTelegramBotAPI

    pyTelegramBotAPI

    Python Telegram bot api.

    TeleBot is the synchronous and asynchronous implementation of Telegram Bot API.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 16
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    OctoMind MCP

    OctoMind MCP

    An MCP server for octomind tools, resources and prompts

    The Octomind MCP Server is designed to integrate Octomind's end-to-end testing tools and resources into local development environments. It enables AI-powered interfaces to create, execute, and manage e2e tests, enhancing the testing workflow. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    DBHub

    DBHub

    Universal database MCP server connecting to MySQL, PostgreSQL

    DBHub is a universal database gateway that implements the MCP server interface so assistants and IDEs can explore and query databases through typed tools. It supports multiple transports—stdio for desktop clients and HTTP for networked scenarios—making it flexible to embed or deploy. Configuration is environment-variable driven, with a DSN and per-engine settings covering Postgres, MySQL, MariaDB, SQL Server, and SQLite. Operational flags include read-only mode, row limits, and even SSH...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Modelence

    Modelence

    Modelence is an all-in-one TypeScript platform

    ...The project is built to support modern application workflows where product teams want to move quickly without stitching together many separate services and libraries. It includes scaffolding and tooling to create a new application quickly, then run a local development server with a predictable structure that’s easy to extend. Modelence also focuses on “standard features” that most apps require, so developers can spend more time on product logic rather than setup and glue code.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 20
    Desktop Commander MCP

    Desktop Commander MCP

    AI-powered MCP server for desktop file and terminal automation

    Desktop Commander MCP is an advanced Model Context Protocol server designed to extend AI assistants with direct control over a user’s local machine, including the file system and terminal. It integrates with clients like Claude Desktop to enable AI-driven workflows such as editing files, executing commands, and automating development tasks from a single conversational interface. Desktop Commander MCP builds on top of an MCP filesystem server and enhances it with powerful search, replace, and code editing capabilities tailored for real-world development environments. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    ComfyUI Desktop

    ComfyUI Desktop

    The desktop app for ComfyUI

    ComfyUI Desktop is a packaged desktop application that simplifies the installation and usage of ComfyUI, a node-based interface for generative AI workflows such as image generation using diffusion models. It is designed to eliminate the complexity of manual setup by bundling all required dependencies, including Python environments, frontend components, and management tools, into a single installable application. Upon launch, the application automatically installs dependencies, initializes...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 22
    Gemma 4 Browser Assistant

    Gemma 4 Browser Assistant

    On-device AI agent Chrome extension powered by Transformers.js

    Gemma 4 Browser Assistant is an open-source browser extension that embeds an AI assistant directly into the browsing experience, powered by on-device machine learning models. It uses Transformers.js and Gemma models to run inference locally in the browser, eliminating the need for external servers and preserving user privacy. The extension includes a side panel interface that allows users to interact with the AI while browsing, enabling tasks such as summarizing pages and answering...
    Downloads: 57 This Week
    Last Update:
    See Project
  • 23
    QMD

    QMD

    mini cli search engine for your docs, knowledge bases, etc.

    QMD is a powerful and lightweight command-line tool that acts as an on-device search engine for your personal knowledge base, allowing you to index and search files like Markdown notes, meeting transcripts, technical documentation, and other text collections without depending on cloud services. Designed to keep all search activity local, it combines classic full-text search techniques with modern semantic features such as vector similarity and hybrid ranking so that queries return not just...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    PasteGuard

    PasteGuard

    Masks sensitive data and secrets before they reach AI

    PasteGuard is an open-source privacy proxy that protects sensitive information like personal data and API secrets by detecting and masking them before they reach large language model APIs such as OpenAI or Anthropic Claude. It sits between an application and the LLM provider, automatically replacing names, emails, tokens, and other personally identifiable information (PII) with placeholders so that external services never see raw sensitive values, and then optionally unmasking them in the...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    Laminar

    Laminar

    Open-source all-in-one platform for engineering AI products

    Laminar is an open source all-in-one platform for engineering best-in-class LLM products. Data governs the quality of your LLM application. Laminar helps you collect it, understand it, and use it. When you trace your LLM application, you get a clear picture of every step of execution and simultaneously collect invaluable data. You can use it to set up better evaluations, as dynamic few-shot examples, and for fine-tuning. All traces are sent in the background via gRPC with minimal overhead....
    Downloads: 3 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB