Showing 129 open source projects for "local host server"

View related business solutions
  • Auth0 B2B Essentials: SSO, MFA, and RBAC Built In Icon
    Auth0 B2B Essentials: SSO, MFA, and RBAC Built In

    Unlimited organizations, 3 enterprise SSO connections, role-based access control, and pro MFA included. Dev and prod tenants out of the box.

    Auth0's B2B Essentials plan gives you everything you need to ship secure multi-tenant apps. Unlimited orgs, enterprise SSO, RBAC, audit log streaming, and higher auth and API limits included. Add on M2M tokens, enterprise MFA, or additional SSO connections as you scale.
    Sign Up Free
  • Compliant and Reliable File Transfers Backed by Top Security Certifications Icon
    Compliant and Reliable File Transfers Backed by Top Security Certifications

    Cerberus FTP Server delivers SOC 2 Type II certified security and FIPS 140-2 validated encryption.

    Stop relying on non-certified, legacy file transfer tools that creak under the weight of modern security demands. Get full audit trails, advanced access controls and more supported by an award-winning team of experts. Start your free 25-day trial today.
    Start Free Trial
  • 1
    OpenVINO Model Server

    OpenVINO Model Server

    A scalable inference server for models optimized with OpenVINO

    OpenVINO™ Model Server is a high-performance inference serving system designed to host and serve machine learning models that have been optimized with the OpenVINO toolkit. It’s implemented in C++ for scalability and efficiency, making it suitable for both edge and cloud deployments where inference workloads must be reliable and high throughput. The server exposes model inference via standard network protocols like REST and gRPC, allowing any client that speaks those protocols to request predictions remotely, abstracting away the complexity of where and how the model runs. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    Last9 MCP Server

    Last9 MCP Server

    Last9 MCP Server

    The Last9 MCP Server is a Model Context Protocol server implementation for Last9, enabling AI agents to seamlessly bring real-time production context—logs, metrics, and traces—into local environments to auto-fix code faster. ​
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    csghub-server

    csghub-server

    csghub-server is the backend server for CSGHub

    csghub-server is the backend component of the CSGHub platform, an open-source infrastructure designed to manage and operate large language models, datasets, and AI development workflows within a private deployment environment. The server acts as a centralized management layer that allows teams to store, organize, and operate AI assets such as models, datasets, and machine learning applications in a manner similar to artifact repositories used in software engineering.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    WhatsApp MCP Server

    WhatsApp MCP Server

    WhatsApp MCP server enabling AI access to chats and messaging

    whatsapp-mcp is an open source Model Context Protocol (MCP) server that enables AI agents to interact directly with a user’s WhatsApp account through a structured interface. It acts as a bridge between WhatsApp and large language models, allowing controlled access to messages, chats, and contacts. whatsapp-mcp is composed of two main components: a Go-based bridge that connects to the WhatsApp Web API and stores data locally, and a Python-based MCP server that exposes tools for AI interaction. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • 5
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 6
    OpenJarvis

    OpenJarvis

    Personal AI, On Personal Devices

    OpenJarvis is an open-source framework designed to build personal AI agents that run primarily on local devices rather than relying on cloud infrastructure. Developed as part of the Intelligence Per Watt research initiative, it focuses on improving the efficiency and practicality of on-device AI systems. The framework provides shared primitives for building local-first agents, along with evaluation tools that measure performance using metrics such as energy consumption, latency, cost, and...
    Downloads: 273 This Week
    Last Update:
    See Project
  • 7
    Lemonade

    Lemonade

    Lemonade helps users run local LLMs with the highest performance

    Lemonade is a local LLM runtime that aims to deliver the highest possible performance on your own hardware by auto-configuring state-of-the-art inference engines for both NPUs and GPUs. The project positions itself as a “local LLM server” you can run on laptops and workstations, abstracting away backend differences while giving you a single place to serve and manage models.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 8
    MiniMax-MCP

    MiniMax-MCP

    Official MiniMax Model Context Protocol (MCP) server

    ...The server is written in Python and distributed under the MIT license, with a pyproject.toml and uv-based workflow that makes installation and execution reproducible. Configuration is handled through JSON files that tell MCP clients how to launch the server (typically via uvx minimax-mcp) and which environment variables to use for the API key, host, and output directory.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    Open Responses

    Open Responses

    Specification for multi-provider, interoperable LLM interfaces

    Open Responses is an open-source implementation of an API compatible with the OpenAI Responses API that lets developers self-host a drop-in alternative endpoint for AI interactions while preserving compatibility with existing Agents SDKs and model workflows. It enables you to run a local or private server that speaks the standard Responses API, so tools, applications, and agents built against that API can operate without contacting OpenAI’s cloud and can instead route calls to any large language model provider you choose, such as Claude, Qwen, Ollama, or others. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Build Securely on Azure with Proven Frameworks Icon
    Build Securely on Azure with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 10
    Farfalle

    Farfalle

    AI search engine - self-host with local or cloud LLMs

    ...The project integrates large language models with multiple search APIs so that the system can gather information from external sources and synthesize responses into concise answers. It can run either with local language models or with cloud-based providers, allowing developers to deploy it privately or integrate with hosted AI services. The architecture separates the frontend and backend, using modern web technologies such as Next.js and FastAPI to deliver an interactive interface and scalable server logic. Farfalle also includes an agent-based search workflow that plans queries and executes multiple search steps to produce more accurate results than traditional keyword searches. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Perplexica

    Perplexica

    Perplexica is an AI-powered answering engine.

    Perplexica is a privacy-focused AI answering engine like Perplexity that you can self-host on your own hardware for private, source-cited web research. It combines live internet search results with AI models, letting you use local LLMs via Ollama or connect to providers like OpenAI, Claude, Gemini, and Groq. Powered by SearxNG, it aggregates results from multiple search engines while keeping your identity and queries private.
    Downloads: 15 This Week
    Last Update:
    See Project
  • 12
    ds4.c

    ds4.c

    DeepSeek 4 Flash local inference engine for Metal

    ds4.c is a specialized local inference engine created by antirez for running DeepSeek V4 Flash models directly on Apple Silicon hardware using Metal acceleration. Unlike general-purpose inference runtimes, the project is intentionally optimized for a specific model family, enabling highly efficient execution and simplified architecture. The engine includes DS4-specific model loading, KV cache management, prompt rendering, and OpenAI-compatible server APIs for local deployment workflows. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    AingDesk is an open-source desktop and server-based AI assistant platform designed to provide a user-friendly environment for interacting with language models and building AI-powered tools. The software enables users to run local AI models or connect to external model APIs through a unified interface. One of its primary goals is to simplify the process of building knowledge-based assistants by allowing users to create local knowledge bases that the AI can search and analyze. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Lemon AI

    Lemon AI

    Full-stack Open-source Self-Evolving General AI Agent

    ...A key component of the framework is a virtual machine sandbox environment that safely executes code generated by the agent without affecting the host system.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Ollamac

    Ollamac

    Mac app for Ollama

    Ollamac is an open-source native macOS application that provides a graphical interface for interacting with local large language models running through the Ollama inference framework. The project was created to simplify the process of using local AI models, which typically require command-line interaction, by offering a clean and intuitive desktop interface. Through this interface, users can run and chat with a variety of LLM models installed through Ollama directly on their own machines....
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    web-eval-agent MCP Server

    web-eval-agent MCP Server

    An MCP server that autonomously evaluates web applications

    ...Marketing and README material emphasize supercharging local debugging loops by combining live browser execution with LLM-driven hypotheses and fixes. Activity on the repo shows steady iteration, with issues and PRs centered on reliability and developer experience. In short, it wraps autonomous, in-editor web testing and diagnosis behind a predictable MCP interface.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture intact. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    kMCP

    kMCP

    Kubernetes Controller for building, testing and deploying MCP servers

    KMCP is a companion toolchain for building, testing, and deploying MCP servers with a workflow that spans local development through Kubernetes production deployments. It includes a CLI for day-to-day development tasks like scaffolding new MCP projects, managing tools, building container images, and running an MCP server locally for validation. For cluster operations, it includes a Kubernetes controller that manages MCP server lifecycles using a dedicated Custom Resource Definition (CRD), allowing MCP servers to be represented as native Kubernetes objects you can operate with familiar kubectl-driven patterns. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    Jan.ai

    Jan.ai

    Open source alternative to ChatGPT that runs 100% offline

    Jan.ai is an open-source, privacy-focused AI assistant that serves as an alternative to ChatGPT, running completely locally on your device. It allows you to download and run LLMs (local language models) offline while also offering optional integration with cloud-based model providers—giving you full control over your data and AI interactions. Download and run LLMs (Llama, Gemma, Qwen, GPT-oss etc.) from HuggingFace. Connect to GPT models via OpenAI, Claude models via Anthropic, Mistral,...
    Downloads: 58 This Week
    Last Update:
    See Project
  • 20
    Shadcn UI v4 MCP Server

    Shadcn UI v4 MCP Server

    A mcp server to allow LLMS gain context about shadcn ui component

    ...The server supports multiple frontend frameworks including React, Svelte, Vue, and React Native, making it highly versatile for cross-platform development. It includes smart caching and efficient GitHub API usage to optimize performance and handle rate limits during component retrieval. The system also supports multiple transport modes such as standard input/output and Server-Sent Events, enabling both local and distributed deployments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    mcpo

    mcpo

    A simple, secure MCP-to-OpenAPI proxy server

    mcpo is a minimal bridge that exposes any MCP tool as an OpenAPI-compatible HTTP server. Instead of writing glue code, you point mcpo at an MCP server command and it generates REST endpoints and an OpenAPI spec that other systems (or LLM agent frameworks) can call immediately. This design lets you reuse a growing library of MCP servers with platforms that only understand HTTP+OpenAPI, unifying tool access across ecosystems.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    webclaw

    webclaw

    Fast, local-first web content extraction for LLMs

    ...The tool addresses a major inefficiency in AI workflows by removing irrelevant elements like navigation menus, ads, and scripts, significantly reducing token usage when feeding data into language models. It supports multiple modes of operation, including CLI usage, REST API access, and an MCP server for direct integration with agent-based systems. Webclaw also provides advanced capabilities such as recursive crawling, structured JSON extraction, summarization, and content comparison, making it suitable for research and data pipelines. Its local-first architecture ensures privacy and eliminates the need for API keys.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    Dive

    Dive

    Dive is an open-source MCP Host Desktop Application

    ...One-click access to managed MCP servers via OAPHub.ai - eliminates complex local deployments. Modern Tauri version alongside traditional Electron version for optimal performance.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    Groq Desktop

    Groq Desktop

    Local Groq Desktop chat app with MCP support

    Groq Desktop is a cross-platform (Windows / macOS / Linux) local desktop application that provides a graphical chat interface for interacting with Groq-hosted, function-call-capable models. It bundles a built-in MCP (Model Context Protocol) server enabling smart function calling, letting users chat with an AI, send images, or interact with richer multimodal inputs — all from a friendly desktop UI.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 25
    Chatbot UI

    Chatbot UI

    AI chat for any model

    Chatbot UI is an open-source, full-featured chat interface for AI models that aims to lower the barrier for anyone wanting to run their own AI-powered chatbot — whether using remote LLM APIs or local/self-hosted models. It bundles a modern frontend (Next.js + TypeScript + Tailwind) with a backend (Supabase by default) for persistent storage of chats, history, and user settings, replacing earlier local-storage-only approaches. The project supports multiple models/providers (e.g. OpenAI, local models via Ollama), and by switching env-vars you can self-host the whole stack or deploy in the cloud (e.g. via Vercel + Supabase) for personal or shared use. ...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB