Showing 314 open source projects for "web server local"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 1
    Last9 MCP Server

    Last9 MCP Server

    Last9 MCP Server

    The Last9 MCP Server is a Model Context Protocol server implementation for Last9, enabling AI agents to seamlessly bring real-time production context—logs, metrics, and traces—into local environments to auto-fix code faster. ​
    Downloads: 10 This Week
    Last Update:
    See Project
  • 2
    web-eval-agent MCP Server

    web-eval-agent MCP Server

    An MCP server that autonomously evaluates web applications

    web-eval-agent is a Model Context Protocol (MCP) server that spins up a browser-use–capable debugging agent to autonomously run and evaluate web apps straight from your editor. It’s positioned as a “let the coding agent debug itself” companion: the agent launches the app, navigates flows, captures evidence, and iterates on failures without manual copy-pasting of logs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    The Web MCP

    The Web MCP

    A powerful Model Context Protocol (MCP) server

    Bright Data’s Web MCP server gives AI assistants robust, real-time web capabilities through an MCP interface designed to avoid blocks, rate limits, and CAPTCHAs. It presents search, crawl, navigate, and extraction tools that agents can call directly, replacing brittle scraping prompts with typed operations. The README markets it as a “gateway” to the live web so assistants don’t fall back to stale training data.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 4
    WhatsApp MCP Server

    WhatsApp MCP Server

    WhatsApp MCP server enabling AI access to chats and messaging

    whatsapp-mcp is an open source Model Context Protocol (MCP) server that enables AI agents to interact directly with a user’s WhatsApp account through a structured interface. It acts as a bridge between WhatsApp and large language models, allowing controlled access to messages, chats, and contacts. whatsapp-mcp is composed of two main components: a Go-based bridge that connects to the WhatsApp Web API and stores data locally, and a Python-based MCP server that exposes tools for AI interaction. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    MCP Server RAG Web Browser

    MCP Server RAG Web Browser

    A MCP Server for the RAG Web Browser Actor

    The MCP Server for the RAG Web Browser Actor allows AI assistants and LLMs to perform web searches and extract information from web pages. It facilitates interaction with the web, enabling up-to-date context retrieval for AI applications. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    MCP Server Playwright

    MCP Server Playwright

    MCP server for browser automation using Playwright

    An MCP (Model Context Protocol) server that leverages Playwright to provide browser automation capabilities, enabling large language models (LLMs) to interact with web pages, take screenshots, and execute JavaScript within a real browser environment. ​
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Tiledesk Server

    Tiledesk Server

    Tiledesk Server is the main API component of the Tiledesk platform

    Tiledesk Server is the backend component of the Tiledesk platform, providing a comprehensive open-source live chat system with integrated chatbot capabilities for customer support and engagement.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 9
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    AingDesk is an open-source desktop and server-based AI assistant platform designed to provide a user-friendly environment for interacting with language models and building AI-powered tools. The software enables users to run local AI models or connect to external model APIs through a unified interface. One of its primary goals is to simplify the process of building knowledge-based assistants by allowing users to create local knowledge bases that the AI can search and analyze. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 10
    OpenVINO Model Server

    OpenVINO Model Server

    A scalable inference server for models optimized with OpenVINO

    OpenVINO™ Model Server is a high-performance inference serving system designed to host and serve machine learning models that have been optimized with the OpenVINO toolkit. It’s implemented in C++ for scalability and efficiency, making it suitable for both edge and cloud deployments where inference workloads must be reliable and high throughput. The server exposes model inference via standard network protocols like REST and gRPC, allowing any client that speaks those protocols to request predictions remotely, abstracting away the complexity of where and how the model runs. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 11
    Markdownify MCP Server

    Markdownify MCP Server

    Convert files and web content into clean, usable Markdown easily

    Markdownify MCP is a Model Context Protocol server that converts many types of files and web content into clean Markdown. It supports formats such as PDFs, images, audio with transcription, DOCX, XLSX, and PPTX, along with web sources like YouTube transcripts, Bing results, and general webpages. Markdownify MCP is designed to simplify content extraction and make data easier to read, share, and reuse in structured workflows.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 12
    Actors MCP Server

    Actors MCP Server

    Model Context Protocol (MCP) Server for Apify's Actors

    The Apify Actors MCP Server is a Model Context Protocol (MCP) server that enables AI assistants to interact with Apify Actors. This integration allows AI models to utilize various web scraping and automation tools provided by Apify, facilitating tasks such as data extraction and web automation. ​
    Downloads: 9 This Week
    Last Update:
    See Project
  • 13
    Text Generation Web UI

    Text Generation Web UI

    Oobabooga - The definitive Web UI for local AI, with powerful features

    A gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, Pythia, OPT, and GALACTICA. Dropdown menu for switching between models. Notebook mode that resembles OpenAI's playground. Chat mode for conversation and role playing. Instruct mode compatible with Alpaca and Open Assistant formats. Nice HTML output for GPT-4chan. Markdown output for GALACTICA, including LaTeX rendering. Custom chat characters. Advanced chat features (send images, get audio responses with TTS)....
    Downloads: 61 This Week
    Last Update:
    See Project
  • 14
    Firecrawl MCP Server

    Firecrawl MCP Server

    Adds powerful web scraping and search to Cursor and Claude

    firecrawl-mcp-server is the official MCP integration for Firecrawl that brings high-recall web scraping, crawling, and search into IDEs and agent runtimes. It exposes tools for single-page scrape, multi-URL batch jobs, site discovery, and search enrichment, returning cleaned, structured content suitable for downstream LLM reasoning. The server is designed to run with Firecrawl’s hosted API or self-hosted deployments, making it flexible for enterprise data-governance requirements. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 15
    RAG Web UI

    RAG Web UI

    RAG Web UI is an intelligent dialogue system based on RAG

    RAG Web UI is an open-source intelligent dialogue system built on retrieval-augmented generation technology, designed to enable users to create AI-powered question answering systems grounded in their own knowledge bases. It combines document retrieval with large language models to provide accurate, context-aware responses based on indexed data rather than generic model knowledge. The platform supports ingestion of multiple document formats, including PDFs, Word files, Markdown, and plain...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Browser Use MCP Server

    Browser Use MCP Server

    Browse the web, directly from Cursor etc.

    A browser automation server implementing the Model Context Protocol, designed to allow AI assistants to browse the web directly from applications like Cursor. It supports natural language commands for web navigation and interaction. ​
    Downloads: 5 This Week
    Last Update:
    See Project
  • 17
    Lemonade

    Lemonade

    Lemonade helps users run local LLMs with the highest performance

    Lemonade is a local LLM runtime that aims to deliver the highest possible performance on your own hardware by auto-configuring state-of-the-art inference engines for both NPUs and GPUs. The project positions itself as a “local LLM server” you can run on laptops and workstations, abstracting away backend differences while giving you a single place to serve and manage models.
    Downloads: 15 This Week
    Last Update:
    See Project
  • 18
    Browserbase MCP Server

    Browserbase MCP Server

    Allow LLMs to control a browser with Browserbase and Stagehand

    Browserbase MCP Server is a server implementation of the Model Context Protocol (MCP) that enables large language models to interact with web browsers programmatically through cloud-based automation. The project provides a standardized interface for connecting AI systems to real-world web environments, allowing them to navigate pages, extract structured data, and perform user-like actions such as clicking, typing, and form submission.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    OpenJarvis

    OpenJarvis

    Personal AI, On Personal Devices

    OpenJarvis is an open-source framework designed to build personal AI agents that run primarily on local devices rather than relying on cloud infrastructure. Developed as part of the Intelligence Per Watt research initiative, it focuses on improving the efficiency and practicality of on-device AI systems. The framework provides shared primitives for building local-first agents, along with evaluation tools that measure performance using metrics such as energy consumption, latency, cost, and...
    Downloads: 95 This Week
    Last Update:
    See Project
  • 20
    OpenClaw Studio

    OpenClaw Studio

    A clean web dashboard for OpenClaw

    ...Built with a control-plane architecture, it handles communication between the browser and the gateway through server-managed connections. OpenClaw Studio supports both local and cloud deployments, making it flexible for different development and production environments. It is ideal for developers and teams looking to streamline agent management and improve visibility into autonomous AI workflows.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    Deep Research

    Deep Research

    Use any LLMs (Large Language Models) for Deep Research

    Deep Research is a local-first research agent that orchestrates multiple LLMs to generate in-depth reports in minutes. It combines “thinking” and “task” model roles with live internet access to plan, search, read, and synthesize findings into structured outputs. The project emphasizes privacy: processing and storage happen locally, avoiding server-side retention of your queries and notes.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 22
    Groq Desktop

    Groq Desktop

    Local Groq Desktop chat app with MCP support

    Groq Desktop is a cross-platform (Windows / macOS / Linux) local desktop application that provides a graphical chat interface for interacting with Groq-hosted, function-call-capable models. It bundles a built-in MCP (Model Context Protocol) server enabling smart function calling, letting users chat with an AI, send images, or interact with richer multimodal inputs — all from a friendly desktop UI.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 23
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture intact. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    ChatMCP

    ChatMCP

    ChatMCP is an AI chat client implementing the Model Context Protocol

    ChatMCP is a cross‑platform AI chat client that implements the Model Context Protocol (MCP) to provide unified chat experiences across environments—including desktop, mobile, and web—with synchronization and protocol support tailored for MCP. You can install MCP Server from MCP Server Market, MCP Server Market is a collection of MCP Server, you can use it to chat with different data. Tested on major distributions: Ubuntu, Fedora, Arch Linux, openSUSE. Improved Experience: Latest versions include better dark theme support, unified data storage following XDG Base Directory Specification, and an optimized UI layout for Linux desktop environments is planned.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 25
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    ...For user-facing testing, mistral.rs can provide a built-in web UI, and it also offers a dedicated lightweight web chat interface that supports richer interaction patterns.
    Downloads: 5 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB