131 projects for "simple web server" with 2 filters applied:

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    Markdownify MCP Server

    Markdownify MCP Server

    Convert files and web content into clean, usable Markdown easily

    Markdownify MCP is a Model Context Protocol server that converts many types of files and web content into clean Markdown. It supports formats such as PDFs, images, audio with transcription, DOCX, XLSX, and PPTX, along with web sources like YouTube transcripts, Bing results, and general webpages. Markdownify MCP is designed to simplify content extraction and make data easier to read, share, and reuse in structured workflows.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Browserbase MCP Server

    Browserbase MCP Server

    Allow LLMs to control a browser with Browserbase and Stagehand

    Browserbase MCP Server is a server implementation of the Model Context Protocol (MCP) that enables large language models to interact with web browsers programmatically through cloud-based automation. The project provides a standardized interface for connecting AI systems to real-world web environments, allowing them to navigate pages, extract structured data, and perform user-like actions such as clicking, typing, and form submission.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    WhatsApp MCP Server

    WhatsApp MCP Server

    WhatsApp MCP server enabling AI access to chats and messaging

    whatsapp-mcp is an open source Model Context Protocol (MCP) server that enables AI agents to interact directly with a user’s WhatsApp account through a structured interface. It acts as a bridge between WhatsApp and large language models, allowing controlled access to messages, chats, and contacts. whatsapp-mcp is composed of two main components: a Go-based bridge that connects to the WhatsApp Web API and stores data locally, and a Python-based MCP server that exposes tools for AI interaction. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    ArXiv MCP Server

    ArXiv MCP Server

    A Model Context Protocol server for searching and analyzing arXiv

    arxiv-mcp-server bridges AI assistants and the arXiv repository through a clean MCP interface, enabling search, metadata retrieval, and content access without bespoke scraping. With simple tools like “search” and “fetch,” an agent can find papers, pull abstracts, and download PDFs for downstream summarization or analysis. The project includes packaging and CI to publish to PyPI, plus tests and linting for reliability.
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 5
    Live API Web Console

    Live API Web Console

    A react-based starter app for using the Live API over websockets

    ...Configuration lives in a simple .env file and the project boots with standard web tooling, letting you experiment quickly with models, system prompts, and tool declarations. It ships with demo branches that show grounded search, function calling, and visualization—one example has the model calling a function that renders Vega/Altair graphs directly in the UI. Under the hood there’s an event-emitting WebSocket client, an audio in/out processing layer, and a minimal scaffolded view so you can focus on your app logic rather than wiring.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Pinchtab

    Pinchtab

    High-performance browser automation bridge and orchestrator

    Pinchtab is a lightweight browser automation backend built specifically for AI agents that need efficient, programmatic web control. Implemented as a small standalone HTTP server, it allows any agent or script to interact with web pages using simple API calls instead of heavyweight browser frameworks. The tool emphasizes accessibility-first snapshots that dramatically reduce token usage compared to screenshot-based approaches, making it cost-effective for large-scale automation. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 7
    mcpo

    mcpo

    A simple, secure MCP-to-OpenAPI proxy server

    mcpo is a minimal bridge that exposes any MCP tool as an OpenAPI-compatible HTTP server. Instead of writing glue code, you point mcpo at an MCP server command and it generates REST endpoints and an OpenAPI spec that other systems (or LLM agent frameworks) can call immediately. This design lets you reuse a growing library of MCP servers with platforms that only understand HTTP+OpenAPI, unifying tool access across ecosystems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Agent Chat UI

    Agent Chat UI

    Web app for interacting with any LangGraph agent (PY & TS) via a chat

    Agent Chat UI is an open-source web application that provides a graphical interface for interacting with AI agents built using LangGraph and related frameworks. The project is implemented as a modern Next.js application and allows users to chat with agent workflows running on remote or local LangGraph servers. Through a simple configuration process, developers can connect the interface to a deployed agent by specifying the server URL, assistant identifier, and authentication credentials. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Pocket TTS

    Pocket TTS

    A TTS that fits in your CPU (and pocket)

    Pocket TTS is a lightweight text-to-speech project designed to run efficiently on CPUs, targeting developers who want local speech generation without depending on GPUs or hosted web APIs. It is built to feel practical in everyday applications, where installation and usage should be as simple as adding a dependency and calling a function. The project focuses on keeping the runtime footprint manageable while still producing natural-sounding speech, which makes it attractive for offline tools, prototypes, and privacy-sensitive workflows. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Happy Coder

    Happy Coder

    Mobile and Web client for Codex and Claude Code, with realtime voice

    ...The project includes components like a dedicated backend server for encrypted sync, a rich front-end experience across web and native apps, and support for push notifications when your coding agent encounters permission requests or errors. Happy prioritizes security with end-to-end encryption so your code and interactions remain private and auditable.
    Downloads: 34 This Week
    Last Update:
    See Project
  • 12
    Actionbook

    Actionbook

    Browser action engine for AI agents. 10× faster, resilient by design

    ...This design makes browsing up to 10× faster and far more resilient than ad-hoc approaches that break on minor page changes, because the action manuals codify expected flows and DOM targets. It provides multiple integration paths — a Rust-based CLI, MCP server support for AI IDEs, and a JavaScript SDK — so developers can plug it into a wide range of agent pipelines and toolchains.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    Gemini MCP Tool

    Gemini MCP Tool

    MCP server that enables AI assistants to interact with Google Gemini

    Gemini MCP Tool is a lightweight MCP server that enables AI assistants to interface directly with the Google Gemini CLI, effectively extending the capabilities of one AI system by leveraging another. Its primary purpose is to allow tools like Claude or other MCP-compatible agents to offload complex or large-scale analysis tasks to Gemini, particularly when working with large files or codebases.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    Index

    Index

    The SOTA Open-Source Browser Agent

    ...The project is built to integrate easily with applications through a simple programming interface, allowing developers to embed browser automation capabilities directly into their software systems. Index can perform tasks such as navigating pages, filling forms, collecting data, and analyzing web content without requiring manual scripting for each website.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Reader LLM

    Reader LLM

    Convert any URL to an LLM-friendly input with a simple prefix

    Reader LLM is an open-source tool designed to convert web content into formats that are easier for large language models to process. The system works by transforming a webpage into a clean text or Markdown representation that removes unnecessary formatting and highlights the core information within the page. Developers can use a simple URL prefix to retrieve a version of a webpage that has been optimized for machine consumption, making it suitable for use in AI agents or retrieval-augmented generation pipelines. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Superglue

    Superglue

    Builds integrations and tools from natural language

    ...The platform also includes self-healing capabilities, meaning it can detect and repair failures in workflows when upstream APIs change or break. Superglue supports multiple interfaces, including a web application, SDK, and MCP server, allowing it to be used in both development and production environments.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    BrowserAI

    BrowserAI

    Run local LLMs like llama, deepseek, kokoro etc. inside your browser

    BrowserAI is a cutting-edge platform that allows users to run large language models (LLMs) directly in their web browser without the need for a server. It leverages WebGPU for accelerated performance and supports offline functionality, making it a highly efficient and privacy-conscious solution. The platform provides a developer-friendly SDK with pre-configured popular models, and it allows for seamless switching between MLC and Transformer engines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Reader 3

    Reader 3

    Quick illustration of how one can easily read books together with LLMs

    This project is a minimalist, self-hosted EPUB reader designed to help users browse and read EPUB books one chapter at a time through a lightweight local server, making it especially easy to extract or work with chapters in external tools like large language models. It was created primarily as a simple demonstration of how to combine local book reading with LLM workflows without heavy dependencies or complicated setup, and it runs with just a small Python script and a basic HTTP server.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    ClawHost

    ClawHost

    Deploy OpenClaw with one click

    ClawHost is an open-source, self-hostable cloud hosting platform designed to simplify the deployment of OpenClaw onto a dedicated VPS in minutes, giving users full control over their AI infrastructure without relying on shared or managed services. It automates server provisioning, DNS configuration, SSL certificates, and firewall setup, so developers can focus on running their AI workloads rather than configuring infrastructure manually. The platform includes a user-friendly web dashboard and REST API that let users select regions, manage servers, and configure SSH keys with minimal friction. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    wllama

    wllama

    WebAssembly binding for llama.cpp - Enabling on-browser LLM inference

    wllama is a WebAssembly-based library that enables large language model inference directly inside a web browser. Built as a binding for the llama.cpp inference engine, the project allows developers to run LLM models locally without requiring a server backend or dedicated GPU hardware. The library leverages WebAssembly SIMD capabilities to achieve efficient execution within modern browsers while maintaining compatibility across platforms.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    ...The system supports additional features such as web search, intelligent agent workflows, and multi-model conversations within a single session. AingDesk can be deployed locally on personal machines or installed as a server using containerized environments. Its design emphasizes accessibility, making it suitable for both beginners and experienced developers who want to experiment with AI tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Chandra

    Chandra

    OCR model for complex documents with layout-aware structured outputs

    ...It is capable of handling over 40 languages and is optimized to read difficult inputs such as messy handwriting and multi-column layouts. Chandra can be run locally using transformer-based inference or deployed with a high-performance server setup for large-scale processing. It also includes command-line tools and optional web-based interfaces to simplify interaction and batch processing workflows.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    FastRTC

    FastRTC

    The python library for real-time communication

    FastRTC is a Python library designed to simplify real-time communication (RTC), especially for audio and video streaming applications. It abstracts away much of the complexity that typically comes with implementing WebRTC by providing a simple interface — e.g. a Stream class — that can be mounted within a web backend (for example a FastAPI application). This makes it particularly well suited for building real-time voice (or video) interfaces for applications such as AI assistants, live chat, or collaborative audio/video tools. FastRTC also integrates nicely with UI frameworks (e.g. via a web demo using Gradio), so developers can rapidly prototype and deploy real-time streaming applications without deep knowledge of low-level WebRTC internals. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Fragments by E2B

    Fragments by E2B

    Open source template for AI-powered code generation apps w/ sandboxes

    ...It supports multiple programming frameworks including Python interpreters, Next.js, Vue.js, Streamlit, and Gradio, allowing generated projects to span from simple scripts to full web applications. It integrates with a wide range of large language model providers and supports streaming responses.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    edge-tts

    edge-tts

    Use Microsoft Edge's online text-to-speech service from Python

    ...From the CLI you can adjust parameters such as speaking rate, volume, and pitch, giving you some control over prosody without diving into SSML. The library is asynchronous under the hood, which makes it efficient for batch jobs or web services that need to synthesize many utterances concurrently.
    Downloads: 19 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB