123 projects for "simple web server" with 2 filters applied:

  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 1
    Markdownify MCP Server

    Markdownify MCP Server

    Convert files and web content into clean, usable Markdown easily

    Markdownify MCP is a Model Context Protocol server that converts many types of files and web content into clean Markdown. It supports formats such as PDFs, images, audio with transcription, DOCX, XLSX, and PPTX, along with web sources like YouTube transcripts, Bing results, and general webpages. Markdownify MCP is designed to simplify content extraction and make data easier to read, share, and reuse in structured workflows.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Browserbase MCP Server

    Browserbase MCP Server

    Allow LLMs to control a browser with Browserbase and Stagehand

    Browserbase MCP Server is a server implementation of the Model Context Protocol (MCP) that enables large language models to interact with web browsers programmatically through cloud-based automation. The project provides a standardized interface for connecting AI systems to real-world web environments, allowing them to navigate pages, extract structured data, and perform user-like actions such as clicking, typing, and form submission.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    WhatsApp MCP Server

    WhatsApp MCP Server

    WhatsApp MCP server enabling AI access to chats and messaging

    whatsapp-mcp is an open source Model Context Protocol (MCP) server that enables AI agents to interact directly with a user’s WhatsApp account through a structured interface. It acts as a bridge between WhatsApp and large language models, allowing controlled access to messages, chats, and contacts. whatsapp-mcp is composed of two main components: a Go-based bridge that connects to the WhatsApp Web API and stores data locally, and a Python-based MCP server that exposes tools for AI interaction. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    ArXiv MCP Server

    ArXiv MCP Server

    A Model Context Protocol server for searching and analyzing arXiv

    arxiv-mcp-server bridges AI assistants and the arXiv repository through a clean MCP interface, enabling search, metadata retrieval, and content access without bespoke scraping. With simple tools like “search” and “fetch,” an agent can find papers, pull abstracts, and download PDFs for downstream summarization or analysis. The project includes packaging and CI to publish to PyPI, plus tests and linting for reliability.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    Live API Web Console

    Live API Web Console

    A react-based starter app for using the Live API over websockets

    ...Configuration lives in a simple .env file and the project boots with standard web tooling, letting you experiment quickly with models, system prompts, and tool declarations. It ships with demo branches that show grounded search, function calling, and visualization—one example has the model calling a function that renders Vega/Altair graphs directly in the UI. Under the hood there’s an event-emitting WebSocket client, an audio in/out processing layer, and a minimal scaffolded view so you can focus on your app logic rather than wiring.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Pinchtab

    Pinchtab

    High-performance browser automation bridge and orchestrator

    Pinchtab is a lightweight browser automation backend built specifically for AI agents that need efficient, programmatic web control. Implemented as a small standalone HTTP server, it allows any agent or script to interact with web pages using simple API calls instead of heavyweight browser frameworks. The tool emphasizes accessibility-first snapshots that dramatically reduce token usage compared to screenshot-based approaches, making it cost-effective for large-scale automation. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 7
    mcpo

    mcpo

    A simple, secure MCP-to-OpenAPI proxy server

    mcpo is a minimal bridge that exposes any MCP tool as an OpenAPI-compatible HTTP server. Instead of writing glue code, you point mcpo at an MCP server command and it generates REST endpoints and an OpenAPI spec that other systems (or LLM agent frameworks) can call immediately. This design lets you reuse a growing library of MCP servers with platforms that only understand HTTP+OpenAPI, unifying tool access across ecosystems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Agent Chat UI

    Agent Chat UI

    Web app for interacting with any LangGraph agent (PY & TS) via a chat

    Agent Chat UI is an open-source web application that provides a graphical interface for interacting with AI agents built using LangGraph and related frameworks. The project is implemented as a modern Next.js application and allows users to chat with agent workflows running on remote or local LangGraph servers. Through a simple configuration process, developers can connect the interface to a deployed agent by specifying the server URL, assistant identifier, and authentication credentials. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Pocket TTS

    Pocket TTS

    A TTS that fits in your CPU (and pocket)

    Pocket TTS is a lightweight text-to-speech project designed to run efficiently on CPUs, targeting developers who want local speech generation without depending on GPUs or hosted web APIs. It is built to feel practical in everyday applications, where installation and usage should be as simple as adding a dependency and calling a function. The project focuses on keeping the runtime footprint manageable while still producing natural-sounding speech, which makes it attractive for offline tools, prototypes, and privacy-sensitive workflows. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 10
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Gemini MCP Tool

    Gemini MCP Tool

    MCP server that enables AI assistants to interact with Google Gemini

    Gemini MCP Tool is a lightweight MCP server that enables AI assistants to interface directly with the Google Gemini CLI, effectively extending the capabilities of one AI system by leveraging another. Its primary purpose is to allow tools like Claude or other MCP-compatible agents to offload complex or large-scale analysis tasks to Gemini, particularly when working with large files or codebases.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    Happy Coder

    Happy Coder

    Mobile and Web client for Codex and Claude Code, with realtime voice

    ...The project includes components like a dedicated backend server for encrypted sync, a rich front-end experience across web and native apps, and support for push notifications when your coding agent encounters permission requests or errors. Happy prioritizes security with end-to-end encryption so your code and interactions remain private and auditable.
    Downloads: 33 This Week
    Last Update:
    See Project
  • 13
    Index

    Index

    The SOTA Open-Source Browser Agent

    ...The project is built to integrate easily with applications through a simple programming interface, allowing developers to embed browser automation capabilities directly into their software systems. Index can perform tasks such as navigating pages, filling forms, collecting data, and analyzing web content without requiring manual scripting for each website.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Reader LLM

    Reader LLM

    Convert any URL to an LLM-friendly input with a simple prefix

    Reader LLM is an open-source tool designed to convert web content into formats that are easier for large language models to process. The system works by transforming a webpage into a clean text or Markdown representation that removes unnecessary formatting and highlights the core information within the page. Developers can use a simple URL prefix to retrieve a version of a webpage that has been optimized for machine consumption, making it suitable for use in AI agents or retrieval-augmented generation pipelines. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Actionbook

    Actionbook

    Browser action engine for AI agents. 10× faster, resilient by design

    ...This design makes browsing up to 10× faster and far more resilient than ad-hoc approaches that break on minor page changes, because the action manuals codify expected flows and DOM targets. It provides multiple integration paths — a Rust-based CLI, MCP server support for AI IDEs, and a JavaScript SDK — so developers can plug it into a wide range of agent pipelines and toolchains.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 16
    BrowserAI

    BrowserAI

    Run local LLMs like llama, deepseek, kokoro etc. inside your browser

    BrowserAI is a cutting-edge platform that allows users to run large language models (LLMs) directly in their web browser without the need for a server. It leverages WebGPU for accelerated performance and supports offline functionality, making it a highly efficient and privacy-conscious solution. The platform provides a developer-friendly SDK with pre-configured popular models, and it allows for seamless switching between MLC and Transformer engines.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Reader 3

    Reader 3

    Quick illustration of how one can easily read books together with LLMs

    This project is a minimalist, self-hosted EPUB reader designed to help users browse and read EPUB books one chapter at a time through a lightweight local server, making it especially easy to extract or work with chapters in external tools like large language models. It was created primarily as a simple demonstration of how to combine local book reading with LLM workflows without heavy dependencies or complicated setup, and it runs with just a small Python script and a basic HTTP server.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    ClawHost

    ClawHost

    Deploy OpenClaw with one click

    ClawHost is an open-source, self-hostable cloud hosting platform designed to simplify the deployment of OpenClaw onto a dedicated VPS in minutes, giving users full control over their AI infrastructure without relying on shared or managed services. It automates server provisioning, DNS configuration, SSL certificates, and firewall setup, so developers can focus on running their AI workloads rather than configuring infrastructure manually. The platform includes a user-friendly web dashboard and REST API that let users select regions, manage servers, and configure SSH keys with minimal friction. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    Anthropic SDK TypeScript

    Anthropic SDK TypeScript

    Access to Anthropic's safety-first language model APIs

    anthropic-sdk-typescript is the TypeScript / JavaScript client library for the Anthropic REST API, enabling backend or Node.js usage of models like Claude. It wraps API endpoints for creating messages, streaming responses, and managing parameters in a type-safe TS environment. The library is designed for server-side use, interfacing with REST, and is stable for integration in web services or backend agents. Example usage shows how to instantiate the Anthropic client, call client.messages.create(...), and obtain responses. It supports streaming endpoints as well. Because TypeScript provides type safety, it helps avoid common errors in JSON interplay. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 20
    wllama

    wllama

    WebAssembly binding for llama.cpp - Enabling on-browser LLM inference

    wllama is a WebAssembly-based library that enables large language model inference directly inside a web browser. Built as a binding for the llama.cpp inference engine, the project allows developers to run LLM models locally without requiring a server backend or dedicated GPU hardware. The library leverages WebAssembly SIMD capabilities to achieve efficient execution within modern browsers while maintaining compatibility across platforms.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    ...The system supports additional features such as web search, intelligent agent workflows, and multi-model conversations within a single session. AingDesk can be deployed locally on personal machines or installed as a server using containerized environments. Its design emphasizes accessibility, making it suitable for both beginners and experienced developers who want to experiment with AI tools.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    FastRTC

    FastRTC

    The python library for real-time communication

    FastRTC is a Python library designed to simplify real-time communication (RTC), especially for audio and video streaming applications. It abstracts away much of the complexity that typically comes with implementing WebRTC by providing a simple interface — e.g. a Stream class — that can be mounted within a web backend (for example a FastAPI application). This makes it particularly well suited for building real-time voice (or video) interfaces for applications such as AI assistants, live chat, or collaborative audio/video tools. FastRTC also integrates nicely with UI frameworks (e.g. via a web demo using Gradio), so developers can rapidly prototype and deploy real-time streaming applications without deep knowledge of low-level WebRTC internals. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Chandra

    Chandra

    OCR model for complex documents with layout-aware structured outputs

    ...It is capable of handling over 40 languages and is optimized to read difficult inputs such as messy handwriting and multi-column layouts. Chandra can be run locally using transformer-based inference or deployed with a high-performance server setup for large-scale processing. It also includes command-line tools and optional web-based interfaces to simplify interaction and batch processing workflows.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 24
    Poco Claw

    Poco Claw

    A more beautiful and easier-to-use alternative to OpenClaw

    Poco Claw is an AI agent platform designed as a more user-friendly and visually polished alternative to traditional OpenClaw implementations. It focuses on improving usability by providing a modern web interface combined with enhanced interaction capabilities such as built-in messaging and project organization tools. The system operates on a sandboxed runtime, ensuring that tasks executed by the agent are isolated from the host environment, which improves security and reliability. It extends beyond simple chatbot functionality by supporting structured workflows, task planning modes, and multi-step execution pipelines. ...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 25
    edge-tts

    edge-tts

    Use Microsoft Edge's online text-to-speech service from Python

    ...From the CLI you can adjust parameters such as speaking rate, volume, and pitch, giving you some control over prosody without diving into SSML. The library is asynchronous under the hood, which makes it efficient for batch jobs or web services that need to synthesize many utterances concurrently.
    Downloads: 21 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB