Showing 41 open source projects for "web server local"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    AingDesk is an open-source desktop and server-based AI assistant platform designed to provide a user-friendly environment for interacting with language models and building AI-powered tools. The software enables users to run local AI models or connect to external model APIs through a unified interface. One of its primary goals is to simplify the process of building knowledge-based assistants by allowing users to create local knowledge bases that the AI can search and analyze. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    webclaw

    webclaw

    Fast, local-first web content extraction for LLMs

    ...The tool addresses a major inefficiency in AI workflows by removing irrelevant elements like navigation menus, ads, and scripts, significantly reducing token usage when feeding data into language models. It supports multiple modes of operation, including CLI usage, REST API access, and an MCP server for direct integration with agent-based systems. Webclaw also provides advanced capabilities such as recursive crawling, structured JSON extraction, summarization, and content comparison, making it suitable for research and data pipelines. Its local-first architecture ensures privacy and eliminates the need for API keys.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    ChatGLM3

    ChatGLM3

    ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat

    ...It keeps the series’ smooth dialog and low deployment cost while adding native tool use (function calling), a built-in code interpreter, and agent-style workflows. The family includes base and long-context variants (8K/32K/128K). The repo ships Python APIs, CLI and web demos (Gradio/Streamlit), an OpenAI-format API server, and a compact fine-tuning kit. Quantization (4/8-bit), CPU/MPS support, and accelerator backends (TensorRT-LLM, OpenVINO, chatglm.cpp) enable lightweight local or edge deployment.
    Downloads: 4 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    ...For user-facing testing, mistral.rs can provide a built-in web UI, and it also offers a dedicated lightweight web chat interface that supports richer interaction patterns.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Agent Chat UI

    Agent Chat UI

    Web app for interacting with any LangGraph agent (PY & TS) via a chat

    Agent Chat UI is an open-source web application that provides a graphical interface for interacting with AI agents built using LangGraph and related frameworks. The project is implemented as a modern Next.js application and allows users to chat with agent workflows running on remote or local LangGraph servers. Through a simple configuration process, developers can connect the interface to a deployed agent by specifying the server URL, assistant identifier, and authentication credentials. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Farfalle

    Farfalle

    AI search engine - self-host with local or cloud LLMs

    ...The project integrates large language models with multiple search APIs so that the system can gather information from external sources and synthesize responses into concise answers. It can run either with local language models or with cloud-based providers, allowing developers to deploy it privately or integrate with hosted AI services. The architecture separates the frontend and backend, using modern web technologies such as Next.js and FastAPI to deliver an interactive interface and scalable server logic. Farfalle also includes an agent-based search workflow that plans queries and executes multiple search steps to produce more accurate results than traditional keyword searches. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    wllama

    wllama

    WebAssembly binding for llama.cpp - Enabling on-browser LLM inference

    wllama is a WebAssembly-based library that enables large language model inference directly inside a web browser. Built as a binding for the llama.cpp inference engine, the project allows developers to run LLM models locally without requiring a server backend or dedicated GPU hardware. The library leverages WebAssembly SIMD capabilities to achieve efficient execution within modern browsers while maintaining compatibility across platforms.
    Downloads: 1 This Week
    Last Update:
    See Project
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 10
    Anyquery

    Anyquery

    Query anything (GitHub, Notion, +40 more) with SQL and let LLMs

    ...Built on top of SQLite, the engine uses a plugin architecture that allows it to extend support to dozens of external services and data sources. Users can query structured files such as CSV, JSON, and Parquet as well as remote data sources like SaaS APIs, cloud storage services, and local applications. The platform also supports querying multiple data sources simultaneously and joining them together within a single SQL query, enabling powerful cross-system analysis. In addition to operating as a local query engine, the system can run as a MySQL-compatible server so that traditional database tools can connect to it.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    PowerInfer

    PowerInfer

    High-speed Large Language Model Serving for Local Deployment

    ...PowerInfer incorporates specialized algorithms and sparse operators to manage neuron activation patterns and minimize data transfers between hardware components. As a result, it enables powerful language models to run on consumer hardware while achieving performance comparable to more expensive server-grade systems.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    LocalAI

    LocalAI

    The free, Open Source alternative to OpenAI, Claude and others

    LocalAI is an open-source platform that allows users to run large language models and other AI systems locally on their own hardware. It acts as a drop-in replacement for APIs such as OpenAI, enabling developers to build AI-powered applications without relying on external cloud services. The platform supports a wide range of model types, including text generation, image creation, speech processing, and embeddings. LocalAI can run on consumer-grade hardware and does not necessarily require a...
    Downloads: 60 This Week
    Last Update:
    See Project
  • 13
    Reader 3

    Reader 3

    Quick illustration of how one can easily read books together with LLMs

    This project is a minimalist, self-hosted EPUB reader designed to help users browse and read EPUB books one chapter at a time through a lightweight local server, making it especially easy to extract or work with chapters in external tools like large language models. It was created primarily as a simple demonstration of how to combine local book reading with LLM workflows without heavy dependencies or complicated setup, and it runs with just a small Python script and a basic HTTP server.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Extractous

    Extractous

    Fast and efficient unstructured data extraction

    Extractous is a Rust-based unstructured data extraction library focused on fast local parsing of documents and other content-heavy files. Its purpose is to extract text and metadata efficiently from formats such as PDF, Word, HTML, email archives, images, and more, without depending on external APIs or separate parsing servers. The project emphasizes performance and low memory usage, and its maintainers describe it as a local-first alternative to heavier extraction stacks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    LangChain-ChatGLM-Webui

    LangChain-ChatGLM-Webui

    Automatic question answering for local knowledge bases based on LLM

    LangChain-ChatGLM-Webui is an open-source web interface that integrates the ChatGLM large language model with the LangChain framework to create an interactive conversational AI platform. The project provides a graphical interface that allows users to interact with language models through chat sessions while also connecting those models to external knowledge sources. It supports retrieval-augmented generation workflows that enable the system to answer questions based on local documents or knowledge bases. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Hollama

    Hollama

    A minimal LLM chat app that runs entirely in your browser

    Hollama is a lightweight open-source chat application designed to run entirely within the browser while interacting with large language model servers. The project provides a minimal but powerful user interface for communicating with local or remote LLMs, including servers powered by Ollama or OpenAI-compatible APIs. Because the application runs as a static web interface, it does not require complex backend infrastructure and can be easily deployed or self-hosted. Hollama supports both text-based and multimodal interactions, allowing users to work with models that process images as well as text. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 17
    node-llama-cpp

    node-llama-cpp

    Run AI models locally on your machine with node.js bindings for llama

    node-llama-cpp is a JavaScript and Node.js binding that allows developers to run large language models locally using the high-performance inference engine provided by llama.cpp. The library enables applications built with Node.js to interact directly with local LLM models without requiring a remote API or external service. By using native bindings and optimized model execution, the framework allows developers to integrate advanced language model capabilities into desktop applications, server software, and command-line tools. The system automatically detects the available hardware on a machine and selects the most appropriate compute backend, including CPU or GPU acceleration. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    MLC LLM

    MLC LLM

    Universal LLM Deployment Engine with ML Compilation

    MLC LLM is a machine learning compiler and deployment framework designed to enable efficient execution of large language models across a wide range of hardware platforms. The project focuses on compiling models into optimized runtimes that can run natively on devices such as GPUs, mobile processors, browsers, and edge hardware. By leveraging machine learning compilation techniques, mlc-llm produces high-performance inference engines that maintain consistent APIs across platforms. The system...
    Downloads: 22 This Week
    Last Update:
    See Project
  • 19
    LLocalSearch

    LLocalSearch

    LLocalSearch is a completely locally running search aggregator

    LLocalSearch is an open-source search engine framework designed to run entirely on local infrastructure using large language model agents to gather and synthesize information from the web. The system allows users to submit natural language questions, after which a chain of LLM-driven agents recursively searches for relevant information and compiles a response. Unlike many AI search tools, LLocalSearch operates without requiring external cloud APIs or proprietary services, making it suitable for privacy-focused or offline environments. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    PasteGuard

    PasteGuard

    Masks sensitive data and secrets before they reach AI

    PasteGuard is an open-source privacy proxy that protects sensitive information like personal data and API secrets by detecting and masking them before they reach large language model APIs such as OpenAI or Anthropic Claude. It sits between an application and the LLM provider, automatically replacing names, emails, tokens, and other personally identifiable information (PII) with placeholders so that external services never see raw sensitive values, and then optionally unmasking them in the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    SeaGOAT

    SeaGOAT

    local-first semantic code search engine

    SeaGOAT is an open-source semantic code search engine designed to help developers explore and understand large codebases more efficiently. Instead of relying solely on traditional keyword search, it uses vector embeddings to represent the meaning of code and queries, allowing users to perform semantic searches that find relevant code even when the exact keywords are not present. The tool runs locally on a developer’s machine and processes repositories using a combination of embedding models...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    AI as Workspace

    AI as Workspace

    An elegant AI chat client. Full-featured, lightweight

    AI as Workspace, short for AI as Workspace, is an open-source AI client application that provides a unified interface for interacting with multiple large language models and AI tools within a single workspace environment. The platform is designed as a lightweight yet powerful desktop or web application that organizes AI interactions through structured workspaces. Instead of managing individual chat sessions separately, users can group conversations, artifacts, and tasks within customizable...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 23
    repo2txt

    repo2txt

    Web-based tool converts GitHub repository contents

    repo2txt is an open-source developer tool that converts the contents of a code repository into a single structured text file that can be easily consumed by large language models. The tool is designed to address the challenge of analyzing entire codebases with AI assistants, where code is normally distributed across many files and directories. By collecting repository contents and formatting them into a single text document, repo2txt allows developers to feed complete projects into AI systems...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    LLM Council

    LLM Council

    LLM Council works together to answer your hardest questions

    LLM Council is a creative open-source web application by Andrej Karpathy that lets you consult multiple large language models together to answer questions more reliably than querying a single model. Instead of relying on one provider, this application sends your query simultaneously to several LLMs supported via OpenRouter, collects each model’s independent response, and then orchestrates a multi-stage evaluation where the models critique and rank each other’s outputs anonymously. After this...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    oterm

    oterm

    the terminal client for Ollama

    Oterm is an open-source terminal client designed to provide a lightweight command-line interface for interacting with large language models through the Ollama ecosystem. The tool allows users to chat with local AI models directly from the terminal without needing a graphical interface or web application. Its interface is designed to be simple and intuitive, enabling developers to launch conversations quickly using a single command. Oterm supports persistent chat sessions that store conversations, system prompts, and parameter configurations locally in a database. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB