Showing 105 open source projects for "server local"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 1
    pyTelegramBotAPI

    pyTelegramBotAPI

    Python Telegram bot api.

    TeleBot is the synchronous and asynchronous implementation of Telegram Bot API.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 2
    JMusicBot

    JMusicBot

    A Discord music bot that's easy to set up and run yourself

    A cross-platform Discord music bot with a clean interface, and that is easy to set up and run yourself! Easy to run (just make sure Java is installed, and run!) Fast loading of songs. No external keys are needed (besides a Discord Bot token) Smooth playback. Server-specific setup for the "DJ" role that can moderate the music. Clean and beautiful menus. Supports many sites, including Youtube, Soundcloud, and more. Supports many online radio/streams. Supports local files. Playlist support (both web/youtube and local) This bot (and the source code here) might not be easy to edit for inexperienced programmers. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    MongoDB Lens

    MongoDB Lens

    MongoDB Lens: Full Featured MCP Server for MongoDB Databases

    MongoDB Lens is a local Model Context Protocol (MCP) server offering full-featured access to MongoDB databases using natural language via large language models (LLMs). It enables users to perform queries, run aggregations, optimize performance, and more through conversational interactions. ​
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    XcodeBuildMCP

    XcodeBuildMCP

    A Model Context Protocol (MCP) server

    ...The server aims to be “agent-ready,” surfacing capabilities (build, clean, test, archive, install, run, log collection) with explicit schemas instead of brittle prompt instructions. It supports MCP transports suitable for local IDEs and service deployments, and pairs with a public website that positions it as a bridge between Xcode and autonomous assistants.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    Desktop Commander MCP

    Desktop Commander MCP

    AI-powered MCP server for desktop file and terminal automation

    Desktop Commander MCP is an advanced Model Context Protocol server designed to extend AI assistants with direct control over a user’s local machine, including the file system and terminal. It integrates with clients like Claude Desktop to enable AI-driven workflows such as editing files, executing commands, and automating development tasks from a single conversational interface. Desktop Commander MCP builds on top of an MCP filesystem server and enhances it with powerful search, replace, and code editing capabilities tailored for real-world development environments. ...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 6
    ChatMCP

    ChatMCP

    ChatMCP is an AI chat client implementing the Model Context Protocol

    ChatMCP is a cross‑platform AI chat client that implements the Model Context Protocol (MCP) to provide unified chat experiences across environments—including desktop, mobile, and web—with synchronization and protocol support tailored for MCP. You can install MCP Server from MCP Server Market, MCP Server Market is a collection of MCP Server, you can use it to chat with different data. Tested on major distributions: Ubuntu, Fedora, Arch Linux, openSUSE. Improved Experience: Latest versions...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 7
    stt

    stt

    Voice Recognition to Text Tool

    ...It leverages open-source speech models such as Faster-Whisper to recognize and transcribe human speech into plain text, structured JSON objects, or subtitle files with time codes, making it suitable for both personal and professional transcription tasks. The project is designed to be easy to deploy: you can run a local Python server that exposes an HTTP API for uploading audio/video files and retrieving transcriptions in different formats. It supports GPU acceleration if available, enabling faster processing on compatible hardware but still offers reliable performance on CPUs alone.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Pocket TTS

    Pocket TTS

    A TTS that fits in your CPU (and pocket)

    ...Because it is CPU-oriented, it fits well in server environments where GPU access is limited, in desktop apps, or in edge deployments where simplicity matters more than maximum throughput. It also emphasizes developer ergonomics, providing a straightforward API surface that can be integrated into pipelines, assistants, accessibility tools, or batch generation scripts.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 9
    Reader 3

    Reader 3

    Quick illustration of how one can easily read books together with LLMs

    This project is a minimalist, self-hosted EPUB reader designed to help users browse and read EPUB books one chapter at a time through a lightweight local server, making it especially easy to extract or work with chapters in external tools like large language models. It was created primarily as a simple demonstration of how to combine local book reading with LLM workflows without heavy dependencies or complicated setup, and it runs with just a small Python script and a basic HTTP server.
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    BrowserAI

    BrowserAI

    Run local LLMs like llama, deepseek, kokoro etc. inside your browser

    BrowserAI is a cutting-edge platform that allows users to run large language models (LLMs) directly in their web browser without the need for a server. It leverages WebGPU for accelerated performance and supports offline functionality, making it a highly efficient and privacy-conscious solution. The platform provides a developer-friendly SDK with pre-configured popular models, and it allows for seamless switching between MLC and Transformer engines. Additionally, it supports features such as...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 11
    Open Responses

    Open Responses

    Specification for multi-provider, interoperable LLM interfaces

    Open Responses is an open-source implementation of an API compatible with the OpenAI Responses API that lets developers self-host a drop-in alternative endpoint for AI interactions while preserving compatibility with existing Agents SDKs and model workflows. It enables you to run a local or private server that speaks the standard Responses API, so tools, applications, and agents built against that API can operate without contacting OpenAI’s cloud and can instead route calls to any large language model provider you choose, such as Claude, Qwen, Ollama, or others. This makes it a powerful option for teams or individuals who want full control over their AI infrastructure, prioritize privacy, or need to standardize inference calls across multiple backends without rewriting their code.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    Extractous

    Extractous

    Fast and efficient unstructured data extraction

    Extractous is a Rust-based unstructured data extraction library focused on fast local parsing of documents and other content-heavy files. Its purpose is to extract text and metadata efficiently from formats such as PDF, Word, HTML, email archives, images, and more, without depending on external APIs or separate parsing servers. The project emphasizes performance and low memory usage, and its maintainers describe it as a local-first alternative to heavier extraction stacks. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    OctoMind MCP

    OctoMind MCP

    An MCP server for octomind tools, resources and prompts

    The Octomind MCP Server is designed to integrate Octomind's end-to-end testing tools and resources into local development environments. It enables AI-powered interfaces to create, execute, and manage e2e tests, enhancing the testing workflow. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Open Claude Cowork

    Open Claude Cowork

    Open Source version of Claude Cowork built with Claude Code

    Open Claude Cowork is an open source desktop chat application that brings the power of autonomous, agent-style AI workflows to your local machine by combining the Claude Agent SDK with the Composio Tool Router, enabling developers and power users to build intelligent assistants that can interact with a vast array of external tools and services. It offers a native Electron-based interface for macOS, Windows, and Linux that feels familiar and modern, supporting persistent, multi-session...
    Downloads: 139 This Week
    Last Update:
    See Project
  • 16
    QMD

    QMD

    mini cli search engine for your docs, knowledge bases, etc.

    QMD is a powerful and lightweight command-line tool that acts as an on-device search engine for your personal knowledge base, allowing you to index and search files like Markdown notes, meeting transcripts, technical documentation, and other text collections without depending on cloud services. Designed to keep all search activity local, it combines classic full-text search techniques with modern semantic features such as vector similarity and hybrid ranking so that queries return not just...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 17
    Secret Llama

    Secret Llama

    Fully private LLM chatbot that runs entirely with a browser

    Secret Llama is a privacy-first large-language-model chatbot that runs entirely inside your web browser, meaning no server is required and your conversation data never leaves your device. It focuses on open-source model support, letting you load families like Llama and Mistral directly in the client for fully local inference. Because everything happens in-browser, it can work offline once models are cached, which is helpful for air-gapped environments or travel.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    DBHub

    DBHub

    Universal database MCP server connecting to MySQL, PostgreSQL

    DBHub is a universal database gateway that implements the MCP server interface so assistants and IDEs can explore and query databases through typed tools. It supports multiple transports—stdio for desktop clients and HTTP for networked scenarios—making it flexible to embed or deploy. Configuration is environment-variable driven, with a DSN and per-engine settings covering Postgres, MySQL, MariaDB, SQL Server, and SQLite. Operational flags include read-only mode, row limits, and even SSH...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    ChatTTS webUI & API

    ChatTTS webUI & API

    A simple native web interface that uses ChatTTS to synthesize text

    ChatTTS-ui is a local web interface and API wrapper around the ChatTTS speech synthesis system, designed to make advanced TTS models easy to use from a browser. It runs a small backend server (Python + Torch + ffmpeg) and exposes a simple webpage where you can type text, adjust parameters, and generate audio. The project supports Chinese, English, and mixed text with digits and control symbols, making it suitable for bilingual content and numerically heavy text like announcements or prompts. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 20
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    FastKoko

    FastKoko

    Dockerized FastAPI wrapper for Kokoro-82M text-to-speech model

    FastKoko is a self-hosted text-to-speech server built around the Kokoro-82M model and exposed through a FastAPI backend. It is designed to be easy to deploy via Docker, with separate CPU and GPU images so that users can choose between pure CPU inference and NVIDIA GPU acceleration. The project exposes an OpenAI-compatible speech endpoint, which means existing code that talks to the OpenAI audio API can often be pointed at a Kokoro-FastAPI instance with minimal changes. It supports multiple...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 22
    MCP Proxy

    MCP Proxy

    A TypeScript SSE proxy for MCP servers that use stdio transport

    mcp-proxy is a lightweight bridge that converts between MCP transports, letting you run a server on stdio and expose it over Streamable HTTP (SSE) or do the reverse. This enables existing desktop-style MCP servers to be reused by web services and IDEs that prefer HTTP, without modifying the server. The tool can multiplex multiple named STDIO servers behind one proxy instance, simplifying fleet deployments or local development with many tools.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    OpenCode

    OpenCode

    The open source coding agent

    OpenCode is an open-source AI coding agent designed to assist developers directly from the terminal, desktop, or IDE environments. It functions as a locally running assistant that can analyze codebases, execute development tasks, and interact with external tools while remaining highly configurable. The system is built around an agent model that can perform actions such as file manipulation, shell execution, and web access with user confirmation. Its provider-agnostic design allows developers...
    Downloads: 75 This Week
    Last Update:
    See Project
  • 24
    OpenAI-Compatible Edge-TTS API

    OpenAI-Compatible Edge-TTS API

    Free, high-quality text-to-speech API endpoint to replace OpenAI

    ...Because it relies on Edge’s TTS, the audio generation itself is free, and the project essentially acts as a smart proxy that handles formatting and streaming. The server supports Server-Sent Events (SSE) for streaming audio, enabling low-latency playback in chat UIs and other interactive tools. A Docker image is provided for one-command deployment, and environment variables can be used to configure default voice, language, response format, authentication, and logging options.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 25
    Farfalle

    Farfalle

    AI search engine - self-host with local or cloud LLMs

    ...The project integrates large language models with multiple search APIs so that the system can gather information from external sources and synthesize responses into concise answers. It can run either with local language models or with cloud-based providers, allowing developers to deploy it privately or integrate with hosted AI services. The architecture separates the frontend and backend, using modern web technologies such as Next.js and FastAPI to deliver an interactive interface and scalable server logic. Farfalle also includes an agent-based search workflow that plans queries and executes multiple search steps to produce more accurate results than traditional keyword searches. ...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB