Showing 71 open source projects for "server performance"

View related business solutions
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    Velocity server

    Velocity server

    The modern, next-generation Minecraft server proxy

    ...The software is designed with a focus on performance, scalability, and modern architecture, allowing it to handle thousands of simultaneous players efficiently. Velocity also includes a plugin API that allows developers to extend the proxy with custom functionality and integrate it with existing server tools. Compared with older proxy solutions, the project emphasizes improved performance, reliability, and cleaner software design.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    Tiledesk Server

    Tiledesk Server

    Tiledesk Server is the main API component of the Tiledesk platform

    Tiledesk Server is the backend component of the Tiledesk platform, providing a comprehensive open-source live chat system with integrated chatbot capabilities for customer support and engagement.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    VikingDB MCP Server

    VikingDB MCP Server

    A mcp server for vikingdb store and search

    An MCP server that interfaces with VikingDB, a high-performance vector database developed by ByteDance, enabling efficient vector storage and search capabilities. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    DB MCP Server

    DB MCP Server

    A powerful multi-database server implementing the MCP

    The DB MCP Server is a powerful multi-database server implementing the Model Context Protocol (MCP) to provide AI assistants with structured access to databases. Built on the FreePeak/cortex framework, it enables execution of SQL queries, transaction management, schema exploration, and performance analysis across different database systems through a unified interface. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    csghub-server

    csghub-server

    csghub-server is the backend server for CSGHub

    csghub-server is the backend component of the CSGHub platform, an open-source infrastructure designed to manage and operate large language models, datasets, and AI development workflows within a private deployment environment. The server acts as a centralized management layer that allows teams to store, organize, and operate AI assets such as models, datasets, and machine learning applications in a manner similar to artifact repositories used in software engineering. Built primarily in the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    MCP Server MySQL

    MCP Server MySQL

    A Model Context Protocol server

    A Model Context Protocol server that provides access to MySQL databases, enabling Large Language Models to inspect database schemas and execute SQL queries, facilitating seamless database interactions. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    OpenVINO Model Server

    OpenVINO Model Server

    A scalable inference server for models optimized with OpenVINO

    OpenVINO™ Model Server is a high-performance inference serving system designed to host and serve machine learning models that have been optimized with the OpenVINO toolkit. It’s implemented in C++ for scalability and efficiency, making it suitable for both edge and cloud deployments where inference workloads must be reliable and high throughput. The server exposes model inference via standard network protocols like REST and gRPC, allowing any client that speaks those protocols to request predictions remotely, abstracting away the complexity of where and how the model runs. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    Browserbase MCP Server

    Browserbase MCP Server

    Allow LLMs to control a browser with Browserbase and Stagehand

    Browserbase MCP Server is a server implementation of the Model Context Protocol (MCP) that enables large language models to interact with web browsers programmatically through cloud-based automation. The project provides a standardized interface for connecting AI systems to real-world web environments, allowing them to navigate pages, extract structured data, and perform user-like actions such as clicking, typing, and form submission.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Google Workspace MCP Server

    Google Workspace MCP Server

    Control Gmail, Google Calendar, Docs, Sheets, Slides, Chat, Forms

    Google Workspace MCP is an open-source server that connects AI assistants to Google Workspace services through the Model Context Protocol (MCP), allowing large language models to interact directly with productivity tools. The project exposes a wide set of Google services including Gmail, Google Drive, Docs, Sheets, Slides, Calendar, Chat, and other Workspace components as structured tools that an AI system can call programmatically. By acting as a bridge between AI clients and the Google...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 10
    Lemonade

    Lemonade

    Lemonade helps users run local LLMs with the highest performance

    Lemonade is a local LLM runtime that aims to deliver the highest possible performance on your own hardware by auto-configuring state-of-the-art inference engines for both NPUs and GPUs. The project positions itself as a “local LLM server” you can run on laptops and workstations, abstracting away backend differences while giving you a single place to serve and manage models. Its README emphasizes real-world adoption across startups, research groups, and large companies, signaling a focus on practical deployments rather than toy demos. ...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 11
    Chrome DevTools MCP

    Chrome DevTools MCP

    Chrome DevTools for coding agents

    ...The repository spells out environment requirements and cautions that exposing a live browser to agents grants powerful access, so sensitive data should be handled carefully. Beyond static inspection, it exposes operational tools like starting a performance trace that an agent can later analyze to propose optimizations. The server is intended to slot into MCP-capable assistants and IDEs, giving them reliable, typed tools and resource endpoints rather than ad-hoc automation. Documentation from the Chrome team explains how the server augments agents with real debugging capabilities.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 12
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    Shadcn UI v4 MCP Server

    Shadcn UI v4 MCP Server

    A mcp server to allow LLMS gain context about shadcn ui component

    ...The server supports multiple frontend frameworks including React, Svelte, Vue, and React Native, making it highly versatile for cross-platform development. It includes smart caching and efficient GitHub API usage to optimize performance and handle rate limits during component retrieval. The system also supports multiple transport modes such as standard input/output and Server-Sent Events, enabling both local and distributed deployments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Nestia

    Nestia

    NestJS Helper + AI Chatbot Development

    Nestia is a high-performance toolkit and ecosystem built on top of NestJS that enhances backend development by introducing strongly typed APIs, automated SDK generation, and advanced tooling for scalable server applications. It is designed to eliminate much of the boilerplate typically associated with API development by leveraging pure TypeScript types to automatically generate validation logic, API documentation, and client SDKs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and device-mapping strategies, helping users get strong performance without manual tuning. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    MongoDB Lens

    MongoDB Lens

    MongoDB Lens: Full Featured MCP Server for MongoDB Databases

    MongoDB Lens is a local Model Context Protocol (MCP) server offering full-featured access to MongoDB databases using natural language via large language models (LLMs). It enables users to perform queries, run aggregations, optimize performance, and more through conversational interactions. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Angel

    Angel

    A Flexible and Powerful Parameter Server for large-scale ML

    Angel is a high-performance distributed machine learning and graph computing platform based on the philosophy of Parameter Server. It is tuned for performance with big data from Tencent and has a wide range of applicability and stability, demonstrating an increasing advantage in handling higher-dimension models. Angel is jointly developed by Tencent and Peking University, taking account of both high availability in industry and innovation in academia.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    FastSD CPU

    FastSD CPU

    Fast stable diffusion on CPU and AI PC

    ...The repository contains multiple interfaces including a desktop GUI for simple generation, an advanced web-based UI with support for extensions like LoRA and ControlNet, and a command-line interface for scripted usage or server deployments. With support for performance-oriented libraries such as OpenVINO and hardware acceleration on platforms like Intel AI PCs, FastSD CPU aims to shrink generation times dramatically compared with naive CPU implementations.
    Downloads: 43 This Week
    Last Update:
    See Project
  • 19
    Text Generation Inference

    Text Generation Inference

    Large Language Model Text Generation Inference

    Text Generation Inference is a high-performance inference server for text generation models, optimized for Hugging Face's Transformers. It is designed to serve large language models efficiently with optimizations for performance and scalability.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Zed

    Zed

    High-performance, multiplayer code editor from the creators of Atom

    Zed is a next-generation code editor designed for high-performance collaboration with humans and AI. Written from scratch in Rust to efficiently leverage multiple CPU cores and your GPU. Integrate upcoming LLMs into your workflow to generate, transform, and analyze code. Chat with teammates, write notes together, and share your screen and project. Multibuffers compose excerpts from across the codebase in one editable surface.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 21
    Chat2DB

    Chat2DB

    AI-driven database tool and SQL client

    Save time by working with data. Connect to all your data sources, and instantly generate optimal SQL for fast lightning information. If you don't know SQL well, you can get instant information without writing SQL. Generate high-performance SQL for your complicated queries using natural language, as well as correcting errors and getting AI suggestions to optimize the performance of SQL queries. Developers can write complex SQL queries quickly and accurately with the help of the AI SQL editor,...
    Downloads: 15 This Week
    Last Update:
    See Project
  • 22
    Arcade AI

    Arcade AI

    Arcade Tool Development Kit (TDK), Worker, Evals, and CLI

    ...This repository contains the core Arcade libraries, organized as separate packages for maximum flexibility and modularity. Evaluation framework for testing tool performance. Test your MCP server's tools, resources, prompts, elicitation, and OAuth 2. MCPJam is compliant with the latest MCP specs. Connect to any MCP server. MCPJam inspector supports STDIO, SSE, and Streamable HTTP transports.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 23
    BrowserAI

    BrowserAI

    Run local LLMs like llama, deepseek, kokoro etc. inside your browser

    BrowserAI is a cutting-edge platform that allows users to run large language models (LLMs) directly in their web browser without the need for a server. It leverages WebGPU for accelerated performance and supports offline functionality, making it a highly efficient and privacy-conscious solution. The platform provides a developer-friendly SDK with pre-configured popular models, and it allows for seamless switching between MLC and Transformer engines. Additionally, it supports features such as speech recognition, text-to-speech, structured output generation, and Web Worker support for non-blocking UI performance.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Telegram Drive

    Telegram Drive

    Telegram Drive

    Telegram Drive is a powerful utility that enables you to organize your telegram files and much more. Teldrive stands out among similar tools, thanks to its implementation in Go, a language known for its efficiency. Its performance surpasses alternatives written in Python and other languages, with the exception of Rust. Teldrive not only excels in speed but also offers an intuitive user interface for efficient file interaction which other tool lacks. Its compatibility with Rclone further...
    Downloads: 57 This Week
    Last Update:
    See Project
  • 25
    MCP Image Compression

    MCP Image Compression

    A high-performance image compression microservice based on MCP

    The MCP Image Compression server is a high-performance microservice based on the Model Context Protocol architecture. It focuses on providing fast and high-quality image compression capabilities to help developers optimize image resources for websites and applications, improving loading speed and user experience. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB