Showing 185 open source projects for "openssh-server"

View related business solutions
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 1
    TensorRT LLM

    TensorRT LLM

    TensorRT LLM provides users with an easy-to-use Python API

    ...It includes support for cutting-edge inference methods like speculative decoding and inflight batching, enabling real-time and large-scale AI applications. TensorRT-LLM integrates seamlessly with NVIDIA’s broader inference ecosystem, including Triton Inference Server and distributed deployment frameworks, making it suitable for production environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Reader 3

    Reader 3

    Quick illustration of how one can easily read books together with LLMs

    This project is a minimalist, self-hosted EPUB reader designed to help users browse and read EPUB books one chapter at a time through a lightweight local server, making it especially easy to extract or work with chapters in external tools like large language models. It was created primarily as a simple demonstration of how to combine local book reading with LLM workflows without heavy dependencies or complicated setup, and it runs with just a small Python script and a basic HTTP server.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Archon

    Archon

    The knowledge and task management backbone for AI coding assistants

    Archon is an open-source “command center” designed to enhance AI coding assistant workflows by giving developers a centralized environment for knowledge management, context engineering, and task coordination across AI agents. It acts as a backend (including an MCP server) that allows different AI coding tools and assistants to share the same structured context, knowledge base, and task lists, improving consistency, productivity, and collaboration across multi-agent interactions. Users can import documentation, project files, and external knowledge so that assistants like Claude Code, Cursor, or other LLM-powered tools work with up-to-date, project-specific context rather than relying on limited prompt memory. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    PaddleSpeech

    PaddleSpeech

    Easy-to-use Speech Toolkit including Self-Supervised Learning model

    ...Via the easy-to-use, efficient, flexible and scalable implementation, our vision is to empower both industrial application and academic research, including training, inference & testing modules, and deployment process. Low barriers to install, CLI, Server, and Streaming Server is available to quick-start your journey. We provide high-speed and ultra-lightweight models, and also cutting-edge technology. We provide production ready streaming asr and streaming tts system. Our frontend contains Text Normalization and Grapheme-to-Phoneme (G2P, including Polyphone and Tone Sandhi). Moreover, we use self-defined linguistic rules to adapt Chinese context.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    nanobot

    nanobot

    🐈 nanobot: The Ultra-Lightweight Clawdbot / OpenClaw

    nanobot is an ultra-lightweight personal AI assistant designed to deliver powerful agent capabilities without unnecessary complexity. Built in just ~4,000 lines of clean, readable code, it offers a minimalist alternative to heavyweight agent frameworks while retaining core intelligence and extensibility. nanobot is optimized for speed and efficiency, enabling fast startup times and low resource usage across environments. Its research-ready architecture makes it easy for developers to...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 6
    RealtimeSTT

    RealtimeSTT

    A robust, efficient, low-latency speech-to-text library

    RealtimeSTT is a Python-based realtime speech-to-text engine emphasizing low latency, wake-word detection, voice activity detection, and automatic speech segmentation. It provides asynchronous callbacks, nanosecond-precision timestamps, and CLI tools, suitable for building voice assistants, meeting transcribers, or live caption systems.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    OpenAI-Compatible Edge-TTS API

    OpenAI-Compatible Edge-TTS API

    Free, high-quality text-to-speech API endpoint to replace OpenAI

    ...Because it relies on Edge’s TTS, the audio generation itself is free, and the project essentially acts as a smart proxy that handles formatting and streaming. The server supports Server-Sent Events (SSE) for streaming audio, enabling low-latency playback in chat UIs and other interactive tools. A Docker image is provided for one-command deployment, and environment variables can be used to configure default voice, language, response format, authentication, and logging options.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    NVIDIA cuOpt

    NVIDIA cuOpt

    GPU accelerated decision optimization

    ...Built primarily in C++, cuOpt leverages NVIDIA GPUs to deliver near real-time solutions for optimization tasks involving millions of variables and constraints. The platform provides multiple interfaces, including C, Python, and server APIs, allowing developers to integrate optimization capabilities into applications and services. cuOpt is designed for high-performance environments and can be deployed across cloud, hybrid, or on-premise infrastructures. By combining GPU acceleration with scalable APIs, cuOpt enables organizations to solve large optimization challenges in logistics, operations research, and decision-making systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    LMCache

    LMCache

    Supercharge Your LLM with the Fastest KV Cache Layer

    ...Its design supports reuse beyond strict prefix matching and enables sharing across serving instances, improving efficiency under real multi-tenant traffic. The broader project includes examples, tests, a server component, and public posts describing cross-engine sharing and inter-GPU KV transfers. These capabilities aim to lower latency, cut GPU cycles, and stabilize performance for production workloads with overlapping prompts or retrieval-augmented contexts. The end result is a cache fabric for LLMs that complements engines rather than replacing them.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Fully Managed MySQL, PostgreSQL, and SQL Server Icon
    Fully Managed MySQL, PostgreSQL, and SQL Server

    Automatic backups, patching, replication, and failover. Focus on your app, not your database.

    Cloud SQL handles your database ops end to end, so you can focus on your app.
    Try Free
  • 10
    Serena

    Serena

    Agent toolkit providing semantic retrieval and editing capabilities

    Serena is a coding-focused agent toolkit that turns an LLM into a practical software-engineering agent with semantic retrieval and editing over real repositories. It operates as an MCP server (and other integrations), exposing IDE-like tools so agents can locate symbols, reason about code structure, make targeted edits, and validate changes. The toolkit is LLM-agnostic and framework-agnostic, positioning itself as a drop-in capability for different chat UIs, orchestrators, or custom agent stacks. It emphasizes symbol-level understanding rather than naive file-wide diffs, enabling more precise refactors and additions. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Nerve

    Nerve

    The Simple Agent Development Kit

    Nerve is a developer-friendly Agent Development Kit (ADK) that utilizes YAML and a CLI to define, run, orchestrate, and evaluate LLM-driven agents. It supports declarative setups, tool integration, workflow pipelines, and both MCP client and server roles. Nerve is a simple yet powerful Agent Development Kit (ADK) to build, run, evaluate, and orchestrate LLM-based agents using just YAML and a CLI. It’s designed for technical users who want programmable, auditable, and reproducible automation using large language models. Define agents using a clean YAML format: system prompt, task, tools, and variables — all in one file.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Binary Ninja MCP

    Binary Ninja MCP

    A Binary Ninja plugin, MCP server

    The Binary Ninja MCP is a plugin and bridge that integrates Binary Ninja with Large Language Model clients via the Model Context Protocol, enhancing reverse engineering workflows with AI assistance. ​
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Giskard

    Giskard

    Collaborative & Open-Source Quality Assurance for all AI models

    The testing framework dedicated to ML models, from tabular to LLMs. Giskard is an open-source testing framework dedicated to ML models, from tabular models to LLMs. Testing Machine Learning applications can be tedious. Since ML models depend on data, testing scenarios depend on the domain specificities and are often infinite. At Giskard, we believe that Machine Learning needs its own testing framework. Created by ML engineers for ML engineers, Giskard enables you to scan your model to find...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 14
    ChatTTS webUI & API

    ChatTTS webUI & API

    A simple native web interface that uses ChatTTS to synthesize text

    ChatTTS-ui is a local web interface and API wrapper around the ChatTTS speech synthesis system, designed to make advanced TTS models easy to use from a browser. It runs a small backend server (Python + Torch + ffmpeg) and exposes a simple webpage where you can type text, adjust parameters, and generate audio. The project supports Chinese, English, and mixed text with digits and control symbols, making it suitable for bilingual content and numerically heavy text like announcements or prompts. From version 0.96 onward, ffmpeg installation is required for deployment, and previous CSV/PT voice tables are no longer valid, so users instead work with updated “voice value” parameters. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 15
    FastAPI-MCP

    FastAPI-MCP

    Expose your FastAPI endpoints as Model Context Protocol (MCP) tools

    ...Rather than acting as a thin converter, it’s built as a native FastAPI extension that understands dependency injection, so you can reuse Depends() for authentication and authorization across your MCP tools. The server speaks directly to your app over its ASGI interface, avoiding extra HTTP hops between the MCP layer and your API, which reduces latency and simplifies deployment. A tiny bootstrap is enough to stand up an MCP server and, if desired, mount an HTTP transport for remote clients. The docs emphasize a FastAPI-first workflow: keep your schemas, reuse your middleware, and surface endpoints to agents without rewriting controllers. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Style-Bert-VITS2

    Style-Bert-VITS2

    Style-Bert-VITS2: Bert-VITS2 with more controllable voice styles

    Style-Bert-VITS2 is a text-to-speech system based on Bert-VITS2 that focuses on highly controllable voice styles and emotional expression. It takes the original Bert-VITS2 v2.1 and its Japanese-Extra variant and extends them so you can control emotion and speaking style with fine-grained intensity, not just choose a generic tone. The project targets both power users and beginners: Windows users without Git or Python can install and run it using bundled .bat scripts, while advanced users can...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 17
    Qwen-Agent

    Qwen-Agent

    Agent framework and applications built upon Qwen>=3.0

    ...It provides components for instruction following, tool usage (function calling), planning, memory, RAG (retrieval augmented generation), code interpreter, etc. It ships with example applications (Browser Assistant, Code Interpreter, Custom Assistant), supports GUI front-ends, backends, server setups. Agent workflow can maintain context / memory to perform multi-turn or more complex logic over time. It acts as the backend for Qwen Chat among other use cases. Built-in Code Interpreter tool that can execute code (locally) as part of agent workflows.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    mcp-use

    mcp-use

    A solution to build and deploy MCP agents and applications

    ...It enables connection to multiple MCP servers, each exposing specific tool capabilities like browsing, file operations, or specialized integrations, through a unified MCPClient. Developers can create custom agents (via MCPAgent) that dynamically select the most appropriate server for each task using configurable pipelines or a built-in server manager. It simplifies authentication, access control, audit logging, observability, sandboxed runtime environments, and deployment workflows, whether self-hosted or managed, making MCP development production-ready. With integrations for popular frameworks like LangChain (Python) and LangChain.js (TypeScript), mcp-use accelerates the creation of tool-enabled AI agents.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    MCP Timeplus

    MCP Timeplus

    Execute SQL queries and manage databases seamlessly with Timeplus

    An MCP server designed for integration with Timeplus, enabling real-time data streaming and analytics through natural language interactions. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    MCP Text Editor

    MCP Text Editor

    Provides line-oriented text file editing capabilities

    The MCP Text Editor Server provides line-oriented text file editing capabilities through a standardized API, optimized for integration with Large Language Models (LLMs). It enables efficient partial file access, minimizing token usage while ensuring safe concurrent editing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    stt

    stt

    Voice Recognition to Text Tool

    ...It leverages open-source speech models such as Faster-Whisper to recognize and transcribe human speech into plain text, structured JSON objects, or subtitle files with time codes, making it suitable for both personal and professional transcription tasks. The project is designed to be easy to deploy: you can run a local Python server that exposes an HTTP API for uploading audio/video files and retrieving transcriptions in different formats. It supports GPU acceleration if available, enabling faster processing on compatible hardware but still offers reliable performance on CPUs alone.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Arcade AI

    Arcade AI

    Arcade Tool Development Kit (TDK), Worker, Evals, and CLI

    ...Evaluation framework for testing tool performance. Test your MCP server's tools, resources, prompts, elicitation, and OAuth 2. MCPJam is compliant with the latest MCP specs. Connect to any MCP server. MCPJam inspector supports STDIO, SSE, and Streamable HTTP transports.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    FastMCP

    FastMCP

    The fast, Pythonic way to build Model Context Protocol servers

    FastMCP is a fast, Pythonic framework for building servers and clients using the Model Context Protocol (MCP). It abstracts away protocol complexity like serialization, validation, and error handling, letting developers focus entirely on their business logic. With simple decorators, you can expose Python functions as tools, resources, or prompts that AI agents can safely and efficiently use. FastMCP introduces clear abstractions—components, providers, and transforms—that make it easy to...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Agent Payments Protocol (AP2)

    Agent Payments Protocol (AP2)

    Building a Secure and Interoperable Future for AI-Driven Payments

    ...It includes “types” definitions (the core message and object schema) and example agent implementations to demonstrate the mechanics of agent-to-agent and agent-to-server interactions. The design emphasizes flexibility: although their samples use a particular Agent Development Kit (ADK) or runtime, the protocol is intended to be independent of those choices.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Appfl

    Appfl

    Advanced Privacy-Preserving Federated Learning framework

    APPFL (Advanced Privacy-Preserving Federated Learning) is a Python framework enabling researchers to easily build and benchmark privacy-aware federated learning solutions. It supports flexible algorithm development, differential privacy, secure communications, and runs efficiently on HPC and multi-GPU setups.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB