Showing 189 open source projects for "telnet-server"

View related business solutions
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 1
    ChatGLM3

    ChatGLM3

    ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat

    ...It keeps the series’ smooth dialog and low deployment cost while adding native tool use (function calling), a built-in code interpreter, and agent-style workflows. The family includes base and long-context variants (8K/32K/128K). The repo ships Python APIs, CLI and web demos (Gradio/Streamlit), an OpenAI-format API server, and a compact fine-tuning kit. Quantization (4/8-bit), CPU/MPS support, and accelerator backends (TensorRT-LLM, OpenVINO, chatglm.cpp) enable lightweight local or edge deployment.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    Reader 3

    Reader 3

    Quick illustration of how one can easily read books together with LLMs

    This project is a minimalist, self-hosted EPUB reader designed to help users browse and read EPUB books one chapter at a time through a lightweight local server, making it especially easy to extract or work with chapters in external tools like large language models. It was created primarily as a simple demonstration of how to combine local book reading with LLM workflows without heavy dependencies or complicated setup, and it runs with just a small Python script and a basic HTTP server.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    GPT4Free

    GPT4Free

    The official gpt4free repository

    gpt4free is an open-source project offering free, unrestricted access to GPT‑4–style language models without requiring an API key. The repository includes scripts and server implementations designed to replicate OpenAI’s GPT‑4 API behavior by leveraging publicly available or self-hosted models. It’s licensed under GPL‑v3.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 4
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    Chroma MCP

    Chroma MCP

    A Model Context Protocol (MCP) server implementation

    Chroma MCP Server is an implementation of the Model Context Protocol (MCP) designed to integrate large language model (LLM) applications with external data sources or tools. It offers a standardized framework to seamlessly provide LLMs with the context they require for effective operation. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Kimi Code CLI

    Kimi Code CLI

    Kimi Code CLI is your next CLI agent

    ...The tool includes integration with Zsh so that users can activate AI assistance via a hotkey while staying within their favorite shell environment, and it can serve as an Agent Client Protocol (ACP) server to bridge AI functionality into compatible IDEs and editors. Its support for well-established MCP tool configuration conventions lets developers connect the CLI to external tools and services during workflows, expanding its capabilities beyond simple queries into orchestrated development tasks.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 7
    Nexa SDK

    Nexa SDK

    Nexa SDK is a comprehensive toolkit for supporting ONNX and GGML

    ...It supports text generation, image generation, vision-language models (VLM), and speech-to-text (ASR), and text-to-speech (TTS) capabilities. Additionally, it offers an OpenAI-compatible API server with JSON schema mode for function calling and streaming support, and a user-friendly Streamlit UI. Users can run Nexa SDK in any device with Python environment, and GPU acceleration is supported, including CUDA, Metal, and ROCm. An executable version is also available.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 8
    Legion MCP

    Legion MCP

    A server that helps people access and query data in databases

    The Legion MCP Server is designed to help users access and query data in databases using the Legion Query Runner, integrated with the Model Context Protocol (MCP) Python SDK. It facilitates efficient data retrieval and analysis through standardized interfaces. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Paperless-ngx

    Paperless-ngx

    A community-supported supercharged version of paperless

    Paperless-ngx is a community-supported open-source document management system that transforms your physical documents into a searchable online archive so you can keep, well, less paper.
    Downloads: 12 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 10
    Colab-MCP

    Colab-MCP

    An MCP server for interacting with Google Colab

    Colab-MCP is an open-source Model Context Protocol server developed by Google that enables AI agents to directly interact with and control Google Colab environments programmatically, transforming Colab into a fully automated, agent-accessible workspace. Instead of relying on manual notebook usage, the system allows MCP-compatible agents to execute code, manage files, install dependencies, and orchestrate entire development workflows within Colab’s cloud infrastructure.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Speech-AI-Forge

    Speech-AI-Forge

    Speech-AI-Forge is a project developed around TTS generation model

    Speech-AI-Forge is a full-stack project built around modern text-to-speech generation models, providing both an API server and a Gradio-based web UI for interactive use. At its core, it acts as a hub that wires together multiple speech-related capabilities, including TTS, speech-to-text and LLM-based control flows, behind a consistent interface. The system is designed to be deployed in several ways: you can try it online via hosted demos, spin it up in a one-click Colab environment, run it in Docker containers, or set it up locally with its environment preparation scripts. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 12
    OpenVoice

    OpenVoice

    Instant voice cloning by MIT and MyShell. Audio foundation model

    OpenVoice is a versatile instant voice cloning system that can replicate a speaker’s tone color from just a short audio clip and then generate speech in multiple languages. It is designed not only to match the timbre of the reference voice, but also to give granular control over style parameters such as emotion, accent, rhythm, pauses, and intonation. The model supports cross-lingual and even zero-shot cross-lingual voice cloning, so a speaker recorded in one language can be made to speak...
    Downloads: 17 This Week
    Last Update:
    See Project
  • 13
    LazyLLM

    LazyLLM

    Easiest and laziest way for building multi-agent LLMs applications

    LazyLLM is an optimized, lightweight LLM server designed for easy and fast deployment of large language models. It is fully compatible with the OpenAI API specification, enabling developers to integrate their own models into applications that normally rely on OpenAI’s endpoints. LazyLLM emphasizes low resource usage and fast inference while supporting multiple models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    TensorFlow

    TensorFlow

    TensorFlow is an open source library for machine learning

    Originally developed by Google for internal use, TensorFlow is an open source platform for machine learning. Available across all common operating systems (desktop, server and mobile), TensorFlow provides stable APIs for Python and C as well as APIs that are not guaranteed to be backwards compatible or are 3rd party for a variety of other languages. The platform can be easily deployed on multiple CPUs, GPUs and Google's proprietary chip, the tensor processing unit (TPU). TensorFlow expresses its computations as dataflow graphs, with each node in the graph representing an operation. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 15
    TensorRT LLM

    TensorRT LLM

    TensorRT LLM provides users with an easy-to-use Python API

    ...It includes support for cutting-edge inference methods like speculative decoding and inflight batching, enabling real-time and large-scale AI applications. TensorRT-LLM integrates seamlessly with NVIDIA’s broader inference ecosystem, including Triton Inference Server and distributed deployment frameworks, making it suitable for production environments.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Chandra

    Chandra

    OCR model for complex documents with layout-aware structured outputs

    ...It is capable of handling over 40 languages and is optimized to read difficult inputs such as messy handwriting and multi-column layouts. Chandra can be run locally using transformer-based inference or deployed with a high-performance server setup for large-scale processing. It also includes command-line tools and optional web-based interfaces to simplify interaction and batch processing workflows.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    nanobot

    nanobot

    🐈 nanobot: The Ultra-Lightweight Clawdbot / OpenClaw

    nanobot is an ultra-lightweight personal AI assistant designed to deliver powerful agent capabilities without unnecessary complexity. Built in just ~4,000 lines of clean, readable code, it offers a minimalist alternative to heavyweight agent frameworks while retaining core intelligence and extensibility. nanobot is optimized for speed and efficiency, enabling fast startup times and low resource usage across environments. Its research-ready architecture makes it easy for developers to...
    Downloads: 14 This Week
    Last Update:
    See Project
  • 18
    PaddleSpeech

    PaddleSpeech

    Easy-to-use Speech Toolkit including Self-Supervised Learning model

    ...Via the easy-to-use, efficient, flexible and scalable implementation, our vision is to empower both industrial application and academic research, including training, inference & testing modules, and deployment process. Low barriers to install, CLI, Server, and Streaming Server is available to quick-start your journey. We provide high-speed and ultra-lightweight models, and also cutting-edge technology. We provide production ready streaming asr and streaming tts system. Our frontend contains Text Normalization and Grapheme-to-Phoneme (G2P, including Polyphone and Tone Sandhi). Moreover, we use self-defined linguistic rules to adapt Chinese context.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    MCP UI

    MCP UI

    SDK for building interactive UI components over MCP for AI tools

    ...It includes both client-side and server-side SDKs, making it possible to define UI elements on the backend and handle user interactions on the frontend. It supports multiple programming environments, including TypeScript, Python, and Ruby, broadening its accessibility for developers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    MCP Text Editor

    MCP Text Editor

    Provides line-oriented text file editing capabilities

    The MCP Text Editor Server provides line-oriented text file editing capabilities through a standardized API, optimized for integration with Large Language Models (LLMs). It enables efficient partial file access, minimizing token usage while ensuring safe concurrent editing.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    FlowLens MCP

    FlowLens MCP

    Open-source MCP server that gives your coding agent

    ...The MCP server then loads this captured “flow” and exposes it to the AI agent via the Model Context Protocol (MCP), letting the agent examine, search, filter, and reason about the session just as a human developer would, without needing the agent to re-run the flow or rely on minimal reproduction data (logs, screenshots).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    LMCache

    LMCache

    Supercharge Your LLM with the Fastest KV Cache Layer

    ...Its design supports reuse beyond strict prefix matching and enables sharing across serving instances, improving efficiency under real multi-tenant traffic. The broader project includes examples, tests, a server component, and public posts describing cross-engine sharing and inter-GPU KV transfers. These capabilities aim to lower latency, cut GPU cycles, and stabilize performance for production workloads with overlapping prompts or retrieval-augmented contexts. The end result is a cache fabric for LLMs that complements engines rather than replacing them.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    ContextForge MCP Gateway

    ContextForge MCP Gateway

    A Model Context Protocol (MCP) Gateway & Registry

    MCP Context Forge is a feature-rich gateway and registry that federates Model Context Protocol (MCP) servers and traditional REST services behind a single, governed endpoint. It exposes an MCP-compliant interface to clients while handling discovery, authentication, rate limiting, retries, and observability on the server side. The gateway scales horizontally, supports multi-cluster deployments on Kubernetes, and uses Redis for federation and caching across instances. Operators can define virtual servers, wire multiple transports, and optionally enable an admin UI for management and monitoring. Packaged for quick starts via PyPI and Docker, it targets production reliability with health checks, metrics, and structured logs. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Nerve

    Nerve

    The Simple Agent Development Kit

    Nerve is a developer-friendly Agent Development Kit (ADK) that utilizes YAML and a CLI to define, run, orchestrate, and evaluate LLM-driven agents. It supports declarative setups, tool integration, workflow pipelines, and both MCP client and server roles. Nerve is a simple yet powerful Agent Development Kit (ADK) to build, run, evaluate, and orchestrate LLM-based agents using just YAML and a CLI. It’s designed for technical users who want programmable, auditable, and reproducible automation using large language models. Define agents using a clean YAML format: system prompt, task, tools, and variables — all in one file.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    ClearML

    ClearML

    Streamline your ML workflow

    ...The ClearML Python Package for integrating ClearML into your existing scripts by adding just two lines of code, and optionally extending your experiments and other workflows with ClearML powerful and versatile set of classes and methods. The ClearML Server storing experiment, model, and workflow data, and supports the Web UI experiment manager, and ML-Ops automation for reproducibility and tuning. It is available as a hosted service and open source for you to deploy your own ClearML Server. The ClearML Agent for ML-Ops orchestration, experiment and workflow reproducibility, and scalability.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB