Showing 17 open source projects for "proxy server windows"

View related business solutions
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    More flexibility. More control.

    Generate interest, access liquidity without selling, and execute trades seamlessly. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 1
    rtk

    rtk

    CLI proxy that reduces LLM token consumption

    rtk is an open-source command-line proxy designed to optimize interactions between AI coding agents and the terminal by reducing unnecessary token consumption. When AI assistants execute shell commands during software development tasks, the resulting terminal output often contains large amounts of repetitive or irrelevant information that can overwhelm the model’s context window. RTK intercepts these command outputs and compresses them into concise summaries before sending them to the...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 2
    Arch Gateway

    Arch Gateway

    The AI-native (edge and LLM) proxy for agents

    Arch is an AI-native proxy designed to facilitate the development of agentic applications by handling complex tasks such as input clarification, agent routing, and seamless integration of prompts with tools for common tasks. It provides unified access and observability of Large Language Models (LLMs), enabling developers to build applications more efficiently. Arch supports both edge and LLM deployments, offering flexibility in various environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Rust Docs MCP Server

    Rust Docs MCP Server

    Prevents outdated Rust code suggestions from AI assistants

    The Rust Docs MCP Server fetches documentation for specified Rust crates, generates embeddings for the content, and provides an MCP tool to answer questions about the crate based on the documentation context. ​
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Plano

    Plano

    Delivery infrastructure for agentic apps

    Plano is an AI-native proxy and data plane designed to simplify the infrastructure required to deploy and operate agentic applications in production environments. It removes repetitive plumbing work from application code by centralizing capabilities such as agent routing, orchestration, guardrails, observability, and model selection. Built on modern proxy technology and compatible with any language or AI framework, Plano enables developers to focus on core agent logic instead of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Cake

    Cake

    Distributed LLM and StableDiffusion inference

    Cake is a compact, powerful toolkit that combines a flexible TCP/UDP proxy, port forwarding system, and connection manager designed for both development and penetration testing scenarios. It enables users to create complex networking flows where traffic can be proxied, relayed, and manipulated between endpoints — useful for debugging networked applications, inspecting protocols, or tunneling traffic through different hops. The tool is designed to work with multiple protocols and supports...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 6
    Zed

    Zed

    High-performance, multiplayer code editor from the creators of Atom

    Zed is a next-generation code editor designed for high-performance collaboration with humans and AI. Written from scratch in Rust to efficiently leverage multiple CPU cores and your GPU. Integrate upcoming LLMs into your workflow to generate, transform, and analyze code. Chat with teammates, write notes together, and share your screen and project. Multibuffers compose excerpts from across the codebase in one editable surface. Evaluate code inline via Jupyter runtimes and collaboratively edit...
    Downloads: 27 This Week
    Last Update:
    See Project
  • 7
    MagicAPI AI Gateway

    MagicAPI AI Gateway

    Built for demanding AI workflows

    The world's fastest AI Gateway proxy, written in Rust and optimized for maximum performance. This high-performance API gateway routes requests to various AI providers (OpenAI, GROQ) with streaming support, making it perfect for developers who need reliable and blazing-fast AI API access.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    agentgateway

    agentgateway

    Next Generation Agentic Proxy for AI Agents and MCP servers

    Agentgateway is an open-source “data plane” built specifically for agentic AI connectivity, focusing on how agents talk to other agents and to tools across different frameworks and environments. It presents itself as a complete connectivity solution that adds drop-in security, observability, and governance to agent-to-agent and agent-to-tool communication without requiring you to rebuild your agent stack. The project supports interoperable protocols designed for this ecosystem, including...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 9
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • 10
    mistral.rs

    mistral.rs

    Fast, flexible LLM inference

    mistral.rs is a fast and flexible LLM inference engine implemented in Rust, designed to run and serve modern language models with an emphasis on performance and practical deployment. It provides multiple entry points for developers, including a CLI for running models locally and an HTTP server that exposes an OpenAI-compatible API surface for easy integration with existing clients. The project includes hardware-aware tooling that can benchmark a system and choose sensible quantization and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems. It provides an API interface that allows developers to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Google Workspace CLI

    Google Workspace CLI

    Command-line tool for Drive, Gmail, Calendar, Sheets, Docs, Chat, etc.

    Google Workspace CLI (gws) is a command-line tool designed to interact with Google Workspace services such as Drive, Gmail, Calendar, Sheets, and more from a single interface. It dynamically generates its command structure using Google’s Discovery Service, allowing it to automatically support new API endpoints as they become available. The tool eliminates the need for manual REST API calls by providing structured commands and built-in help for each resource and method. It outputs structured...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 13
    ort

    ort

    Fast ML inference & training for ONNX models in Rust

    ort is a high-performance Rust library that provides bindings to ONNX Runtime, enabling developers to run machine learning inference and training workflows directly within Rust applications using the standardized ONNX model format. It is designed to bridge the gap between modern machine learning frameworks and systems programming by offering a safe, ergonomic API for executing models originally built in ecosystems like PyTorch, TensorFlow, or scikit-learn. The library emphasizes speed and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Monoio

    Monoio

    Rust async runtime based on io-uring

    Monoio is a Rust asynchronous runtime designed for high-performance I/O-bound servers and applications, built around native OS async I/O primitives (e.g. io_uring on Linux, epoll / kqueue on other Unix-like systems), rather than layering atop an existing runtime. Its design philosophy centers on a “thread-per-core” model where each core runs its own event loop, minimizing cross-thread synchronization needs, avoiding the overhead and complexity of task scheduling, and letting developers write...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Phantasm

    Phantasm

    Toolkits to create a human-in-the-loop approval layer

    Phantasm offers toolkits to create a human-in-the-loop approval layer to monitor and guide AI agents' workflows in real-time, ensuring safety and reliability in AI operations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Code2Prompt

    Code2Prompt

    Convert codebases into structured prompts optimized for LLM analysis

    code2prompt is an open source command line tool designed to convert an entire codebase into a structured prompt that can be easily used with large language models. It analyzes a project directory, gathers relevant source files, and formats them into a single prompt that includes the source tree and code content. This approach helps developers quickly provide full project context to AI models without manually copying files or assembling prompts. code2prompt is built in Rust and focuses on...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Extractous

    Extractous

    Fast and efficient unstructured data extraction

    Extractous is a Rust-based unstructured data extraction library focused on fast local parsing of documents and other content-heavy files. Its purpose is to extract text and metadata efficiently from formats such as PDF, Word, HTML, email archives, images, and more, without depending on external APIs or separate parsing servers. The project emphasizes performance and low memory usage, and its maintainers describe it as a local-first alternative to heavier extraction stacks. For broader format...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB