Showing 8 open source projects for "multicast streaming server"

View related business solutions
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    Desktop Commander MCP

    Desktop Commander MCP

    AI-powered MCP server for desktop file and terminal automation

    Desktop Commander MCP is an advanced Model Context Protocol server designed to extend AI assistants with direct control over a user’s local machine, including the file system and terminal. It integrates with clients like Claude Desktop to enable AI-driven workflows such as editing files, executing commands, and automating development tasks from a single conversational interface. Desktop Commander MCP builds on top of an MCP filesystem server and enhances it with powerful search, replace, and code editing capabilities tailored for real-world development environments. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    OpenAI Assistants Quickstart

    OpenAI Assistants Quickstart

    OpenAI Assistants API quickstart with Next.js

    openai-assistants-quickstart is a template for using the Assistants API in a Next.js app, demonstrating streaming, tool use, and function calling in one place. The repository includes multiple example pages that each showcase specific capabilities, while all examples share the same underlying assistant with all capabilities enabled. The primary chat logic lives in the Chat component at app/components/chat.tsx, which manages rendering, streaming, and forwarding function calls. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Open Claude Cowork

    Open Claude Cowork

    Open Source version of Claude Cowork built with Claude Code

    Open Claude Cowork is an open source desktop chat application that brings the power of autonomous, agent-style AI workflows to your local machine by combining the Claude Agent SDK with the Composio Tool Router, enabling developers and power users to build intelligent assistants that can interact with a vast array of external tools and services. It offers a native Electron-based interface for macOS, Windows, and Linux that feels familiar and modern, supporting persistent, multi-session...
    Downloads: 123 This Week
    Last Update:
    See Project
  • 4
    Deep Research

    Deep Research

    Use any LLMs (Large Language Models) for Deep Research

    Deep Research is a local-first research agent that orchestrates multiple LLMs to generate in-depth reports in minutes. It combines “thinking” and “task” model roles with live internet access to plan, search, read, and synthesize findings into structured outputs. The project emphasizes privacy: processing and storage happen locally, avoiding server-side retention of your queries and notes. A simple web UI lets you enter topics and configure models, while the backend streams progress as...
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 5
    Antigravity Claude Proxy

    Antigravity Claude Proxy

    Proxy that exposes Antigravity provided claude / gemini models

    ...This makes it easier to integrate Claude into existing toolchains, scripts, notebooks, or agent frameworks that do not have built-in support for Anthropic’s native SDKs. It abstracts away key differences like authentication choreography, request schema quirks, and streaming protocols so client code can remain unchanged when switching between models.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    LangServe

    LangServe

    Helps developers deploy LangChain runnables and chains as a REST API

    LangServe is an open-source deployment framework designed to expose LangChain applications as production-ready REST APIs. The tool simplifies the process of turning language-model pipelines, chains, and agents into web services that can be accessed by external applications. Instead of manually writing API endpoints, developers can use LangServe to automatically generate a server that exposes LangChain workflows through HTTP interfaces. The framework is built on top of FastAPI and uses...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Basaran

    Basaran

    Basaran, an open-source alternative to the OpenAI text completion API

    ...Support both decoder-only and encoder-decoder models. Detokenizer that handles surrogates and whitespace. Multi-GPU support with optional 8-bit quantization. Real-time partial progress using server-sent events. Compatible with OpenAI API and client libraries. Comes with a fancy web-based playground. Docker images are available on Docker Hub and GitHub Packages.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    ChatLLM Web

    ChatLLM Web

    Chat with LLM like Vicuna totally in your browser with WebGPU

    Chat with LLM like Vicuna totally in your browser with WebGPU, safely, privately, and with no server. Powered By web-llm. To use this app, you need a browser that supports WebGPU, such as Chrome 113 or Chrome Canary. Chrome versions ≤ 112 are not supported. You will need a GPU with about 6.4GB of memory. If your GPU has less memory, the app will still run, but the response time will be slower. The first time you use the app, you will need to download the model. For the Vicuna-7b model that...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB