Showing 25 open source projects for "routing"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 1
    NVIDIA cuOpt

    NVIDIA cuOpt

    GPU accelerated decision optimization

    NVIDIA cuOpt is a GPU-accelerated optimization engine designed to solve complex mathematical optimization problems at large scale. It supports a range of optimization models including linear programming (LP), mixed integer linear programming (MILP), quadratic programming (QP), and vehicle routing problems (VRP). Built primarily in C++, cuOpt leverages NVIDIA GPUs to deliver near real-time solutions for optimization tasks involving millions of variables and constraints. The platform provides multiple interfaces, including C, Python, and server APIs, allowing developers to integrate optimization capabilities into applications and services. cuOpt is designed for high-performance environments and can be deployed across cloud, hybrid, or on-premise infrastructures. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 2
    OpenAI Forward

    OpenAI Forward

    An efficient forwarding service designed for LLMs

    ...Its main purpose is to make model access more manageable and efficient by adding operational controls such as request rate limiting, token rate limiting, caching, logging, routing, and key management around existing LLM endpoints. The project can proxy both local and cloud-hosted language model services, which makes it useful for teams that want a single control layer regardless of whether they are using something like LocalAI or a hosted provider compatible with OpenAI-style APIs. A major emphasis of the repository is asynchronous performance, using tools such as uvicorn, aiohttp, and asyncio to support high-throughput forwarding workloads.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    MoBA

    MoBA

    MoBA: Mixture of Block Attention for Long-Context LLMs

    ...Instead of forcing each token to attend to every other token in the sequence, MoBA divides the context into blocks and dynamically routes queries to only the most relevant segments of information. This routing strategy reduces the computational cost associated with traditional attention while preserving performance on reasoning and long-context tasks. The approach allows language models to scale to significantly longer input contexts without the quadratic computational cost normally associated with transformer attention mechanisms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Claude Cognitive

    Claude Cognitive

    Persistent context and multi-instance coordination

    ...It introduces an attention-based context router that prioritizes files and content relevant to the current development discussion — tagging them as HOT, WARM, or COLD based on recency and keyword activation — so Claude Code doesn’t waste token budget rereading irrelevant code. This context routing dramatically reduces redundant token usage and accelerates large codebase interactions by focusing only on what truly matters to the current task. Additionally, Claude-Cognitive includes a pool coordinator to share state across multiple Claude Code instances, preserving what’s been learned or completed and preventing repetitive debugging or redundant exploration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 5
    DreamO

    DreamO

    A Unified Framework for Image Customization

    ...Built on a diffusion-transformer (DiT) backbone, it supports a diverse set of tasks — including identity preservation, virtual “try-on” (e.g. clothing, accessories), style transfer, IP adaptation (objects/characters), and layout/condition-aware customizations — all handled within the same unified architecture. DreamO’s design introduces a feature routing constraint that helps disentangle different control conditions (like identity, style, clothing) when more than one is specified, which significantly reduces conflicts and artifacts when combining controls. It also uses a “placeholder strategy” to precisely align conditional inputs (e.g. where to place clothing or objects) in generated images, giving users fine-grained control over composition.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    ChatGPT Clone

    ChatGPT Clone

    ChatGPT interface with better UI

    ...The project is useful for prototyping assistants, documentation bots, and internal developer tools without committing to a specific vendor or UI framework. Configuration is kept simple so newcomers can get a working chat in minutes and then dial in features like authentication or multi-model routing. While it illustrates how to hook into third-party LLM endpoints, it is typically positioned as an educational, self-hosted starter that you should operate responsibly and within provider's terms of use.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 7
    DeepClaude

    DeepClaude

    Unleash Next-Level AI

    ...The system commonly pairs models such as DeepSeek R1 with Claude or Gemini, leveraging their complementary strengths to produce results that outperform individual models in benchmarks and real-world usage scenarios. DeepClaude is designed with compatibility in mind, supporting OpenAI-style APIs and allowing integration with various third-party model providers and routing services. It includes a graphical configuration interface and Docker-based deployment options, making it accessible to both developers and non-technical users who want to run advanced AI systems locally.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Every Code

    Every Code

    Local AI coding agent CLI with multi-agent orchestration tools

    Every Code (often referred to simply as Code) is a fast, local AI-powered coding agent designed to run directly in the terminal environment. It is a community-driven fork of the Codex CLI, with a strong emphasis on improving real-world developer ergonomics and workflows. Every Code enhances the traditional coding assistant model by introducing multi-agent orchestration, allowing multiple AI agents to collaborate, compare solutions, and refine outputs in parallel. It supports integration with...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 9
    AgentScope

    AgentScope

    Build and run agents you can see, understand and trust

    AgentScope is a production-ready agent framework designed to help developers build, deploy, and scale intelligent agentic applications. It provides essential abstractions that evolve with advancing LLM capabilities, emphasizing reasoning, tool use, and flexible orchestration rather than rigid prompt constraints. With built-in support for ReAct agents, memory, planning, human-in-the-loop control, and real-time voice interaction, developers can create powerful agents in minutes. AgentScope...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8 Monitoring Tools in One APM. Install in 5 Minutes. Icon
    8 Monitoring Tools in One APM. Install in 5 Minutes.

    Errors, performance, logs, uptime, hosts, anomalies, dashboards, and check-ins. One interface.

    AppSignal works out of the box for Ruby, Elixir, Node.js, Python, and more. 30-day free trial, no credit card required.
    Start Free
  • 10
    Pruna AI

    Pruna AI

    Pruna is a model optimization framework built for developers

    Pruna is an open-source, self-hostable AI inference engine designed to help teams deploy and manage large language models (LLMs) efficiently across private or hybrid infrastructures. Built with performance and developer ergonomics in mind, Pruna simplifies inference workflows by enabling multi-model orchestration, autoscaling, GPU resource allocation, and compatibility with popular open-source models. It is ideal for companies or teams looking to reduce reliance on external APIs while...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Lingua-Py

    Lingua-Py

    The most accurate natural language detection library for Python

    ...This is very useful as a preprocessing step for linguistic data in natural language processing applications such as text classification and spell checking. Other use cases, for instance, might include routing e-mails to the right geographically located customer service department, based on the e-mails' languages. Language detection is often done as part of large machine learning frameworks or natural language processing applications. In cases where you don't need the full-fledged functionality of those systems or don't want to learn the ropes of those, a small flexible library comes in handy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    KServe

    KServe

    Standardized Serverless ML Inference Platform on Kubernetes

    KServe provides a Kubernetes Custom Resource Definition for serving machine learning (ML) models on arbitrary frameworks. It aims to solve production model serving use cases by providing performant, high abstraction interfaces for common ML frameworks like Tensorflow, XGBoost, ScikitLearn, PyTorch, and ONNX. It encapsulates the complexity of autoscaling, networking, health checking, and server configuration to bring cutting edge serving features like GPU Autoscaling, Scale to Zero, and...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Parallax

    Parallax

    Parallax is a distributed model serving framework

    Parallax is a decentralized inference framework designed to run large language models across distributed computing resources. Instead of relying on centralized GPU clusters in data centers, the system allows multiple heterogeneous machines to collaborate in serving AI inference workloads. Parallax divides model layers across different nodes and dynamically coordinates them to form a complete inference pipeline. A two-stage scheduling architecture determines how model layers are allocated to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    ZAPI

    ZAPI

    ZAPI by Adopt AI is an open-source Python library

    ZAPI is a developer-centric API framework that streamlines building, testing, and deploying APIs with strong type safety and minimal boilerplate, helping teams deliver backend services faster with fewer errors. It emphasizes a declarative router and schema model that uses types to define request and response formats, providing clear contracts for frontend and backend teams while automatically generating documentation. Zapi abstracts many repetitive tasks such as validation, authentication...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    SkillForge

    SkillForge

    Ultimate meta-skill for generating best-in-class Claude Code skills

    SkillForge is a systematic methodology and tooling framework for creating high-quality AI “skills” specifically optimized for Claude Code integrations, treating skill creation as an engineering discipline rather than an ad-hoc art form. It introduces a multi-phase architecture where every input or request is triaged intelligently, analyzed deeply through structured lenses, specified formally, synthesized with automated generation, and finally subjected to multi-agent review before...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Semantic Router

    Semantic Router

    Superfast AI decision making and processing of multi-modal data

    Semantic Router is a superfast decision-making layer for your LLMs and agents. Rather than waiting for slow, unreliable LLM generations to make tool-use or safety decisions, we use the magic of semantic vector space — routing our requests using semantic meaning. Combining LLMs with deterministic rules means we can be confident that our AI systems behave as intended. Cramming agent tools into the limited context window is expensive, slow, and fundamentally limited. Semantic Router enables lightning-fast and cheap tool usage that can scale to many thousands of tools. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Cactus

    Cactus

    Low-latency AI inference engine optimized for mobile devices

    Cactus is a low-latency, energy-efficient AI inference framework designed specifically for mobile devices and wearables, enabling advanced machine learning capabilities directly on-device. It provides a full-stack architecture composed of an inference engine, a computation graph system, and highly optimized hardware kernels tailored for ARM-based processors. Cactus emphasizes efficient memory usage through techniques such as zero-copy computation graphs and quantized model formats, allowing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    ContextForge MCP Gateway

    ContextForge MCP Gateway

    A Model Context Protocol (MCP) Gateway & Registry

    MCP Context Forge is a feature-rich gateway and registry that federates Model Context Protocol (MCP) servers and traditional REST services behind a single, governed endpoint. It exposes an MCP-compliant interface to clients while handling discovery, authentication, rate limiting, retries, and observability on the server side. The gateway scales horizontally, supports multi-cluster deployments on Kubernetes, and uses Redis for federation and caching across instances. Operators can define...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    MiniMax-01

    MiniMax-01

    Large-language-model & vision-language-model based on Linear Attention

    MiniMax-01 is the official repository for two flagship models: MiniMax-Text-01, a long-context language model, and MiniMax-VL-01, a vision-language model built on top of it. MiniMax-Text-01 uses a hybrid attention architecture that blends Lightning Attention, standard softmax attention, and Mixture-of-Experts (MoE) routing to achieve both high throughput and long-context reasoning. It has 456 billion total parameters with 45.9 billion activated per token and is trained with advanced parallel strategies such as LASP+, varlen ring attention, and Expert Tensor Parallelism, enabling a training context of 1 million tokens and up to 4 million tokens at inference. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Warlock-Studio

    Warlock-Studio

    AI Suite for upscaling, interpolating & restoring images/videos

    v6.0. Warlock-Studio is a Windows application that uses Real-ESRGAN, BSRGAN, IRCNN, GFPGAN, RealESRNet, RealESRAnime and RIFE Artificial Intelligence models to upscale, restore faces, interpolate frames and reduce noise in images and videos. the application supports GPU acceleration (including multi-GPU setups) and offers batch processing for large workloads. It includes drag-and-drop handling for single or multiple files, optional pre-resize functions, and an automatic tiling system...
    Leader badge
    Downloads: 20 This Week
    Last Update:
    See Project
  • 21
    LLaMA-MoE

    LLaMA-MoE

    Building Mixture-of-Experts from LLaMA with Continual Pre-training

    LLaMA-MoE is an open-source project that builds mixture-of-experts language models from LLaMA through expert partitioning and continual pre-training. The repository is centered on making MoE research more accessible by offering smaller and more affordable models with only about 3.0 to 3.5 billion activated parameters, which helps reduce deployment and experimentation costs. Its architecture works by splitting LLaMA feed-forward networks into sparse experts and adding gating mechanisms so...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    e-Dokyumento

    e-Dokyumento

    e-Dokyumento is web-based Document Management System (DMS)

    e-Dokyumento is opensource web-based Document Management System (DMS) A Document Management which automates the basic office document workflow such as receiving, filing, routing, and approving through capturing (scanning), digitizing (OCR Reading), storing, tagging, and electronically routing and approving (e-signature) of electronic documents. # Demo : https://e-dokyumento.herokuapp.com/ https://edokyu.seillig.com/ (refer to Readme.md for the accounts) #Dockerhub: https://hub.docker.com/r/nelsonmaligro/edokyumento # Install using the ISO: 1. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 23
    CapsGNN

    CapsGNN

    A PyTorch implementation of "Capsule Graph Neural Network"

    ...Inspired by the Capsule Neural Network (CapsNet), we propose the Capsule Graph Neural Network (CapsGNN), which adopts the concept of capsules to address the weakness in existing GNN-based graph embeddings algorithms. By extracting node features in the form of capsules, routing mechanism can be utilized to capture important information at the graph level. As a result, our model generates multiple embeddings for each graph to capture graph properties from different aspects.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Dynamic Routing Between Capsules

    Dynamic Routing Between Capsules

    A PyTorch implementation of the NIPS 2017 paper

    Dynamic Routing Between Capsules is a PyTorch implementation of the Capsule Network architecture originally proposed to address limitations in traditional convolutional neural networks. Capsule networks aim to improve how neural models represent spatial hierarchies and relationships between objects within images. Instead of scalar neuron activations, capsules output vectors that encode both the presence of features and their spatial properties such as orientation or pose.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Open Exchange (OpEx)

    Open Exchange (OpEx)

    The open source Algorithmic Trading System

    OpEx is an application suite that includes the main building blocks of commercial electronic trading systems. All OpEx applications run on distributed system architectures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB