FastRouter
FastRouter is a unified API gateway that enables AI applications to access many large language, image, and audio models (like GPT-5, Claude 4 Opus, Gemini 2.5 Pro, Grok 4, etc.) through a single OpenAI-compatible endpoint. It features automatic routing, which dynamically picks the optimal model per request based on factors like cost, latency, and output quality. It supports massive scale (no imposed QPS limits) and ensures high availability via instant failover across model providers. FastRouter also includes cost control and governance tools to set budgets, rate limits, and model permissions per API key or project, and it delivers real-time analytics on token usage, request counts, and spending trends. The integration process is minimal; you simply swap your OpenAI base URL to FastRouterâs endpoint and configure preferences in the dashboard; the routing, optimization, and failover functions then run transparently.
Learn more
VibeKit
VibeKit is a simple, open source SDK for safely running Codex and Claude Code agents in secure, customizable sandboxes. It enables you to embed coding agents directly in your app or workflow via a dropâin SDK. import VibeKit and VibeKitConfig, and call generateCode with prompts, modes, and streaming callbacks for live output handling. VibeKit runs code in fully isolated private sandboxes, supports customizable environments where you can install packages, and is modelâagnostic, letting you choose any compatible Codex or Claude model. It streams agent output efficiently, maintains full prompt and code history, provides async run handling, integrates with GitHub for commits, branches, and pull requests, and supports telemetry and tracing (via OpenTelemetry). Compatible sandbox providers include E2B (today), with Daytona, Modal, Fly.io, and others coming soon, plus support for any runtime that meets your security needs.
Learn more
Edgee
Edgee is an AI gateway that sits between your application and large language model providers, acting as an edge intelligence layer that compresses prompts before they reach the model to reduce token usage, lower costs, and improve latency without changing your existing code. Applications call Edgee through a single OpenAI-compatible API, and Edgee applies edge-level policies such as intelligent token compression, routing, privacy controls, retries, caching, and cost governance before forwarding requests to the selected provider, including OpenAI, Anthropic, Gemini, xAI, and Mistral. Its token compression engine removes redundant input tokens while preserving semantic intent and context, achieving up to 50% input token reduction, which is especially valuable for long contexts, RAG pipelines, and multi-turn agents. Edgee enables tagging requests with custom metadata to track usage and spending by feature, team, project, or environment, and provides cost alerts when spending spikes.
Learn more
Compress-GLB
Compress-GLB is a simple web app designed for one task: reducing the size of GLB or GLTF 3D models so they load faster without causing strain on browsers or mobile GPUs. The tool enables up to a 90% size reduction, preserving model quality. It's powered by the open-source gltf-transform library for compression, optimizing texture (KTX2/Basis), mesh (Draco), and geometry. Perfect for game developers, web designers, and 3D artists. New users receive 5 free credits. Additional credits are available for a pay-as-you-go model. It offers a simple interface: drag, drop, choose compression levels, and proceed.
Learn more