Showing 7 open source projects for "binary"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 1
    rtk

    rtk

    CLI proxy that reduces LLM token consumption

    ...By minimizing the amount of noise sent to the AI model, the tool improves reasoning quality and allows longer development sessions within the same context window. The system is implemented as a lightweight Rust binary that runs locally and integrates easily with common AI coding environments.
    Downloads: 39 This Week
    Last Update:
    See Project
  • 2
    yek

    yek

    Serialize repositories into LLM-ready context w/ smart prioritization

    Yek is a Rust-based CLI tool designed to serialize text-based files from a repository or directory into a single structured output for large language model use. It scans projects using .gitignore rules to exclude irrelevant files and automatically filters out binary or oversized content. Yek prioritizes files based on Git history, placing more important content later in the output to align with how language models process context. Yek supports multiple directories, individual files, and glob patterns, making it flexible for different workflows. It can stream output when piped or save results to a temporary file, depending on usage. ...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 3
    Moltis

    Moltis

    A Rust-native claw you can trust

    Moltis is an open-source personal AI assistant platform written in Rust that is designed to run as a fully self-hosted, local-first agent environment. It compiles the entire assistant stack, including the web interface, model routing, memory, and tools, into a single self-contained binary with no external runtime dependencies. The system supports multiple large language model providers alongside local models, enabling users to maintain privacy while still accessing cloud capabilities when needed. Moltis emphasizes security through sandboxed execution environments, where commands and browsing tasks run in isolated containers and require explicit approval. ...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 4
    Crabtalk

    Crabtalk

    Agents daemon that hides nothing

    Crabtalk is a composable AI agent runtime designed to provide a minimal yet powerful foundation for building and orchestrating intelligent agents within a single lightweight binary. It is implemented in Rust and focuses on delivering high performance, reliability, and low overhead compared to more complex agent frameworks. The system is built around a small set of core primitives, including skills, memory, context isolation, and extensions, which together enable flexible and modular agent behavior. CrabTalk emphasizes simplicity by avoiding unnecessary abstractions, allowing developers to maintain full control over how agents operate and interact with their environment. ...
    Downloads: 5 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture intact. Shimmy focuses on performance and simplicity, using efficient runtime components to minimize memory usage and startup time compared to heavier inference frameworks. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    ZeroClaw

    ZeroClaw

    Fast, small, and fully autonomous AI assistant infrastructure

    ...It is designed around a trait-based architecture so that model providers, communication channels, memory systems, and tooling integrations can be swapped or extended without rewriting core components, giving engineers flexibility and long-term maintainability. The framework features a compact single binary with fast cold and warm startup times and very low memory overhead, making it suitable even for resource-constrained hardware like small servers or edge devices. Security is a first-class concern, with sandbox controls, encrypted secrets, allowlisted operations, and scoped filesystem access by default, helping reduce risk when running autonomous agents.
    Downloads: 51 This Week
    Last Update:
    See Project
  • 7
    ModelFox

    ModelFox

    ModelFox makes it easy to train, deploy, and monitor ML models

    ...Learn about your models and monitor them in production from your browser. ModelFox makes it easy to train, deploy, and monitor machine learning models. You can install the modelfox CLI by either downloading the binary from the latest GitHub release or by building from source. Train a machine learning model by running modelfox train with the path to a CSV file and the name of the column you want to predict. The CLI automatically transforms your data into features, trains a number of linear and gradient boosted decision tree models to predict the target column, and writes the best model to a .modelfox file. ...
    Downloads: 52 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB