Showing 174 open source projects for "mac runtime"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    ONNX Runtime

    ONNX Runtime

    ONNX Runtime: cross-platform, high performance ML inferencing

    ONNX Runtime is a cross-platform inference and training machine-learning accelerator. ONNX Runtime inference can enable faster customer experiences and lower costs, supporting models from deep learning frameworks such as PyTorch and TensorFlow/Keras as well as classical machine learning libraries such as scikit-learn, LightGBM, XGBoost, etc. ONNX Runtime is compatible with different hardware, drivers, and operating systems, and provides optimal performance by leveraging hardware accelerators...
    Downloads: 79 This Week
    Last Update:
    See Project
  • 2
    IREE

    IREE

    A retargetable MLIR-based machine learning compiler runtime toolkit

    IREE (Intermediate Representation Execution Environment, pronounced as "eerie") is an MLIR-based end-to-end compiler and runtime that lowers Machine Learning (ML) models to a unified IR that scales up to meet the needs of the data center and down to satisfy the constraints and special considerations of mobile and edge deployments.
    Downloads: 67 This Week
    Last Update:
    See Project
  • 3
    Monoio

    Monoio

    Rust async runtime based on io-uring

    Monoio is a Rust asynchronous runtime designed for high-performance I/O-bound servers and applications, built around native OS async I/O primitives (e.g. io_uring on Linux, epoll / kqueue on other Unix-like systems), rather than layering atop an existing runtime. Its design philosophy centers on a “thread-per-core” model where each core runs its own event loop, minimizing cross-thread synchronization needs, avoiding the overhead and complexity of task scheduling, and letting developers write...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 4
    kokoro-onnx

    kokoro-onnx

    TTS with kokoro and onnx runtime

    kokoro-onnx is a text-to-speech toolkit that wraps the Kokoro neural TTS model in an easy-to-use ONNX Runtime interface, so you can generate speech from Python with minimal setup. It focuses on running efficiently on commodity hardware, including macOS with Apple Silicon, while still delivering near real-time performance for many use cases. The project ships prebuilt model files and a simple example script, so you can go from installation to producing an audio.wav file in just a few steps....
    Downloads: 95 This Week
    Last Update:
    See Project
  • Application Monitoring That Won't Slow Your App Down Icon
    Application Monitoring That Won't Slow Your App Down

    AppSignal's Rust-based agent is lightweight and stable. Already running in thousands of production apps.

    Full APM with errors, performance, logs, and uptime monitoring. 99.999% uptime SLA on the platform itself.
    Start Free
  • 5
    Osaurus

    Osaurus

    AI edge infrastructure for macOS. Run local or cloud models

    Osaurus is an open-source AI edge infrastructure platform designed specifically for macOS environments to run and manage AI models locally. The project provides a native runtime that allows applications to access large language models and AI tools directly on the user’s machine without relying entirely on cloud services. Osaurus supports running both local and remote models, enabling developers to build AI-powered applications that can operate offline or leverage external APIs when needed....
    Downloads: 4 This Week
    Last Update:
    See Project
  • 6
    AsmJit

    AsmJit

    Low-latency machine code generation

    AsmJit is a low-level code generation library designed for dynamically creating machine code at runtime, enabling just-in-time (JIT) compilation for performance-critical applications. It provides a high-level API that abstracts away the complexity of writing raw assembly while still allowing fine-grained control over instruction generation. The library supports multiple architectures, including x86 and x64, making it versatile for cross-platform development. It is commonly used in...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    LiteRT

    LiteRT

    LiteRT is the new name for TensorFlow Lite (TFLite)

    LiteRT is an experimental, real-time inference runtime built by Google AI Edge to run lightweight ML models on edge devices with ultra-low latency. It focuses on delivering predictable and consistent performance for models used in time-critical applications like robotics, AR/VR, and IoT. LiteRT is designed to be hardware-agnostic, with minimal dependencies and tight control over execution scheduling.
    Downloads: 17 This Week
    Last Update:
    See Project
  • 8
    Claw Code

    Claw Code

    AI agent harness for AI coding agents

    Claw Code is an open-source AI agent harness project focused on building better tools for orchestrating and managing autonomous coding agents. It originated as a clean-room reimplementation inspired by the architecture of Claude Code, aiming to replicate core concepts without using proprietary code. The project provides a Python-based foundation for experimenting with agent workflows, tool integration, and task execution pipelines. It emphasizes harness engineering—how agents are structured,...
    Downloads: 95 This Week
    Last Update:
    See Project
  • 9
    DeepGEMM

    DeepGEMM

    Clean and efficient FP8 GEMM kernels with fine-grained scaling

    DeepGEMM is a specialized CUDA library for efficient, high-performance general matrix multiplication (GEMM) operations, with particular focus on low-precision formats such as FP8 (and experimental support for BF16). The library is designed to work cleanly and simply, avoiding overly templated or heavily abstracted code, while still delivering performance that rivals expert-tuned libraries. It supports both standard and “grouped” GEMMs, which is useful for architectures like Mixture of...
    Downloads: 196 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    Docker Agent

    Docker Agent

    AI Agent Builder and Runtime by Docker Engineering

    Docker Agent is an open-source multi-agent runtime developed by Docker that enables developers to define, run, and orchestrate AI agents using simple declarative configuration files instead of traditional code-heavy approaches. It introduces a YAML-based configuration model where users describe agent behavior, tools, models, and interaction logic in a single file, significantly reducing complexity in building AI systems. The runtime supports multi-agent collaboration, allowing specialized...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 11
    OpenFang

    OpenFang

    Open-source Agent Operating System

    OpenFang is an open-source agent operating system designed to orchestrate autonomous AI agents and workflows in a structured, production-oriented environment. Written primarily in Rust, the project focuses on building a high-performance runtime where multiple specialized agents can collaborate to complete complex computational or development tasks. It aims to move beyond simple chat-based agents by providing infrastructure for persistent agent memory, task coordination, and scalable...
    Downloads: 51 This Week
    Last Update:
    See Project
  • 12
    SmythOS

    SmythOS

    Cloud-native runtime for agentic AI

    SmythOS SRE (Smyth Runtime Environment) is an open-source runtime and development platform designed for building and operating production-grade AI agents. It provides a foundational infrastructure layer that functions similarly to an operating system for agentic AI systems, managing resources such as language models, storage, vector databases, and caching through a unified interface. Developers can use the runtime to create, deploy, and orchestrate intelligent agents across local machines,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ByteHook

    ByteHook

    ByteHook is an Android PLT hook library

    ByteHook is a ByteDance-hosted project whose name suggests a hooking or instrumentation library, likely used for hooking system calls or API calls for monitoring, sandboxing or instrumentation. The repository appears to aim at low-level hooking/injection capabilities, perhaps to support runtime introspection, behavioral monitoring, or hooking-based instrumentation (e.g. for security, tracing, sandboxing, or debugging). Because hooking is a common technique for intercepting library or system...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    mlx

    mlx

    MLX: An array framework for Apple silicon

    MlX offers a local web interface to browse, download, and run ML models via Hugging Face or local sources. It supports searching by tags or tasks, visualization of model metadata, quick inference demos, automatic setup of runtime environments, and works with PyTorch, TensorFlow, and ONNX. Ideal for researchers exploring and testing models via browser.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 15
    ort

    ort

    Fast ML inference & training for ONNX models in Rust

    ort is a high-performance Rust library that provides bindings to ONNX Runtime, enabling developers to run machine learning inference and training workflows directly within Rust applications using the standardized ONNX model format. It is designed to bridge the gap between modern machine learning frameworks and systems programming by offering a safe, ergonomic API for executing models originally built in ecosystems like PyTorch, TensorFlow, or scikit-learn. The library emphasizes speed and...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 16
    Spice.ai OSS

    Spice.ai OSS

    A self-hostable CDN for databases

    Spice is a portable runtime offering developers a unified SQL interface to materialize, accelerate, and query data from any database, data warehouse, or data lake. Spice connects, fuses, and delivers data to applications, machine-learning models, and AI backends, functioning as an application-specific, tier-optimized Database CDN. The Spice runtime, written in Rust, is built-with industry-leading technologies such as Apache DataFusion, Apache Arrow, Apache Arrow Flight, SQLite, and DuckDB....
    Downloads: 28 This Week
    Last Update:
    See Project
  • 17
    Elkeid

    Elkeid

    Open source solution that can meet the requirements of workloads

    Elkeid is an open-source platform for security and intrusion-detection that aims to support a wide variety of deployment contexts — from bare-metal hosts to containers, Kubernetes clusters, and even serverless environments. It was born out of ByteDance’s internal security best practices, offering for community users a subset of its enterprise-grade capabilities. Elkeid combines kernel-level data collection, user-space agents, and runtime instrumentation (RASP) to detect malicious behavior,...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 18
    NullClaw

    NullClaw

    Fastest, smallest, and fully autonomous AI assistant infrastructure

    NullClaw is the smallest fully autonomous AI assistant infrastructure, built entirely in Zig as a single static binary with zero runtime dependencies. At just 678 KB with ~1 MB peak RAM usage, it boots in under 2 milliseconds and runs on virtually any hardware, including low-cost ARM boards. Despite its size, it delivers a complete AI stack with 22+ model providers, 18+ communication channels, integrated tools, hybrid memory, and sandboxed runtime support. Its architecture is fully modular,...
    Downloads: 21 This Week
    Last Update:
    See Project
  • 19
    OpenShell

    OpenShell

    OpenShell is the safe, private runtime for autonomous AI agents.

    OpenShell is an open-source runtime designed to safely run autonomous AI agents in isolated environments. Developed by NVIDIA, it provides sandboxed execution spaces that protect system resources, credentials, and data from unauthorized access. Each agent runs inside a containerized sandbox governed by declarative YAML security policies that control network access, file permissions, and process behavior. The platform includes a gateway service that manages sandbox lifecycles and routes AI...
    Downloads: 31 This Week
    Last Update:
    See Project
  • 20
    llama.cpp

    llama.cpp

    Port of Facebook's LLaMA model in C/C++

    The llama.cpp project enables the inference of Meta's LLaMA model (and other models) in pure C/C++ without requiring a Python runtime. It is designed for efficient and fast model execution, offering easy integration for applications needing LLM-based capabilities. The repository focuses on providing a highly optimized and portable implementation for running large language models directly within C/C++ environments.
    Downloads: 251 This Week
    Last Update:
    See Project
  • 21
    Torch-TensorRT

    Torch-TensorRT

    PyTorch/TorchScript/FX compiler for NVIDIA GPUs using TensorRT

    Torch-TensorRT is a compiler for PyTorch/TorchScript, targeting NVIDIA GPUs via NVIDIA’s TensorRT Deep Learning Optimizer and Runtime. Unlike PyTorch’s Just-In-Time (JIT) compiler, Torch-TensorRT is an Ahead-of-Time (AOT) compiler, meaning that before you deploy your TorchScript code, you go through an explicit compile step to convert a standard TorchScript program into a module targeting a TensorRT engine. Torch-TensorRT operates as a PyTorch extension and compiles modules that integrate...
    Downloads: 18 This Week
    Last Update:
    See Project
  • 22
    whisper.cpp

    whisper.cpp

    Port of OpenAI's Whisper model in C/C++

    whisper.cpp is a lightweight, C/C++ reimplementation of OpenAI’s Whisper automatic speech recognition (ASR) model—designed for efficient, standalone transcription without external dependencies. The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library. The command downloads the base.en model converted to custom ggml format and runs the inference on all .wav samples in the folder samples....
    Downloads: 360 This Week
    Last Update:
    See Project
  • 23
    langrocks

    langrocks

    Tools like web browser, computer access and code runner for LLMs

    Langrocks is a programming language experimentation toolkit that enables developers to create, test, and optimize custom programming languages.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Ralph AI Agent

    Ralph AI Agent

    AI agent loop that runs repeatedly until all PRD items are complete

    Ralph is a Rust-based AI agent runtime that focuses on safe, modular, and programmable autonomous behavior. It provides a reactive loop where agents can repeatedly assess the current context, reason about the next best action using large language models, and execute actions across integrated tools and services. The runtime emphasizes safety boundaries by sandboxing operations, enforcing time and token limits, and isolating execution layers to prevent unpredictable side effects. Ralph also...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Unity MCP

    Unity MCP

    AI-powered bridge connecting LLMs and advanced AI agents

    Unity-MCP is an open-source integration that connects artificial intelligence assistants with the Unity game development environment through the Model Context Protocol. The project enables AI tools such as coding assistants and autonomous agents to interact directly with Unity projects, allowing them to analyze scenes, modify assets, and generate code within the development environment. By exposing Unity editor functionality through MCP tools, the plugin allows external AI systems to...
    Downloads: 20 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • 5
  • Next
MongoDB Logo MongoDB