Showing 174 open source projects for "mac runtime"

View related business solutions
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • Forever Free Full-Stack Observability | Grafana Cloud Icon
    Forever Free Full-Stack Observability | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    Cog

    Cog

    Package and deploy machine learning models using Docker containers

    Cog is an open source tool designed to package machine learning models into standardized, production-ready containers. It simplifies the process of deploying models by automatically generating Docker images based on a simple configuration file, eliminating the need to manually write complex Dockerfiles. Developers can define the runtime environment, dependencies, and Python versions required for their models, allowing Cog to build a consistent container environment that follows best...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    Dafthunk

    Dafthunk

    A workflow execution platform built on top of the fantastic Cloudflare

    Dafthunk is a browser-first visual workflow automation platform that lets you build, run, and monitor workflows without standing up a traditional server stack. It’s designed around durable, multi-step execution so workflows can keep going (and recover) even when individual steps fail, time out, or need retries. The platform is closely aligned with Cloudflare’s ecosystem, using edge-native building blocks for execution, orchestration, and storage so workflows can run near users with low...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser. The WebLLM package itself does not come...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    Ludwig AI

    Ludwig AI

    Low-code framework for building custom LLMs, neural networks

    Declarative deep learning framework built for scale and efficiency. Ludwig is a low-code framework for building custom AI models like LLMs and other deep neural networks. Declarative YAML configuration file is all you need to train a state-of-the-art LLM on your data. Support for multi-task and multi-modality learning. Comprehensive config validation detects invalid parameter combinations and prevents runtime failures. Automatic batch size selection, distributed training (DDP, DeepSpeed),...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 5
    E2B

    E2B

    Secure open source cloud runtime for AI apps & AI agents

    E2B's Code Interpreter SDK allows you to add code-interpreting capabilities to your AI apps. E2B Sandbox is a secure sandboxed cloud environment made for AI agents and AI apps. Sandboxes allow AI agents and apps to have long-running cloud secure environments. In these environments, large language models can use the same tools as humans do.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    oneDNN

    oneDNN

    oneAPI Deep Neural Network Library (oneDNN)

    This software was previously known as Intel(R) Math Kernel Library for Deep Neural Networks (Intel(R) MKL-DNN) and Deep Neural Network Library (DNNL). oneAPI Deep Neural Network Library (oneDNN) is an open-source cross-platform performance library of basic building blocks for deep learning applications. oneDNN is part of oneAPI. The library is optimized for Intel(R) Architecture Processors, Intel Processor Graphics and Xe Architecture graphics. oneDNN has experimental support for the...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 7
    Poco Claw

    Poco Claw

    A more beautiful and easier-to-use alternative to OpenClaw

    Poco Claw is an AI agent platform designed as a more user-friendly and visually polished alternative to traditional OpenClaw implementations. It focuses on improving usability by providing a modern web interface combined with enhanced interaction capabilities such as built-in messaging and project organization tools. The system operates on a sandboxed runtime, ensuring that tasks executed by the agent are isolated from the host environment, which improves security and reliability. It extends...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    tlm

    tlm

    Local CLI Copilot, powered by Ollama

    tlm is an open-source command-line AI assistant designed to provide intelligent terminal support using locally running large language models. The project functions as a CLI copilot that helps developers generate commands, explain shell instructions, and answer technical questions directly from the terminal. Instead of relying on cloud APIs or paid AI services, TLM runs entirely on the user’s workstation and integrates with local models managed through the Ollama runtime. This approach allows...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    MemOS

    MemOS

    AI memory OS for LLM and Agent systems

    MemOS is an experimental operating system and runtime built around the concept of memory-centric computing, where memory objects are first-class citizens and program execution is organized around efficient, persistent memory access rather than traditional process and file system boundaries. The project explores rethinking system abstractions by tightly coupling computation with memory objects so that programs can operate on large datasets without expensive serialization or context switching....
    Downloads: 1 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    Z80-μLM

    Z80-μLM

    Z80-μLM is a 2-bit quantized language model

    Z80-μLM is a retro-computing AI project that demonstrates a tiny language model (Z80-μLM) engineered to run on an 8-bit Z80 CPU by aggressively quantizing weights down to 2-bit precision. The repository provides a complete workflow where you train or fine-tune conversational models in Python, then export them into a format that can be executed on classic Z80 systems. A key deliverable is producing CP/M-compatible .COM binaries, enabling a genuinely vintage “chat with your computer”...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    Solon

    Solon

    Java enterprise application development framework

    Solon is a full-scenario Java enterprise application framework that positions itself as a lean, high-performance alternative to heavy stacks. It advertises large concurrency gains, lower memory use, much faster startup, and dramatically smaller packages while remaining compatible from Java 8 through Java 24. The framework focuses on restrained APIs and an open ecosystem, with modules that cover web, data, cloud, and microservice patterns. Its messaging emphasizes “replaceable Spring”...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    tf2onnx

    tf2onnx

    Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX

    tf2onnx converts TensorFlow (tf-1.x or tf-2.x), keras, tensorflow.js and tflite models to ONNX via command line or python API. Note: tensorflow.js support was just added. While we tested it with many tfjs models from tfhub, it should be considered experimental. TensorFlow has many more ops than ONNX and occasionally mapping a model to ONNX creates issues. tf2onnx will use the ONNX version installed on your system and installs the latest ONNX version if none is found. We support and test ONNX...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Latent Box

    Latent Box

    A collection of awesome-lists for AI, creativity and art. AI

    Latent Box is an open-source platform focused on managing, deploying, and interacting with local or self-hosted AI models through a unified interface that simplifies experimentation and usage. It is designed to act as a control layer for running generative models, particularly those related to text and multimodal outputs, while maintaining full ownership of data and infrastructure. The platform emphasizes usability by providing a clean user interface that allows users to load models,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    what-to-eat

    what-to-eat

    An AI-based intelligent recipe generation platform

    what-to-eat is an AI-powered recipe generation platform that helps users decide what to cook by generating customized recipes based on ingredients, cuisine preferences, and dietary considerations. The application combines modern frontend technologies with large language models to create a fully interactive cooking assistant that can produce detailed recipes, nutritional analysis, and even visual representations of dishes. It supports a wide range of cuisines, including traditional Chinese...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    NVIDIA NeMo Agent Toolkit

    NVIDIA NeMo Agent Toolkit

    Library for efficiently connecting and optimizing teams of AI agents

    NVIDIA NeMo Agent Toolkit is an open-source framework designed to build, optimize, and manage AI agents across different development ecosystems. It provides enterprise-grade tools for improving agent performance, reliability, and observability throughout the development lifecycle. The toolkit integrates with popular agent frameworks such as LangChain, LlamaIndex, CrewAI, Microsoft Semantic Kernel, and Google ADK. Developers can monitor agent execution, trace workflows, and analyze...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 16
    OpenClaw Opik Observability Plugin

    OpenClaw Opik Observability Plugin

    Official plugin for OpenClaw that exports agent traces to Opik

    OpenClaw Opik Observability Plugin is an open-source plugin designed to add observability and monitoring capabilities to OpenClaw autonomous AI agents by exporting operational traces to the Opik observability platform. The project integrates directly with OpenClaw’s plugin architecture so that developers can capture detailed runtime information about how their agents behave while executing tasks. Each time an AI agent performs an action—such as calling a large language model, invoking a...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    Moltis

    Moltis

    A Rust-native claw you can trust

    Moltis is an open-source personal AI assistant platform written in Rust that is designed to run as a fully self-hosted, local-first agent environment. It compiles the entire assistant stack, including the web interface, model routing, memory, and tools, into a single self-contained binary with no external runtime dependencies. The system supports multiple large language model providers alongside local models, enabling users to maintain privacy while still accessing cloud capabilities when...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 18
    draw-a-ui

    draw-a-ui

    Draw wireframe sketches and generate HTML with AI vision models

    draw-a-ui is an experimental open source application that converts hand-drawn interface wireframes into working HTML code using artificial intelligence. draw-a-ui combines the tldraw canvas drawing tool with a vision-capable language model to interpret user-created mockups and translate them into a single HTML layout styled with Tailwind CSS. When a user sketches a UI on the canvas, the application captures the current drawing as SVG, converts it into a PNG image, and sends that image to a...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    mllm

    mllm

    Fast Multimodal LLM on Mobile Devices

    mllm is an open-source inference engine designed to run multimodal large language models efficiently on mobile devices and edge computing environments. The framework focuses on delivering high-performance AI inference in resource-constrained systems such as smartphones, embedded hardware, and lightweight computing platforms. Implemented primarily in C and C++, it is designed to operate with minimal external dependencies while taking advantage of hardware-specific acceleration technologies...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Coze Loop

    Coze Loop

    Next-generation AI Agent Optimization Platform

    Coze Loop is a developer-oriented platform that provides full lifecycle management for AI agents, covering everything from prompt engineering to production monitoring. The project aims to simplify the increasingly complex workflow of building reliable AI agents by offering integrated tools for debugging, evaluation, observability, and optimization. Through its visual playground, developers can test prompts interactively and compare outputs across different language models. The platform also...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Supertonic

    Supertonic

    Lightning-fast, on-device TTS, running natively via ONNX

    Supertonic is a lightning-fast, on-device text-to-speech system built around ONNX Runtime for maximum speed and portability. It focuses on running entirely locally, eliminating the need for cloud APIs and providing low latency and strong privacy guarantees, even on constrained devices like Raspberry Pi boards and e-readers. The core model is highly compact at around 66 million parameters, yet benchmarks show it can generate speech up to 167× faster than real time on modern consumer hardware...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 22
    Cloudflare Agents

    Cloudflare Agents

    Build and deploy AI Agents on Cloudflare

    Cloudflare Agents is an open-source framework designed to help developers build, deploy, and manage AI agents that run at the network edge. It provides infrastructure for creating stateful, event-driven agents capable of real-time interaction while maintaining low latency through Cloudflare’s distributed platform. The project includes SDKs, templates, and deployment tooling that simplify the process of connecting agents to external APIs, storage systems, and workflows. Its architecture...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    AutoAgent

    AutoAgent

    AutoAgent: Fully-Automated and Zero-Code LLM Agent Framework

    AutoAgent is a fully automated, zero-code LLM agent framework that lets users create agents and workflows using natural language instead of manual coding and configuration. It is structured around modes that cover both “use” and “build” scenarios: a user mode for running a ready-made multi-agent research assistant, plus editors for creating individual agents or multi-agent workflows from conversational requirements. The framework emphasizes self-managing workflow generation, where it can...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Matcha-TTS

    Matcha-TTS

    A fast TTS architecture with conditional flow matching

    Matcha-TTS is a non-autoregressive neural text-to-speech architecture that uses conditional flow matching to generate speech quickly while maintaining natural quality. It models speech as an ODE-based generative process, and conditional flow matching lets it reach high-quality audio in only a few synthesis steps, which greatly reduces latency compared to score-matching diffusion approaches. The model is fully probabilistic, so it can generate diverse realizations of the same text while still...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    SageMaker Training Toolkit

    SageMaker Training Toolkit

    Train machine learning models within Docker containers

    Train machine learning models within a Docker container using Amazon SageMaker. Amazon SageMaker is a fully managed service for data science and machine learning (ML) workflows. You can use Amazon SageMaker to simplify the process of building, training, and deploying ML models. To train a model, you can include your training script and dependencies in a Docker container that runs your training code. A container provides an effectively isolated environment, ensuring a consistent runtime and...
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB