Showing 30 open source projects for "mac runtime"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 1
    Osaurus

    Osaurus

    AI edge infrastructure for macOS. Run local or cloud models

    Osaurus is an open-source AI edge infrastructure platform designed specifically for macOS environments to run and manage AI models locally. The project provides a native runtime that allows applications to access large language models and AI tools directly on the user’s machine without relying entirely on cloud services. Osaurus supports running both local and remote models, enabling developers to build AI-powered applications that can operate offline or leverage external APIs when needed....
    Downloads: 6 This Week
    Last Update:
    See Project
  • 2
    SmythOS

    SmythOS

    Cloud-native runtime for agentic AI

    SmythOS SRE (Smyth Runtime Environment) is an open-source runtime and development platform designed for building and operating production-grade AI agents. It provides a foundational infrastructure layer that functions similarly to an operating system for agentic AI systems, managing resources such as language models, storage, vector databases, and caching through a unified interface. Developers can use the runtime to create, deploy, and orchestrate intelligent agents across local machines,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    llama.cpp

    llama.cpp

    Port of Facebook's LLaMA model in C/C++

    The llama.cpp project enables the inference of Meta's LLaMA model (and other models) in pure C/C++ without requiring a Python runtime. It is designed for efficient and fast model execution, offering easy integration for applications needing LLM-based capabilities. The repository focuses on providing a highly optimized and portable implementation for running large language models directly within C/C++ environments.
    Downloads: 198 This Week
    Last Update:
    See Project
  • 4
    langrocks

    langrocks

    Tools like web browser, computer access and code runner for LLMs

    Langrocks is a programming language experimentation toolkit that enables developers to create, test, and optimize custom programming languages.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • 5
    Unity MCP

    Unity MCP

    AI-powered bridge connecting LLMs and advanced AI agents

    Unity-MCP is an open-source integration that connects artificial intelligence assistants with the Unity game development environment through the Model Context Protocol. The project enables AI tools such as coding assistants and autonomous agents to interact directly with Unity projects, allowing them to analyze scenes, modify assets, and generate code within the development environment. By exposing Unity editor functionality through MCP tools, the plugin allows external AI systems to...
    Downloads: 13 This Week
    Last Update:
    See Project
  • 6
    chatd

    chatd

    Chat with your documents using local AI

    chatd is an open-source desktop application that allows users to interact with their documents through a locally running large language model. The software focuses on privacy and security by ensuring that all document processing and inference occur entirely on the user’s computer without sending data to external cloud services. It includes a built-in integration with the Ollama runtime, which provides a cross-platform environment for running large language models locally. The application...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 7
    Clippy

    Clippy

    Clippy, now with some AI

    Clippy is an open-source desktop assistant that allows users to run modern large language models locally while presenting them through a nostalgic interface inspired by Microsoft’s classic Clippy assistant from the 1990s. The project serves as both a playful homage to the early days of personal computing and a practical demonstration of local AI inference. Clippy integrates with the llama.cpp runtime to run models directly on a user’s computer without requiring cloud-based AI services. It...
    Downloads: 39 This Week
    Last Update:
    See Project
  • 8
    OneFileLLM

    OneFileLLM

    Specify a github or local repo, github pull request

    OneFileLLM is an open-source project designed to simplify the distribution and execution of large language model applications by packaging them into a single portable file. The concept behind the project is to eliminate the complexity normally associated with deploying AI systems, which often require multiple dependencies, frameworks, and configuration steps. Instead, the entire runtime environment, model interface, and application logic are bundled together into a single executable...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    MLC LLM

    MLC LLM

    Universal LLM Deployment Engine with ML Compilation

    MLC LLM is a machine learning compiler and deployment framework designed to enable efficient execution of large language models across a wide range of hardware platforms. The project focuses on compiling models into optimized runtimes that can run natively on devices such as GPUs, mobile processors, browsers, and edge hardware. By leveraging machine learning compilation techniques, mlc-llm produces high-performance inference engines that maintain consistent APIs across platforms. The system...
    Downloads: 30 This Week
    Last Update:
    See Project
  • AI-generated apps that pass security review Icon
    AI-generated apps that pass security review

    Stop waiting on engineering. Build production-ready internal tools with AI—on your company data, in your cloud.

    Retool lets you generate dashboards, admin panels, and workflows directly on your data. Type something like “Build me a revenue dashboard on my Stripe data” and get a working app with security, permissions, and compliance built in from day one. Whether on our cloud or self-hosted, create the internal software your team needs without compromising enterprise standards or control.
    Try Retool free
  • 10
    Vercel AI SDK

    Vercel AI SDK

    Build AI-powered applications with React, Svelte, Vue, and Solid

    The Vercel AI SDK is a library for building AI-powered streaming text and chat UIs.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 11
    OpenLLM

    OpenLLM

    Operating LLMs in production

    An open platform for operating large language models (LLMs) in production. Fine-tune, serve, deploy, and monitor any LLMs with ease. With OpenLLM, you can run inference with any open-source large-language models, deploy to the cloud or on-premises, and build powerful AI apps. Built-in supports a wide range of open-source LLMs and model runtime, including Llama 2, StableLM, Falcon, Dolly, Flan-T5, ChatGLM, StarCoder, and more. Serve LLMs over RESTful API or gRPC with one command, query via...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 12
    Prompt Poet

    Prompt Poet

    Streamlines and simplifies prompt design for both developers

    Prompt Poet is an open-source framework designed to simplify the creation, organization, and maintenance of prompts for large language model applications. The project focuses on transforming prompt engineering into a structured design process rather than ad-hoc string manipulation within application code. It allows developers and non-technical users to build prompts using templated configurations based on YAML and Jinja2, which makes prompts easier to compose, reuse, and modify across...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    PicoLM

    PicoLM

    Run a 1-billion parameter LLM on a $10 board with 256MB RAM

    PicoLM is an open-source inference framework designed to run large language models on extremely constrained hardware environments such as inexpensive single-board computers and embedded systems. The project focuses on enabling efficient local inference by optimizing memory usage, computation, and system dependencies so that relatively large models can operate on devices with minimal RAM. It is written primarily in C and designed with a minimalist architecture that removes unnecessary...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    llama.vscode

    llama.vscode

    VS Code extension for LLM-assisted code/text completion

    llama.vscode is a Visual Studio Code extension that provides AI-assisted coding features powered primarily by locally running language models. The extension is designed to be lightweight and efficient, enabling developers to use AI tools even on consumer-grade hardware. It integrates with the llama.cpp runtime to run language models locally, eliminating the need to rely entirely on external APIs or cloud providers. The extension supports common AI development features such as code...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Emscripten

    Emscripten

    Emscripten: An LLVM-to-WebAssembly Compiler

    Emscripten is a complete open-source compiler toolchain that transforms C, C++, and other LLVM-based source code into WebAssembly (and JavaScript), enabling native‑like applications to run in web browsers, Node.js, and other Wasm environments. While Emscripten mostly focuses on compiling C and C++ using Clang, it can be integrated with other LLVM-using compilers (for example, Rust has Emscripten integration, with the wasm32-unknown-emscripten and asmjs-unknown-emscripten targets). Emscripten...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 16
    OM1

    OM1

    Modular AI runtime for robots

    OM1 is an open-source AI platform designed to build autonomous agents capable of interacting with digital environments and completing complex tasks. The project focuses on creating a modular architecture where language models can coordinate with external tools, APIs, and knowledge sources to accomplish multi-step objectives. Instead of operating as simple conversational systems, OM1 agents can plan actions, retrieve information, and execute tasks across different services. The framework...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 17
    Lobe Icons

    Lobe Icons

    Brings AI/LLM brand logos to your React & React Native apps

    Lobe Icons is an open-source icon library designed to provide developers with a comprehensive collection of logos and visual assets representing popular artificial intelligence platforms, language models, and related technologies. The project focuses on making it easy for developers to include recognizable AI brand icons in applications such as dashboards, AI tools, documentation sites, or developer portals. The library includes icons for a wide range of AI providers and models, allowing...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 18
    SWIFT LLM

    SWIFT LLM

    Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs

    SWIFT LLM is a comprehensive framework developed within the ModelScope ecosystem for training, fine-tuning, evaluating, and deploying large language models and multimodal models. The platform provides a full machine learning pipeline that supports tasks ranging from model pre-training to reinforcement learning alignment techniques. It integrates with popular inference engines such as vLLM and LMDeploy to accelerate deployment and runtime performance. The framework also includes support for...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 19
    tlm

    tlm

    Local CLI Copilot, powered by Ollama

    tlm is an open-source command-line AI assistant designed to provide intelligent terminal support using locally running large language models. The project functions as a CLI copilot that helps developers generate commands, explain shell instructions, and answer technical questions directly from the terminal. Instead of relying on cloud APIs or paid AI services, TLM runs entirely on the user’s workstation and integrates with local models managed through the Ollama runtime. This approach allows...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 20
    Ollamac

    Ollamac

    Mac app for Ollama

    Ollamac is an open-source native macOS application that provides a graphical interface for interacting with local large language models running through the Ollama inference framework. The project was created to simplify the process of using local AI models, which typically require command-line interaction, by offering a clean and intuitive desktop interface. Through this interface, users can run and chat with a variety of LLM models installed through Ollama directly on their own machines....
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    Ludwig AI

    Ludwig AI

    Low-code framework for building custom LLMs, neural networks

    Declarative deep learning framework built for scale and efficiency. Ludwig is a low-code framework for building custom AI models like LLMs and other deep neural networks. Declarative YAML configuration file is all you need to train a state-of-the-art LLM on your data. Support for multi-task and multi-modality learning. Comprehensive config validation detects invalid parameter combinations and prevents runtime failures. Automatic batch size selection, distributed training (DDP, DeepSpeed),...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 22
    SimpleLLM

    SimpleLLM

    950 line, minimal, extensible LLM inference engine built from scratch

    SimpleLLM is a minimal, extensible large language model inference engine implemented in roughly 950 lines of code, built from scratch to serve both as a learning tool and a research platform for novel inference techniques. It provides the core components of an LLM runtime—such as tokenization, batching, and asynchronous execution—without the abstraction overhead of more complex engines, making it easier for developers and researchers to understand and modify. Designed to run efficiently on...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 23
    OpenDAN

    OpenDAN

    OpenDAN is an open source Personal AI OS

    OpenDAN is an open-source Personal AI OS , that consolidates various AI modules in one place for your personal use. The goal of OpenDAN (Open and Do Anything Now with AI) is to create a Personal AI OS , which provides a runtime environment for various Al modules as well as protocols for interoperability between them. With OpenDAN, users can securely collaborate with various AI modules using their private data to create powerful personal AI agents, such as butlers, lawyers, doctors, teachers,...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser. The WebLLM package itself does not come...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    mllm

    mllm

    Fast Multimodal LLM on Mobile Devices

    mllm is an open-source inference engine designed to run multimodal large language models efficiently on mobile devices and edge computing environments. The framework focuses on delivering high-performance AI inference in resource-constrained systems such as smartphones, embedded hardware, and lightweight computing platforms. Implemented primarily in C and C++, it is designed to operate with minimal external dependencies while taking advantage of hardware-specific acceleration technologies...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • Next
MongoDB Logo MongoDB