8 projects for "mac runtime" with 2 filters applied:

  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 1
    MLX Engine

    MLX Engine

    LM Studio Apple MLX engine

    MLX Engine is the Apple MLX-based inference backend used by LM Studio to run large language models efficiently on Apple Silicon hardware. Built on top of the mlx-lm and mlx-vlm ecosystems, the engine provides a unified architecture capable of supporting both text-only and multimodal models. Its design focuses on high-performance on-device inference, leveraging Apple’s MLX stack to accelerate computation on M-series chips. The project introduces modular VisionAddOn components that allow image...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    NVIDIA FLARE

    NVIDIA FLARE

    NVIDIA Federated Learning Application Runtime Environment

    NVIDIA Federated Learning Application Runtime Environment NVIDIA FLARE is a domain-agnostic, open-source, extensible SDK that allows researchers and data scientists to adapt existing ML/DL workflows(PyTorch, TensorFlow, Scikit-learn, XGBoost etc.) to a federated paradigm. It enables platform developers to build a secure, privacy-preserving offering for a distributed multi-party collaboration. NVIDIA FLARE is built on a componentized architecture that allows you to take federated...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 3
    shimmy

    shimmy

    Python-free Rust inference server

    The shimmy project is a lightweight local inference server designed to run large language models with minimal overhead. Written primarily in Rust, the tool provides a small standalone binary that exposes an API compatible with the OpenAI interface, allowing existing applications to interact with local models without significant code changes. This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    rust-bert

    rust-bert

    Rust native ready-to-use NLP pipelines and transformer-based models

    rust-bert is a Rust-based implementation of transformer-based natural language processing models that provides ready-to-use pipelines for tasks such as text classification, summarization, and question answering. The project ports many capabilities of the Hugging Face Transformers ecosystem into the Rust programming language. It allows developers to run state-of-the-art NLP models like BERT, GPT-2, and DistilBERT directly within Rust applications while maintaining high performance and memory...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Let your crypto work for you

    Put idle assets to work with competitive interest rates, borrow without selling, and trade with precision. All in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 5
    KotlinDL

    KotlinDL

    High-level Deep Learning Framework written in Kotlin

    KotlinDL is a high-level Deep Learning API written in Kotlin and inspired by Keras. Under the hood, it uses TensorFlow Java API and ONNX Runtime API for Java. KotlinDL offers simple APIs for training deep learning models from scratch, importing existing Keras and ONNX models for inference, and leveraging transfer learning for tailoring existing pre-trained models to your tasks. This project aims to make Deep Learning easier for JVM and Android developers and simplify deploying deep learning...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    uTensor

    uTensor

    TinyML AI inference library

    uTensor is an embedded machine learning inference framework designed to run neural network models on resource-constrained devices such as microcontrollers and Internet-of-Things hardware. The project focuses on enabling TinyML deployments by translating trained machine learning models into efficient C++ code that can execute directly on embedded systems. Instead of training models on-device, the framework uses an offline workflow that converts trained TensorFlow graphs into optimized...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Weld

    Weld

    High-performance runtime for data analytics applications

    Weld is a programming language and runtime designed to improve the performance of data-intensive applications by optimizing computations across multiple libraries. Instead of optimizing individual functions independently, Weld introduces an intermediate representation that allows different frameworks to share optimization opportunities. This approach reduces data movement between libraries and enables the system to generate highly optimized machine code for parallel execution. Weld is...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    When to use TensorFlowSharp

    When to use TensorFlowSharp

    TensorFlow API for .NET languages

    When to use TensorFlowSharp is a .NET binding for the TensorFlow machine learning framework that allows developers to run TensorFlow models directly from C# and other .NET languages. The project exposes TensorFlow’s native C API through a strongly typed interface designed to integrate naturally with the .NET ecosystem. Its primary purpose is to enable developers working in Microsoft-based environments to load trained TensorFlow models and perform inference or additional training within .NET...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB