Showing 9 open source projects for "gpu hardware"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 1
    CubeCL

    CubeCL

    Multi-platform high-performance compute language extension for Rust

    CubeCL is a low-level compute language and compiler framework designed to simplify and optimize GPU programming for high-performance workloads, particularly in machine learning and numerical computing. It provides an abstraction layer that allows developers to write portable, hardware-efficient compute kernels without directly dealing with complex GPU APIs such as CUDA or OpenCL. CubeCL focuses on delivering predictable performance and composability by exposing explicit control over memory layouts, parallelism, and execution patterns while still maintaining a developer-friendly syntax. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 2
    llmfit

    llmfit

    157 models, 30 providers, one command to find what runs on hardware

    llmfit is a terminal-based utility that helps developers determine which large language models can realistically run on their local hardware by analyzing system resources and model requirements. The tool automatically detects CPU, RAM, GPU, and VRAM specifications, then ranks available models based on performance factors such as speed, quality, and memory fit. It provides both an interactive terminal user interface and a traditional CLI mode, enabling flexible workflows for different user preferences. llmfit also supports advanced configurations including multi-GPU setups, mixture-of-experts architectures, and dynamic quantization recommendations. ...
    Downloads: 53 This Week
    Last Update:
    See Project
  • 3
    Bend

    Bend

    A massively parallel, high-level programming language

    Bend is an interactive programming environment (REPL) built on top of the Kotlin language, designed to allow users to explore, experiment, and learn Kotlin in a live, feedback-driven manner. The tool lets you define variables, functions, or values at the prompt and iteratively refine them—immediately seeing output and types—while preserving state across commands. It emphasizes discoverability and experimentation: users can inspect functions, call them on sample inputs, and evolve logic...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    ort

    ort

    Fast ML inference & training for ONNX models in Rust

    ort is a high-performance Rust library that provides bindings to ONNX Runtime, enabling developers to run machine learning inference and training workflows directly within Rust applications using the standardized ONNX model format. It is designed to bridge the gap between modern machine learning frameworks and systems programming by offering a safe, ergonomic API for executing models originally built in ecosystems like PyTorch, TensorFlow, or scikit-learn. The library emphasizes speed and...
    Downloads: 6 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    uzu

    uzu

    A high-performance inference engine for AI models

    uzu is a high-performance inference engine designed to run artificial intelligence models efficiently on Apple Silicon hardware. Written primarily in Rust and leveraging Apple’s Metal framework, the project focuses on maximizing performance when executing large language models and other AI workloads on devices such as Mac computers with M-series chips. The engine implements a hybrid architecture in which model layers can be executed either as custom GPU kernels or through Apple’s MPSGraph API, allowing it to balance performance and compatibility depending on the workload. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    shimmy

    shimmy

    Python-free Rust inference server

    ...It supports modern model formats such as GGUF and SafeTensors and can automatically discover models stored locally or in common directories used by other AI tools. Advanced capabilities include CPU offloading for Mixture-of-Experts models and GPU acceleration, enabling large models to run on consumer hardware with limited VRAM.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    WezTerm

    WezTerm

    GPU-accelerated cross-platform terminal emulator

    WezTerm is a GPU-accelerated, cross-platform terminal emulator and multiplexer built in Rust by Wez Furlong. Designed for speed and modern hardware, WezTerm offers smooth rendering and excellent performance on Windows, macOS, and Linux. It supports multiple tabs and panes, configurable keybindings, and integrates well with shell environments. WezTerm is highly customizable and focuses on providing a powerful, flexible terminal experience for developers and system administrators. ...
    Downloads: 24 This Week
    Last Update:
    See Project
  • 8
    NebiOS Alternative Mirror (Official)

    NebiOS Alternative Mirror (Official)

    An Ubuntu Linux-based OS that aims to end user.

    ...🛡️ Privacy: NebiOS has no telemetry, no background data services, no tracking daemons, and cannot send any personal data without user action. Full policy: nebisoftware.com/legal/privacy.html === 🚀 NebiOS X out now! ⚠️ Note on Virtual Machines: NebiOS X is optimized and designed for real hardware performance and native GPU acceleration. Virtualization platforms (VMware, VirtualBox, GNOME Boxes, etc.) may show degraded graphics or system performance. For accurate testing, please install NebiOS X on physical hardware. === System Requirements and informations about before installation can be found at https://nebios.org/download (select your version for it)
    Leader badge
    Downloads: 35 This Week
    Last Update:
    See Project
  • 9
    Forma

    Forma

    An efficient vector-graphics renderer

    ...The project aims to achieve portability, performance, simplicity, and small footprint through a streamlined four-stage rendering pipeline. Forma provides both CPU (software) and GPU (hardware) backends, relying on Rust’s SIMD auto-vectorization, Rayon for multithreading, and WebGPU (wgpu) for hardware acceleration. The renderer processes Bézier curves, line segments, and pixels through stages of flattening, rasterization, sorting, and painting, updating only changed tiles for efficiency. This design allows Forma to render complex vector scenes—such as large-scale SVGs—at interactive frame rates even on CPUs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Train ML Models With SQL You Already Know Icon
    Train ML Models With SQL You Already Know

    BigQuery automates data prep, analysis, and predictions with built-in AI assistance.

    Build and deploy ML models using familiar SQL. Automate data prep with built-in Gemini. Query 1 TB and store 10 GB free monthly.
    Try Free
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB