Showing 21 open source projects for "gpu max performance"

View related business solutions
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 1
    Gupax

    Gupax

    GUI Uniting P2Pool And XMRig

    A versatile tool designed for managing and analyzing GPU performance across tasks.
    Downloads: 31 This Week
    Last Update:
    See Project
  • 2
    nviwatch

    nviwatch

    A blazingly fast rust based TUI for managing and monitoring NVIDIA GPU

    NviWatch is an interactive terminal user interface (TUI) application for monitoring NVIDIA GPU devices and processes. Built with Rust, it provides real-time insights into GPU performance metrics, including temperature, utilization, memory usage, and power consumption.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Alacritty

    Alacritty

    A cross-platform, GPU-accelerated terminal emulator

    Alacritty is the fastest open source terminal emulator there is. How is it the fastest? With such a strong focus on simplicity and performance, Alacritty’s included features are very carefully considered, ensuring that it remains blazingly fast. It’s got a GPU for rendering that makes a whole lot of optimizations possible. In various benchmarked terminals, Alacritty has shown to be either faster, or way faster than others. Alacritty requires no additional setup, but still allows configuration of many aspects of the terminal. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    CubeCL

    CubeCL

    Multi-platform high-performance compute language extension for Rust

    CubeCL is a low-level compute language and compiler framework designed to simplify and optimize GPU programming for high-performance workloads, particularly in machine learning and numerical computing. It provides an abstraction layer that allows developers to write portable, hardware-efficient compute kernels without directly dealing with complex GPU APIs such as CUDA or OpenCL. CubeCL focuses on delivering predictable performance and composability by exposing explicit control over memory layouts, parallelism, and execution patterns while still maintaining a developer-friendly syntax. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • 5
    llmfit

    llmfit

    157 models, 30 providers, one command to find what runs on hardware

    llmfit is a terminal-based utility that helps developers determine which large language models can realistically run on their local hardware by analyzing system resources and model requirements. The tool automatically detects CPU, RAM, GPU, and VRAM specifications, then ranks available models based on performance factors such as speed, quality, and memory fit. It provides both an interactive terminal user interface and a traditional CLI mode, enabling flexible workflows for different user preferences. llmfit also supports advanced configurations including multi-GPU setups, mixture-of-experts architectures, and dynamic quantization recommendations. ...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 6
    uzu

    uzu

    A high-performance inference engine for AI models

    ...The engine implements a hybrid architecture in which model layers can be executed either as custom GPU kernels or through Apple’s MPSGraph API, allowing it to balance performance and compatibility depending on the workload. By utilizing Apple’s unified memory architecture, uzu reduces memory copying overhead and improves inference throughput for local AI workloads. The system includes a simple high-level API that enables developers to run models, create inference sessions, and generate outputs with minimal configuration.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    OpenFang

    OpenFang

    Open-source Agent Operating System

    OpenFang is an open-source agent operating system designed to orchestrate autonomous AI agents and workflows in a structured, production-oriented environment. Written primarily in Rust, the project focuses on building a high-performance runtime where multiple specialized agents can collaborate to complete complex computational or development tasks. It aims to move beyond simple chat-based agents by providing infrastructure for persistent agent memory, task coordination, and scalable execution. The system is positioned as a foundation for building advanced AI tooling, particularly in environments that require tight integration with GPU workflows and modern AI pipelines. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 8
    Zed

    Zed

    High-performance, multiplayer code editor from the creators of Atom

    Zed is a next-generation code editor designed for high-performance collaboration with humans and AI. Written from scratch in Rust to efficiently leverage multiple CPU cores and your GPU. Integrate upcoming LLMs into your workflow to generate, transform, and analyze code. Chat with teammates, write notes together, and share your screen and project. Multibuffers compose excerpts from across the codebase in one editable surface.
    Downloads: 27 This Week
    Last Update:
    See Project
  • 9
    three-d

    three-d

    Makes it simple to draw stuff across platforms (including web)

    three-d is a lightweight and modern 3D rendering library written in Rust that targets both native and WebAssembly environments, providing a simple yet powerful abstraction over GPU-based graphics APIs. It is designed to make 3D graphics programming accessible while still offering fine-grained control over rendering pipelines, materials, lighting, and camera systems. The library leverages modern graphics standards such as OpenGL and WebGL to deliver high-performance rendering across platforms, including browsers and desktop applications. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • 10
    Floem

    Floem

    A native Rust UI library with fine-grained reactivity

    Floem is a cross-platform GUI framework for Rust. It aims to be extremely performant while providing world-class developer ergonomics. Supporting both GPU and CPU rendering, Floem gives you performance that's closest to bare metal. Also primitives are provided to help the developer to write performant UI code without too much effect.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    shimmy

    shimmy

    Python-free Rust inference server

    ...This compatibility enables developers to replace remote AI services with locally hosted models while keeping their existing software architecture intact. Shimmy focuses on performance and simplicity, using efficient runtime components to minimize memory usage and startup time compared to heavier inference frameworks. It supports modern model formats such as GGUF and SafeTensors and can automatically discover models stored locally or in common directories used by other AI tools. Advanced capabilities include CPU offloading for Mixture-of-Experts models and GPU acceleration, enabling large models to run on consumer hardware with limited VRAM.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    ort

    ort

    Fast ML inference & training for ONNX models in Rust

    ort is a high-performance Rust library that provides bindings to ONNX Runtime, enabling developers to run machine learning inference and training workflows directly within Rust applications using the standardized ONNX model format. It is designed to bridge the gap between modern machine learning frameworks and systems programming by offering a safe, ergonomic API for executing models originally built in ecosystems like PyTorch, TensorFlow, or scikit-learn. The library emphasizes speed and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Bend

    Bend

    A massively parallel, high-level programming language

    Bend is an interactive programming environment (REPL) built on top of the Kotlin language, designed to allow users to explore, experiment, and learn Kotlin in a live, feedback-driven manner. The tool lets you define variables, functions, or values at the prompt and iteratively refine them—immediately seeing output and types—while preserving state across commands. It emphasizes discoverability and experimentation: users can inspect functions, call them on sample inputs, and evolve logic...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    MusicGPT

    MusicGPT

    Generate music based on natural language prompts using LLMs

    MusicGPT is an open-source application designed to generate music from natural language prompts using locally executed artificial intelligence models. The software allows users to run advanced music generation systems directly on their own devices without requiring heavy dependencies such as Python or full machine learning frameworks. Instead, it provides a lightweight environment capable of executing music generation models locally on CPUs or GPUs while maintaining strong performance across...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 15
    WezTerm

    WezTerm

    GPU-accelerated cross-platform terminal emulator

    WezTerm is a GPU-accelerated, cross-platform terminal emulator and multiplexer built in Rust by Wez Furlong. Designed for speed and modern hardware, WezTerm offers smooth rendering and excellent performance on Windows, macOS, and Linux. It supports multiple tabs and panes, configurable keybindings, and integrates well with shell environments. WezTerm is highly customizable and focuses on providing a powerful, flexible terminal experience for developers and system administrators. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 16
    NebiOS Alternative Mirror (Official)

    NebiOS Alternative Mirror (Official)

    A unique Ubuntu-based Linux OS with its own desktop and ecosystem.

    ...Main site: https://nebios.org 📌 This mirror page is updated less frequently than the main site. For the latest information, visit https://nebios.org ⚠️ Don't download from unknown sources. ⚠️ Virtual machines: NebiOS X requires GPU acceleration. Most hypervisors (VirtualBox, VMware, GNOME Boxes) will show degraded performance. QEMU with 3D acceleration enabled gives the best VM experience. See FAQ for details. System requirements and pre-installation info: https://nebios.org/download (select your version) Full FAQ: https://nebios.org/support/faq
    Leader badge
    Downloads: 195 This Week
    Last Update:
    See Project
  • 17
    pipeless

    pipeless

    A computer vision framework to create and deploy apps in minutes

    ...You can easily use industry-standard models, such as YOLO, or load your custom model in one of the supported inference runtimes. Pipeless ships some of the most popular inference runtimes, such as the ONNX Runtime, allowing you to run inference with high performance on CPU or GPU out-of-the-box. You can deploy your Pipeless application with a single command to edge and IoT devices or the cloud.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    SIG Rust

    SIG Rust

    Rust language bindings for TensorFlow

    SIG Rust provides idiomatic Rust bindings for TensorFlow, making it possible for developers to work with TensorFlow functionality from within the Rust programming language. Rather than replacing TensorFlow itself, it acts as an integration layer that connects Rust applications to the TensorFlow C API. The repository is designed for developers who want Rust’s performance, safety, and systems programming strengths while still accessing TensorFlow’s machine learning capabilities. It includes...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Ambient

    Ambient

    The multiplayer game engine

    Ambient is an open-source, cross-platform runtime and engine for building and deploying high-performance multiplayer games and 3D applications, using a modern stack built on Rust, WebAssembly (WASM), and WebGPU. It aims to make multiplayer game development accessible and flexible, providing an entity-component-system (ECS) at its core that doubles as a real-time in-game database; everything in the game — from world objects to runtime data — is represented as entities + components, which can...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    Forma

    Forma

    An efficient vector-graphics renderer

    Forma is an experimental vector graphics renderer written in Rust, developed by Google to explore high-performance, parallelized rendering techniques across multiple platforms. The project aims to achieve portability, performance, simplicity, and small footprint through a streamlined four-stage rendering pipeline. Forma provides both CPU (software) and GPU (hardware) backends, relying on Rust’s SIMD auto-vectorization, Rayon for multithreading, and WebGPU (wgpu) for hardware acceleration. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    Real-ESRGAN Video Enhance

    Real-ESRGAN Video Enhance

    Real-ESRGAN video upscaler with resumability

    ...It utilizes Real-ESRGAN-can-Vulkan, FFmpeg and MediaInfo under the hood. REVE employs a segment-based approach to video upscaling, allowing it to simultaneously upscale and encode videos. This results in a notable enhancement in performance and enables the feature of reusability. You can download Windows executable file for Intel/AMD/Nvidia GPU. This executable file is portable and includes all the binaries and models required. No CUDA or PyTorch environment is needed.
    Downloads: 16 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB