Showing 10 open source projects for "gpu max performance"

View related business solutions
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 1
    AGI (Android GPU Inspector)

    AGI (Android GPU Inspector)

    Android GPU Inspector

    Android GPU Inspector (AGI) is a desktop tool for profiling, tracing, and debugging graphics workloads running on Android devices. It helps developers analyze Vulkan and OpenGL ES applications at the system, frame, and draw-call levels to uncover GPU and CPU bottlenecks. AGI captures detailed performance counters, timelines, and pipeline state to reveal stalls, overdraw, shader hotspots, and inefficient resource usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    SwiftShader

    SwiftShader

    SwiftShader is a high-performance CPU-based implementation

    SwiftShader is Google’s high-performance CPU-based implementation of the Vulkan 1.3 graphics API, designed to provide a hardware-independent rendering solution for 3D graphics. Unlike traditional GPU drivers, SwiftShader executes graphics commands entirely on the CPU, making it ideal for environments where dedicated graphics hardware is unavailable or unsuitable. It acts as a drop-in replacement for Vulkan drivers, allowing existing applications to run seamlessly by redirecting API calls through its software-based rendering engine. ...
    Downloads: 139 This Week
    Last Update:
    See Project
  • 3
    Beta9

    Beta9

    Run serverless GPU workloads with fast cold starts on bare-metal

    beta9 is a platform that enables running serverless GPU workloads with fast cold starts on bare-metal servers globally. It allows developers to deploy and scale GPU-accelerated applications without managing underlying infrastructure, offering flexibility and efficiency for AI and high-performance computing tasks. beta9 supports various frameworks and provides tools for monitoring and managing deployments effectively.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    ffmpeg-over-ip

    ffmpeg-over-ip

    Connect to remote ffmpeg servers

    ffmpeg-over-ip is a client-server system that enables remote execution of FFmpeg commands on a machine with GPU access while controlling it from another environment such as a container or virtual machine. It allows applications without direct GPU access to offload video transcoding tasks to a remote server, improving performance without requiring complex passthrough setups. The system works by coordinating commands through a lightweight protocol while using a shared filesystem to exchange media data. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • Our Free Plans just got better! | Auth0 Icon
    Our Free Plans just got better! | Auth0

    With up to 25k MAUs and unlimited Okta connections, our Free Plan lets you focus on what you do best—building great apps.

    You asked, we delivered! Auth0 is excited to expand our Free and Paid plans to include more options so you can focus on building, deploying, and scaling applications without having to worry about your security. Auth0 now, thank yourself later.
    Try free now
  • 5
    Nuclio

    Nuclio

    High-Performance Serverless event and data processing platform

    Nuclio is an open source and managed serverless platform used to minimize development and maintenance overhead and automate the deployment of data-science-based applications. Real-time performance running up to 400,000 function invocations per second. Portable across low laptops, edge, on-prem and multi-cloud deployments. The first serverless platform supporting GPUs for optimized utilization and sharing. Automated deployment to production in a few clicks from Jupyter notebook. Deploy one of...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 6
    NVIDIA AI Cluster Runtime (AICR)

    NVIDIA AI Cluster Runtime (AICR)

    Tooling for optimized and reproducible GPU-accelerated AI runtime

    ...Based on its positioning within NVIDIA’s repositories, it is designed to support scalable AI runtime environments, potentially addressing challenges related to orchestration, resource management, or reproducible AI execution. The project likely aligns with NVIDIA’s broader strategy of building modular infrastructure layers that integrate with GPU-accelerated workloads and cloud-native systems. It appears to emphasize automation, consistency, and performance optimization across AI pipelines, potentially targeting enterprise and research use cases. Given NVIDIA’s ecosystem, it may also integrate with containerized environments, Kubernetes, or other orchestration frameworks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Kubeflow Trainer

    Kubeflow Trainer

    Distributed AI Model Training and LLM Fine-Tuning on Kubernetes

    ...One of its key innovations is the integration of MPI-based distributed computing within Kubernetes, allowing efficient communication between nodes for high-performance training. It also includes advanced scheduling capabilities through integrations with tools like Kueue and Volcano, enabling topology-aware resource allocation and multi-cluster job orchestration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    LLaMA.go

    LLaMA.go

    llama.go is like llama.cpp in pure Golang

    llama.go is like llama.cpp in pure Golang. The code of the project is based on the legendary ggml.cpp framework of Georgi Gerganov written in C++ with the same attitude to performance and elegance. Both models store FP32 weights, so you'll needs at least 32Gb of RAM (not VRAM or GPU RAM) for LLaMA-7B. Double to 64Gb for LLaMA-13B.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    MLPACK is a C++ machine learning library with emphasis on scalability, speed, and ease-of-use. Its aim is to make machine learning possible for novice users by means of a simple, consistent API, while simultaneously exploiting C++ language features to provide maximum performance and flexibility for expert users. * More info + downloads: https://mlpack.org * Git repo: https://github.com/mlpack/mlpack
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 10
    Darktile

    Darktile

    Darktile is a GPU rendered terminal emulator

    Darktile is a GPU-rendered terminal emulator specifically designed for tiling window managers. It utilizes GPU acceleration to provide smooth and efficient rendering, supporting Unicode and a variety of themes. Darktile includes features like font ligatures, context-aware overlays (hints), and customizable cursors, enhancing the terminal experience for users who prefer tiling window environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB