Showing 299 open source projects for "linux nvidia"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 1
    TensorRT LLM

    TensorRT LLM

    TensorRT LLM provides users with an easy-to-use Python API

    TensorRT-LLM is an open-source high-performance inference library specifically designed to optimize and accelerate large language model deployment on NVIDIA GPUs. It provides a Python-based API built on top of PyTorch that allows developers to define, customize, and deploy LLMs efficiently across a variety of hardware configurations, from single GPUs to large multi-node clusters. The library focuses on maximizing throughput and minimizing latency through advanced techniques such as...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    TTOS Linux

    TTOS Linux

    Desktop Linux Operating System

    TTOS Linux is a desktop linux operating system based on Debian Linux. It includes many custom utilities TTOS Specific to separate it from other debian variants.
    Leader badge
    Downloads: 69 This Week
    Last Update:
    See Project
  • 3
    OSMO

    OSMO

    The developer-first platform for scaling complex Physical AI workloads

    OSMO is a developer-first orchestration platform designed to scale complex physical AI workflows across heterogeneous computing environments, including cloud GPUs, simulation clusters, and edge devices. It was originally built internally at NVIDIA to support robotics and embodied AI systems, where workflows span multiple stages such as data generation, training, simulation, and hardware testing. The platform addresses what NVIDIA refers to as the “three computer problem” by unifying these...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    clone-voice

    clone-voice

    A sound cloning tool with a web interface, using your voice

    Clone-voice is a local voice-cloning tool that lets you synthesize speech in any target voice or convert one recording into another voice using the same timbre. It is built around Coqui’s XTTS-v2 model, so it inherits multilingual support and modern neural TTS quality while wrapping it in a user-friendly desktop workflow. The app is designed to be very easy to use: you download a precompiled package, double-click app.exe, and it launches a browser-based web interface where you control...
    Downloads: 11 This Week
    Last Update:
    See Project
  • Custom VMs From 1 to 96 vCPUs With 99.95% Uptime Icon
    Custom VMs From 1 to 96 vCPUs With 99.95% Uptime

    General-purpose, compute-optimized, or GPU/TPU-accelerated. Built to your exact specs.

    Live migration and automatic failover keep workloads online through maintenance. One free e2-micro VM every month.
    Try Free
  • 5
    FastKoko

    FastKoko

    Dockerized FastAPI wrapper for Kokoro-82M text-to-speech model

    FastKoko is a self-hosted text-to-speech server built around the Kokoro-82M model and exposed through a FastAPI backend. It is designed to be easy to deploy via Docker, with separate CPU and GPU images so that users can choose between pure CPU inference and NVIDIA GPU acceleration. The project exposes an OpenAI-compatible speech endpoint, which means existing code that talks to the OpenAI audio API can often be pointed at a Kokoro-FastAPI instance with minimal changes. It supports multiple...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 6
    CUDA.jl

    CUDA.jl

    CUDA programming in Julia

    High-performance GPU programming in a high-level language. JuliaGPU is a GitHub organization created to unify the many packages for programming GPUs in Julia. With its high-level syntax and flexible compiler, Julia is well-positioned to productively program hardware accelerators like GPUs without sacrificing performance. The latest development version of CUDA.jl requires Julia 1.8 or higher. If you are using an older version of Julia, you need to use a previous version of CUDA.jl. This will...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    Likwid

    Likwid

    Performance monitoring and benchmarking suite

    Likwid is a simple to install and use toolsuite of command line applications and a library for performance oriented programmers. It works for Intel, AMD, ARMv8 and POWER9 processors on the Linux operating system. There is additional support for Nvidia and AMD GPUs. There is support for ARMv7 and POWER8/9 but there is currently no test machine in our hands to test them properly.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 8
    Archium Linux

    Archium Linux

    Archium is easier take on Arch Linux.

    Archium Linux is an Arch-based Linux distribution designed to simplify installation and initial system setup without sacrificing the flexibility and transparency of Arch Linux. It provides a guided TUI installer, sensible defaults, and optional desktop environments such as KDE Plasma and GNOME, allowing users to get a working system quickly while still maintaining full control. Archium stays close to upstream Arch, avoiding heavy modifications or unnecessary abstraction. The goal is to...
    Leader badge
    Downloads: 97 This Week
    Last Update:
    See Project
  • 9
    Humanoid-Gym

    Humanoid-Gym

    Reinforcement Learning for Humanoid Robot with Zero-Shot Sim2Real

    Humanoid-Gym is a reinforcement learning framework designed to train locomotion and control policies for humanoid robots using high-performance simulation environments. The system is built on top of NVIDIA Isaac Gym, which allows large-scale parallel simulation of robotic environments directly on GPU hardware. Its primary goal is to enable efficient training of humanoid robots in simulation while enabling policies to transfer effectively to real-world hardware without additional training....
    Downloads: 1 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 10
    GPU Hot

    GPU Hot

    Real-time NVIDIA GPU dashboard

    GPU Hot is an open-source, lightweight monitoring dashboard designed to provide real-time visibility into NVIDIA GPU performance across single machines or entire clusters. The project offers a self-hosted web interface that streams hardware metrics directly from GPU servers, enabling developers, ML engineers, and system administrators to observe GPU utilization and system behavior in real time through a browser. The dashboard collects and displays a wide range of performance metrics...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    Proton-GE-Custom

    Proton-GE-Custom

    Compatibility tool for Steam Play based on Wine

    Proton-GE-Custom is an unofficial compatibility tool for Steam Play that builds on Valve’s Proton and Wine codebases to provide additional patches, fixes, and enhancements for running Windows games on Linux systems through Steam, particularly on Steam Deck and other Linux gaming platforms. Unlike the official Proton builds, this custom fork includes community-developed patches such as media foundation support, raw input improvements, AMD and Nvidia optimizations, and broader compatibility tweaks that help certain games launch and run more reliably or with better performance. ...
    Downloads: 55 This Week
    Last Update:
    See Project
  • 12
    NVTOP

    NVTOP

    GPU & Accelerator process monitoring for AMD, Apple, Huawei, Intel

    NVTOP stands for Neat Video card TOP, a (h) top-like task monitor for GPUs and accelerators. It can handle multiple GPUs and print information about them in a htop-familiar way. Currently supported vendors are AMD (Linux AMD GPU driver), Apple (limited M1 & M2 support), Huawei (Ascend), Intel (Linux i915 driver), NVIDIA (Linux proprietary divers), and Qualcomm Adreno (Linux MSM driver).
    Downloads: 8 This Week
    Last Update:
    See Project
  • 13
    CuPy

    CuPy

    A NumPy-compatible array library accelerated by CUDA

    CuPy is an open source implementation of NumPy-compatible multi-dimensional array accelerated with NVIDIA CUDA. It consists of cupy.ndarray, a core multi-dimensional array class and many functions on it. CuPy offers GPU accelerated computing with Python, using CUDA-related libraries to fully utilize the GPU architecture. According to benchmarks, it can even speed up some operations by more than 100X. CuPy is highly compatible with NumPy, serving as a drop-in replacement in most...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 14
    JAX Toolbox

    JAX Toolbox

    Public CI, Docker images for popular JAX libraries

    JAX Toolbox is a development toolkit designed to streamline and optimize the use of JAX for machine learning and high-performance computing on NVIDIA GPUs. It provides prebuilt Docker images, continuous integration pipelines, and optimized example implementations that help developers quickly set up and run JAX workloads without complex configuration. The project supports popular JAX-based frameworks and models, including architectures used for large-scale pretraining such as GPT and LLaMA...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    OptiScaler

    OptiScaler

    OptiScaler bridges upscaling/frame gen across GPUs

    OptiScaler is an open-source modding tool designed to enhance graphics performance and flexibility in modern PC games by allowing users to replace or modify built-in upscaling technologies. Instead of relying on the upscaling method originally integrated by a game developer, the software intercepts the game’s rendering pipeline and redirects it to alternative technologies chosen by the user. This makes it possible to swap technologies such as NVIDIA DLSS, AMD FSR, or Intel XeSS even if the...
    Downloads: 132 This Week
    Last Update:
    See Project
  • 16
    TAME LLM

    TAME LLM

    Traditional Mandarin LLMs for Taiwan

    TAME LLM is an open-source initiative focused on building and releasing large language models optimized for Traditional Mandarin and the linguistic context of Taiwan. The project includes models such as Llama-3-Taiwan-70B, which are fine-tuned versions of large transformer architectures trained on extensive corpora containing both Traditional Mandarin and English text. These models are designed to support applications such as conversational AI, knowledge retrieval, and domain-specific...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    WanGP

    WanGP

    AI video generator optimized for low VRAM and older GPUs use

    Wan2GP is an open source AI video generation toolkit designed to make modern generative models accessible on consumer-grade hardware with limited GPU memory. It acts as a unified interface for running multiple video, image, and audio generation models, including Wan-based models as well as other systems like Hunyuan Video, Flux, and Qwen. A key focus of the project is reducing VRAM requirements, enabling some workflows to run on as little as 6 GB while still supporting older Nvidia and...
    Downloads: 33 This Week
    Last Update:
    See Project
  • 18
    WhiteSur GTK Theme

    WhiteSur GTK Theme

    MacOS like theme for all gtk based desktops

    WhiteSur-gtk-theme brings a macOS Big Sur–inspired look to Linux desktops by providing a polished GTK theme with light and dark variants, rounded shapes, and refined translucency. It includes assets and installer scripts to apply the theme across GTK applications and desktop shells that support GTK theming, aiming for a cohesive, high-contrast interface. The project pays attention to details like window controls, titlebars, selection states, and widget hover effects so apps feel consistent...
    Downloads: 32 This Week
    Last Update:
    See Project
  • 19
    exo

    exo

    Run your own AI cluster at home with everyday devices

    Run your own AI cluster at home with everyday devices. Maintained by exo labs. Forget expensive NVIDIA GPUs, unify your existing devices into one powerful GPU, iPhone, iPad, Android, Mac, Linux, or pretty much any device. Now the default models, run 8B, 70B, and 405B parameter models on your own devices.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    waifu2x ncnn Vulkan

    waifu2x ncnn Vulkan

    waifu2x converter ncnn version, run fast GPU with vulkan

    ncnn implementation of waifu2x converter. Runs fast on Intel/AMD/Nvidia/Apple-Silicon with Vulkan API. waifu2x-ncnn-vulkan uses ncnn project as the universal neural network inference framework.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    FlashAttention

    FlashAttention

    Fast and memory-efficient exact attention

    FlashAttention is a high-performance deep learning optimization library that reimplements the attention mechanism used in transformer models to be significantly faster and more memory-efficient than standard implementations. It achieves this by using IO-aware algorithms that minimize memory reads and writes, reducing the quadratic memory overhead typically associated with attention operations. The project provides implementations of FlashAttention, FlashAttention-2, and newer iterations...
    Downloads: 21 This Week
    Last Update:
    See Project
  • 22
    FlashMLA

    FlashMLA

    FlashMLA: Efficient Multi-head Latent Attention Kernels

    FlashMLA is a high-performance decoding kernel library designed especially for Multi-Head Latent Attention (MLA) workloads, targeting NVIDIA Hopper GPU architectures. It provides optimized kernels for MLA decoding, including support for variable-length sequences, helping reduce latency and increase throughput in model inference systems using that attention style. The library supports both BF16 and FP16 data types, and includes a paged KV cache implementation with a block size of 64 to...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Jellyfin Android TV

    Jellyfin Android TV

    Android TV Client for Jellyfin

    Jellyfin Android TV is a Jellyfin client for Android TV, Nvidia Shield, and Amazon Fire TV devices. We welcome all contributions and pull requests! If you have a larger feature in mind please open an issue so we can discuss the implementation before you start. Jellyfin is the volunteer-built media solution that puts you in control of your media. Stream to any device from your own server, with no strings attached. Your media, your server, your way. Jellyfin enables you to collect, manage, and...
    Downloads: 32 This Week
    Last Update:
    See Project
  • 24
    Apollo

    Apollo

    The easiest way to stream with the native resolution of your client

    Apollo is a self-hosted desktop streaming host designed to enable low-latency game streaming from a personal computer to remote clients using protocols compatible with Moonlight and Artemis. It acts as a server that captures, encodes, and streams desktop or game sessions while supporting hardware acceleration across AMD, Intel, and NVIDIA GPUs. The project includes a web-based interface that allows users to configure streaming settings, manage connected clients, and control application...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 25
    OpenShell

    OpenShell

    OpenShell is the safe, private runtime for autonomous AI agents.

    OpenShell is an open-source runtime designed to safely run autonomous AI agents in isolated environments. Developed by NVIDIA, it provides sandboxed execution spaces that protect system resources, credentials, and data from unauthorized access. Each agent runs inside a containerized sandbox governed by declarative YAML security policies that control network access, file permissions, and process behavior. The platform includes a gateway service that manages sandbox lifecycles and routes AI...
    Downloads: 28 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB