Showing 210 open source projects for "linux nvidia"

View related business solutions
  • $300 in Free Credit Towards Top Cloud Services Icon
    $300 in Free Credit Towards Top Cloud Services

    Build VMs, containers, AI, databases, storage—all in one place.

    Start your project in minutes. After credits run out, 20+ products include free monthly usage. Only pay when you're ready to scale.
    Get Started
  • Go From AI Idea to AI App Fast Icon
    Go From AI Idea to AI App Fast

    One platform to build, fine-tune, and deploy ML models. No MLOps team required.

    Access Gemini 3 and 200+ models. Build chatbots, agents, or custom models with built-in monitoring and scaling.
    Try Free
  • 1
    FastChat

    FastChat

    Open platform for training, serving, and evaluating language models

    ...This can reduce memory usage by around half with slightly degraded model quality. It is compatible with the CPU, GPU, and Metal backend. Vicuna-13B with 8-bit compression can run on a single NVIDIA 3090/4080/T4/V100(16GB) GPU. In addition to that, you can add --cpu-offloading to commands above to offload weights that don't fit on your GPU onto the CPU memory. This requires 8-bit compression to be enabled and the bitsandbytes package to be installed, which is only available on linux operating systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    DxTonia

    DxTonia

    A tool for detecting the presence of leg dystonia from videos.

    ...The software has been verified on two datasets and achieved higher accuracy than routine clinical care. Stay tuned for the improved version. Note: Currently only works on Windows PC with Nvidia GPU. Soon, mac and Linux CPU only version will be released. The requirment files are in old-version_scripts folder. Instructions to use DxTonia https://www.rishabh-bajpai.com/dxtoniainstallation https://www.medrxiv.org/content/10.1101/2024.10.15.24315574
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Kompute

    Kompute

    General purpose GPU compute framework built on Vulkan

    General purpose GPU compute framework built on Vulkan to support 1000s of cross-vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). Blazing fast, mobile-enabled, asynchronous, and optimized for advanced GPU data processing use cases. Backed by the Linux Foundation.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    This is a freedom software and it obeys opensource initiative. The main purpose of this project is to bring extra updates to the existing Octave OCL project with more supporting PC devices like Intel GPU, AMD APUs, and Apple Silicon GPU as well as mobile devices like Qualcomm Adreno and ARM Mali GPUs. Due to the fact tremendous time and energy has been spent on its working on many devices, the project name is changing to AnyArray to reflect it is really working on many devices in 9 March...
    Downloads: 0 This Week
    Last Update:
    See Project
  • AI-powered service management for IT and enterprise teams Icon
    AI-powered service management for IT and enterprise teams

    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity. Maximize operational efficiency with refreshingly simple, AI-powered Freshservice.
    Try it Free
  • 5
    BackgroundMattingV2

    BackgroundMattingV2

    Real-Time High-Resolution Background Matting

    Official repository for the paper Real-Time High-Resolution Background Matting. Our model requires capturing an additional background image and produces state-of-the-art matting results at 4K 30fps and HD 60fps on an Nvidia RTX 2080 TI GPU.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    Transformers4Rec

    Transformers4Rec

    Transformers4Rec is a flexible and efficient library

    Transformers4Rec is an advanced recommendation system library that leverages Transformer models for sequential and session-based recommendations. The library works as a bridge between natural language processing (NLP) and recommender systems (RecSys) by integrating with one of the most popular NLP frameworks, Hugging Face Transformers (HF). Transformers4Rec makes state-of-the-art transformer architectures available for RecSys researchers and industry practitioners. Traditional recommendation...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    solo-learn

    solo-learn

    Library of self-supervised methods for visual representation

    A library of self-supervised methods for visual representation learning powered by Pytorch Lightning. A library of self-supervised methods for unsupervised visual representation learning powered by PyTorch Lightning. We aim at providing SOTA self-supervised methods in a comparable environment while, at the same time, implementing training tricks. The library is self-contained, but it is possible to use the models outside of solo-learn.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    Downloads: 14 This Week
    Last Update:
    See Project
  • 9
    Downloads: 4 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    Parallel WaveGAN

    Parallel WaveGAN

    Unofficial Parallel WaveGAN

    Parallel WaveGAN is an unofficial PyTorch implementation of several state-of-the-art non-autoregressive neural vocoders, centered on Parallel WaveGAN but also including MelGAN, Multiband-MelGAN, HiFi-GAN, and StyleMelGAN. Its main goal is to provide a real-time neural vocoder that can turn mel spectrograms into high-quality speech audio efficiently. The repository is designed to work hand-in-hand with ESPnet-TTS and NVIDIA Tacotron2-style front ends, so you can build complete TTS or singing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Super Easy AI Installer Tool

    Super Easy AI Installer Tool

    Application that simplifies the installation of AI-related projects

    "Super Easy AI Installer Tool" is a user-friendly application that simplifies the installation process of AI-related repositories for users. The tool is designed to provide an easy-to-use solution for accessing and installing AI repositories with minimal technical hassle to none the tool will automatically handle the installation process, making it easier for users to access and use AI tools. "Super Easy AI Installer Tool" is currently in early development phase and may have a few bugs. But...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    panda

    panda

    Panda Linux (Arch Based)

    Panda Linux is a minimalist distribution, based on Arch Linux. Complete with installers for gaming platforms such as Steam™, Battle.net™ and many others via Lutris software, available in the package manager. Use an image writer such as Balena-Etcher or Rufus to burn the .iso image to a USB flash drive for best results. Download Etcher here: https://etcher.balena.io/#download-etcher Download Rufus here: https://rufus.ie/en/ Login details for the live user session: Username: isouser Password : isouser Nvidia users can install the latest Nvidia driver with the Nvidia-Installer utility. ...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 14
    enhancr

    enhancr

    Video Frame Interpolation & Super Resolution using NVIDIA's TensorRT

    ...The GUI was designed to provide a stunning experience powered by state-of-the-art technologies without feeling clunky and outdated like other alternatives. It features blazing-fast TensorRT inference by NVIDIA, which can speed up AI processes significantly. Pre-packaged, without the need to install Docker or WSL (Windows Subsystem for Linux) - and NCNN inference by Tencent which is lightweight and runs on NVIDIA, AMD and even Apple Silicon - in contrast to the mammoth of an inference PyTorch is, which only runs on NVIDIA GPUs.
    Downloads: 31 This Week
    Last Update:
    See Project
  • 15
    Lightning Bolts

    Lightning Bolts

    Toolbox of models, callbacks, and datasets for AI/ML researchers

    Bolts package provides a variety of components to extend PyTorch Lightning, such as callbacks & datasets, for applied research and production. Torch ORT converts your model into an optimized ONNX graph, speeding up training & inference when using NVIDIA or AMD GPUs. We can introduce sparsity during fine-tuning with SparseML, which ultimately allows us to leverage the DeepSparse engine to see performance improvements at inference time.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    NIM Pools Hub Miner

    NIM Pools Hub Miner

    CPU & GPU Nimiq Miner for NVIDIA & AMD

    Nimiq Miner with a focus on user experience and ease of use.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 17
    Transformer Reinforcement Learning X

    Transformer Reinforcement Learning X

    A repo for distributed training of language models with Reinforcement

    trlX is a distributed training framework designed from the ground up to focus on fine-tuning large language models with reinforcement learning using either a provided reward function or a reward-labeled dataset. Training support for Hugging Face models is provided by Accelerate-backed trainers, allowing users to fine-tune causal and T5-based language models of up to 20B parameters, such as facebook/opt-6.7b, EleutherAI/gpt-neox-20b, and google/flan-t5-xxl. For models beyond 20B parameters,...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    Thrust

    Thrust

    The C++ parallel algorithms library

    Thrust is the C++ parallel algorithms library which inspired the introduction of parallel algorithms to the C++ Standard Library. Thrust's high-level interface greatly enhances programmer productivity while enabling performance portability between GPUs and multicore CPUs. It builds on top of established parallel programming frameworks (such as CUDA, TBB, and OpenMP). It also provides a number of general-purpose facilities similar to those found in the C++ Standard Library. The NVIDIA C++...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19

    NAM-Runner

    Batch file to install and run NAM (neural-amp-modeler) easily.

    A Windows 10 batch file, that installs and runs the NAM model trainer (neural-amp-modeler) by Steven Atkinson right into the GUI application. Fully automated. Custom one-time installation of everything you need to train neural network models of guitar amps and more for the NAM VST plugin, no Conda required. Runs as a launcher afterwards. Portable installation. New pyTorch inclues CUDA runtime for fast Nvidia GPU support. No command line, python or conda knowledge needed! Just double click.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 20
    Knet

    Knet

    Koç University deep learning framework

    Knet.jl is a deep learning package implemented in Julia, so you should be able to run it on any machine that can run Julia. It has been extensively tested on Linux machines with NVIDIA GPUs and CUDA libraries, and it has been reported to work on OSX and Windows. If you would like to try it on your own computer, please follow the instructions on Installation. If you would like to try working with a GPU and do not have access to one, take a look at Using Amazon AWS or Using Microsoft Azure. If you find a bug, please open a GitHub issue. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    FasterTransformer

    FasterTransformer

    Transformer related optimization, including BERT, GPT

    FasterTransformer is a high-performance inference library designed to accelerate transformer-based models such as BERT, GPT, and T5 on NVIDIA GPUs. It provides optimized implementations of transformer encoder and decoder layers using CUDA, cuBLAS, and custom kernels to maximize throughput and minimize latency. The library supports multiple deep learning frameworks, including TensorFlow, PyTorch, and Triton, allowing developers to integrate it into existing pipelines without major changes. It...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22

    Proteus Model Builder

    GUI for training of neural network models for GuitarML Proteus

    GUI for easier installation and training of neural network models for guitar amplifiers and pedals, based on the GuitarML Proteus models. These are usable for Proteus, Chowdhury-DSP BYOD and even NeuralPi, on all platforms incl. Linux and RaspberryPi. What is this? GuitarML's work on Proteus, NeuralPi and Proteusboard (hardware) is amazing. https://github.com/GuitarML Yet, it is not easy to wrap your head around if you are not familiar with programming, AI, machine learning, neuronal...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 23
    NBMiner

    NBMiner

    GPU Miner for ETH, RVN, BEAM, CFX, ZIL, AE, ERGO

    nbminer.com & NBMiner_github are the only 2 officially maintained site for publishing information and new releases of NBMiner. Be aware when you download NBMiner binaries from other sources. For GPUs with Hynix GDDR6 memory, LHR mode is not recommended for the poor performance. Improve performance on Grin29 & AE. Add support for mining SERO, algo progpow_sero. Change devfee in percentage, [0-5]. Set to ‘0’ to turn off devfee with lower hash rate. Otherwise, devfee = max(set_value,...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 24
    AlphaTensor

    AlphaTensor

    AI discovers faster, efficient algorithms for matrix multiplication

    AlphaTensor, developed by Google DeepMind, is the research codebase accompanying the 2022 Nature publication “Discovering faster matrix multiplication algorithms with reinforcement learning.” The project demonstrates how reinforcement learning can be used to automatically discover efficient algorithms for matrix multiplication — a fundamental operation in computer science and numerical computation. The repository is organized into four main components: algorithms, benchmarking,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Hello AI World

    Hello AI World

    Guide to deploying deep-learning inference networks

    Hello AI World is a great way to start using Jetson and experiencing the power of AI. In just a couple of hours, you can have a set of deep learning inference demos up and running for realtime image classification and object detection on your Jetson Developer Kit with JetPack SDK and NVIDIA TensorRT. The tutorial focuses on networks related to computer vision, and includes the use of live cameras. You’ll also get to code your own easy-to-follow recognition program in Python or C++, and train...
    Downloads: 0 This Week
    Last Update:
    See Project