Search Results for "nvidia gpu mod" - Page 4

Showing 155 open source projects for "nvidia gpu mod"

View related business solutions
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • Go from Code to Production URL in Seconds Icon
    Go from Code to Production URL in Seconds

    Cloud Run deploys apps in any language instantly. Scales to zero. Pay only when code runs.

    Skip the Kubernetes configs. Cloud Run handles HTTPS, scaling, and infrastructure automatically. Two million requests free per month.
    Try it free
  • 1
    FastChat

    FastChat

    Open platform for training, serving, and evaluating language models

    ...If you do not have enough memory, you can enable 8-bit compression by adding --load-8bit to the commands above. This can reduce memory usage by around half with slightly degraded model quality. It is compatible with the CPU, GPU, and Metal backend. Vicuna-13B with 8-bit compression can run on a single NVIDIA 3090/4080/T4/V100(16GB) GPU. In addition to that, you can add --cpu-offloading to commands above to offload weights that don't fit on your GPU onto the CPU memory. This requires 8-bit compression to be enabled and the bitsandbytes package to be installed, which is only available on linux operating systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    DxTonia

    DxTonia

    A tool for detecting the presence of leg dystonia from videos.

    ...The software has been verified on two datasets and achieved higher accuracy than routine clinical care. Stay tuned for the improved version. Note: Currently only works on Windows PC with Nvidia GPU. Soon, mac and Linux CPU only version will be released. The requirment files are in old-version_scripts folder. Instructions to use DxTonia https://www.rishabh-bajpai.com/dxtoniainstallation https://www.medrxiv.org/content/10.1101/2024.10.15.24315574
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    oxDNA

    oxDNA

    A code primarily aimed at DNA and RNA coarse-grained simulations

    The oxDNA code has been moved to https://github.com/lorenzo-rovigatti/oxDNA, please go there for new releases.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 4
    TTOS Linux

    TTOS Linux

    Desktop Linux Operating System

    TTOS Linux is a desktop linux operating system based on Debian Linux. It includes many custom utilities TTOS Specific to separate it from other debian variants.
    Leader badge
    Downloads: 65 This Week
    Last Update:
    See Project
  • Enterprise-grade ITSM, for every business Icon
    Enterprise-grade ITSM, for every business

    Give your IT, operations, and business teams the ability to deliver exceptional services—without the complexity.

    Freshservice is an intuitive, AI-powered platform that helps IT, operations, and business teams deliver exceptional service without the usual complexity. Automate repetitive tasks, resolve issues faster, and provide seamless support across the organization. From managing incidents and assets to driving smarter decisions, Freshservice makes it easy to stay efficient and scale with confidence.
    Try it Free
  • 5
    This is a freedom software and it obeys opensource initiative. The main purpose of this project is to bring extra updates to the existing Octave OCL project with more supporting PC devices as well as mobile devices. Due to the fact tremendous time and energy has been spent on its working on many devices, the project name is changing to AnyArray to reflect it is really working on many devices on March 2026. Ever since started providing extra GPU Compute support, it has been for...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    VanitySearch

    VanitySearch

    VanitySearch is a Bitcoin address prefix lookup tool.

    VanitySearch is a Bitcoin address prefix lookup tool. If you want to generate a secure private key, use the `-s` option to enter your passphrase, which will be used to generate the base key conforming to the BIP38 standard (e.g., `VanitySearch.exe -s "my passphrase" 1MyPrefix"`). You can also use `VanitySearch.exe -ps "my passphrase"`, which adds a cryptographically secure seed to your passphrase.Fixed custom address matching errors and private key conversion errors, changed the randomizer,...
    Leader badge
    Downloads: 32 This Week
    Last Update:
    See Project
  • 7
    lamco-rdp-server

    lamco-rdp-server

    A full-featured Wayland-native RDP server for Linux written in Rust.

    lamco-rdp-server is a native Wayland RDP server for Linux desktops. It provides remote desktop access using the standard RDP protocol, so you can connect from any RDP client: Windows Remote Desktop, FreeRDP, Remmina, or others. It supports multiple screen capture and input backends: * xdg-desktop-portal: works with GNOME, KDE Plasma, and other portal-compatible compositors * Native wlroots protocols: works with Sway, Hyprland, and other wlroots-based compositors The server detects...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    BackgroundMattingV2

    BackgroundMattingV2

    Real-Time High-Resolution Background Matting

    Official repository for the paper Real-Time High-Resolution Background Matting. Our model requires capturing an additional background image and produces state-of-the-art matting results at 4K 30fps and HD 60fps on an Nvidia RTX 2080 TI GPU.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Transformers4Rec

    Transformers4Rec

    Transformers4Rec is a flexible and efficient library

    Transformers4Rec is an advanced recommendation system library that leverages Transformer models for sequential and session-based recommendations. The library works as a bridge between natural language processing (NLP) and recommender systems (RecSys) by integrating with one of the most popular NLP frameworks, Hugging Face Transformers (HF). Transformers4Rec makes state-of-the-art transformer architectures available for RecSys researchers and industry practitioners. Traditional recommendation...
    Downloads: 8 This Week
    Last Update:
    See Project
  • Auth0 B2B Essentials: SSO, MFA, and RBAC Built In Icon
    Auth0 B2B Essentials: SSO, MFA, and RBAC Built In

    Unlimited organizations, 3 enterprise SSO connections, role-based access control, and pro MFA included. Dev and prod tenants out of the box.

    Auth0's B2B Essentials plan gives you everything you need to ship secure multi-tenant apps. Unlimited orgs, enterprise SSO, RBAC, audit log streaming, and higher auth and API limits included. Add on M2M tokens, enterprise MFA, or additional SSO connections as you scale.
    Sign Up Free
  • 10
    Parallel WaveGAN

    Parallel WaveGAN

    Unofficial Parallel WaveGAN

    Parallel WaveGAN is an unofficial PyTorch implementation of several state-of-the-art non-autoregressive neural vocoders, centered on Parallel WaveGAN but also including MelGAN, Multiband-MelGAN, HiFi-GAN, and StyleMelGAN. Its main goal is to provide a real-time neural vocoder that can turn mel spectrograms into high-quality speech audio efficiently. The repository is designed to work hand-in-hand with ESPnet-TTS and NVIDIA Tacotron2-style front ends, so you can build complete TTS or singing...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Super Easy AI Installer Tool

    Super Easy AI Installer Tool

    Application that simplifies the installation of AI-related projects

    "Super Easy AI Installer Tool" is a user-friendly application that simplifies the installation process of AI-related repositories for users. The tool is designed to provide an easy-to-use solution for accessing and installing AI repositories with minimal technical hassle to none the tool will automatically handle the installation process, making it easier for users to access and use AI tools. "Super Easy AI Installer Tool" is currently in early development phase and may have a few bugs. But...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    GSvit

    GSvit

    Fast FDTD solver with graphics card support

    Fast FDTD solver with graphics card support. Optimized for nanoscale optics - scanning near field optical microscopy, rough surface scattering and solar cells. Uses CUDA environment for graphics card operation.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 13
    NVIDIA Container Toolkit

    NVIDIA Container Toolkit

    Build and run Docker containers leveraging NVIDIA GPUs

    The NVIDIA Container Toolkit allows users to build and run GPU accelerated Docker containers. The toolkit includes a container runtime library and utilities to automatically configure containers to leverage NVIDIA GPUs. Make sure you have installed the NVIDIA driver and Docker engine for your Linux distribution Note that you do not need to install the CUDA Toolkit on the host system, but the NVIDIA driver needs to be installed.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 14
    NIM Pools Hub Miner

    NIM Pools Hub Miner

    CPU & GPU Nimiq Miner for NVIDIA & AMD

    Nimiq Miner with a focus on user experience and ease of use.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    DLSS2FSR

    DLSS2FSR

    CyberFSR/PotatoFSR (FSR 2.1.2) for Games

    Replacing DLSS 2.x with FSR 2.1.2 for various games such as Cyberpunk 2077, Red Dead Redemption 2, etc. It includes a DLL wrapper/injector (winmm.dll) to disable Nvidia GeForce RTX GPU checking, so AMD Radeon / Nvidia GTX users can enjoy the mod. Download the latest relase from Release Page. Extract the contents of the archive next to the game EXE file in your games folder. Make sure the game is running in DX12 mode. Download the latest relase from Release Page. Extract the contents of the archive next to the game EXE file in your games folder.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 16

    NAM-Runner

    Batch file to install and run NAM (neural-amp-modeler) easily.

    ...Custom one-time installation of everything you need to train neural network models of guitar amps and more for the NAM VST plugin, no Conda required. Runs as a launcher afterwards. Portable installation. New pyTorch inclues CUDA runtime for fast Nvidia GPU support. No command line, python or conda knowledge needed! Just double click.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Knet

    Knet

    Koç University deep learning framework

    Knet.jl is a deep learning package implemented in Julia, so you should be able to run it on any machine that can run Julia. It has been extensively tested on Linux machines with NVIDIA GPUs and CUDA libraries, and it has been reported to work on OSX and Windows. If you would like to try it on your own computer, please follow the instructions on Installation. If you would like to try working with a GPU and do not have access to one, take a look at Using Amazon AWS or Using Microsoft Azure. If you find a bug, please open a GitHub issue. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    FasterTransformer

    FasterTransformer

    Transformer related optimization, including BERT, GPT

    FasterTransformer is a high-performance inference library designed to accelerate transformer-based models such as BERT, GPT, and T5 on NVIDIA GPUs. It provides optimized implementations of transformer encoder and decoder layers using CUDA, cuBLAS, and custom kernels to maximize throughput and minimize latency. The library supports multiple deep learning frameworks, including TensorFlow, PyTorch, and Triton, allowing developers to integrate it into existing pipelines without major changes. It...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    G-Diffuser Bot

    G-Diffuser Bot

    Discord bot and Interface for Stable Diffusion

    The first release of the all-in-one installer version of G-Diffuser is here. This release no longer requires the installation of WSL or Docker and has a systray icon to keep track of and launch G-Diffuser components. The infinite zoom scripts have been updated with some improvements, notably a new compositer script that is hundreds of times faster than before. The first release of the all-in-one installer is here. It notably features much easier "one-click" installation and updating, as well...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    NBMiner

    NBMiner

    GPU Miner for ETH, RVN, BEAM, CFX, ZIL, AE, ERGO

    nbminer.com & NBMiner_github are the only 2 officially maintained site for publishing information and new releases of NBMiner. Be aware when you download NBMiner binaries from other sources. For GPUs with Hynix GDDR6 memory, LHR mode is not recommended for the poor performance. Improve performance on Grin29 & AE. Add support for mining SERO, algo progpow_sero. Change devfee in percentage, [0-5]. Set to ‘0’ to turn off devfee with lower hash rate. Otherwise, devfee = max(set_value,...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    GFPGAN

    GFPGAN

    GFPGAN aims at developing Practical Algorithms

    GFPGAN aims at developing Practical Algorithms for Real-world Face Restoration. Colab Demo for GFPGAN; (Another Colab Demo for the original paper model) Online demo: Huggingface (return only the cropped face) Online demo: Replicate.ai (may need to sign in, return the whole image). Online demo: Baseten.co (backed by GPU, returns the whole image). We provide a clean version of GFPGAN, which can run without CUDA extensions. So that it can run in Windows or on CPU mode. GFPGAN aims at developing...
    Downloads: 72 This Week
    Last Update:
    See Project
  • 22
    BCI

    BCI

    BCI: Breast Cancer Immunohistochemical Image Generation

    Breast Cancer Immunohistochemical Image Generation through Pyramid Pix2pix. We have released the trained model on BCI and LLVIP datasets. We host a competition for breast cancer immunohistochemistry image generation on Grand Challenge. Project pix2pix provides a python script to generate pix2pix training data in the form of pairs of images {A,B}, where A and B are two different depictions of the same underlying scene, these can be pairs {HE, IHC}. Then we can learn to translate A(HE images)...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Fairseq

    Fairseq

    Facebook AI Research Sequence-to-Sequence Toolkit written in Python

    Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks. We provide reference implementations of various sequence modeling papers. Recent work by Microsoft and Google has shown that data parallel training can be made significantly more efficient by sharding the model parameters and optimizer state across data parallel workers. These ideas are encapsulated in the...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    SageMaker MXNet Inference Toolkit

    SageMaker MXNet Inference Toolkit

    Toolkit for allowing inference and serving with MXNet in SageMaker

    ...AWS Deep Learning Containers (DLCs) are a set of Docker images for training and serving models in TensorFlow, TensorFlow 2, PyTorch, and MXNet. Deep Learning Containers provide optimized environments with TensorFlow and MXNet, Nvidia CUDA (for GPU instances), and Intel MKL (for CPU instances) libraries and are available in the Amazon Elastic Container Registry (Amazon ECR). The AWS DLCs are used in Amazon SageMaker as the default vehicles for your SageMaker jobs such as training, inference, transforms etc. They've been tested for machine learning workloads on Amazon EC2, Amazon ECS and Amazon EKS services as well.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Hugging Face Transformer

    Hugging Face Transformer

    CPU/GPU inference server for Hugging Face transformer models

    ...Then, if you spend some time, you can build something over ONNX Runtime and Triton inference server. You will usually get from 2X to 4X faster inference compared to vanilla Pytorch. It's cool! However, if you want the best in class performances on GPU, there is only a single possible combination: Nvidia TensorRT and Triton. You will usually get 5X faster inference compared to vanilla Pytorch.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB