Showing 60 open source projects for "gpu process"

View related business solutions
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • 1
    InvokeAI

    InvokeAI

    InvokeAI is a leading creative engine for Stable Diffusion models

    InvokeAI is an implementation of Stable Diffusion, the open source text-to-image and image-to-image generator. It provides a streamlined process with various new features and options to aid the image generation process. It runs on Windows, Mac and Linux machines, and runs on GPU cards with as little as 4 GB or RAM. InvokeAI is a leading creative engine built to empower professionals and enthusiasts alike. Generate and create stunning visual media using the latest AI-driven technologies. ...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 2
    Unsloth Studio

    Unsloth Studio

    Unified web UI for training and running open models locally

    Unsloth Studio is a web-based interface for running and training AI models locally with a unified and user-friendly experience. It allows users to work with a wide range of models for text, audio, vision, embeddings, and more without relying heavily on cloud infrastructure. Built on top of the Unsloth framework, it focuses on high-performance training with reduced VRAM usage and faster speeds compared to traditional methods. The platform supports fine-tuning, pretraining, and reinforcement...
    Downloads: 18 This Week
    Last Update:
    See Project
  • 3
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser. The WebLLM package itself does not come...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 4
    ArrayFire

    ArrayFire

    ArrayFire, a general purpose GPU library

    ArrayFire is a general-purpose tensor library that simplifies the process of software development for the parallel architectures found in CPUs, GPUs, and other hardware acceleration devices. The library serves users in every technical computing market. Data structures in ArrayFire are smartly managed to avoid costly memory transfers and to take advantage of each performance feature provided by the underlying hardware. The community of ArrayFire developers invites you to build with us if...
    Downloads: 4 This Week
    Last Update:
    See Project
  • Full-stack observability with actually useful AI | Grafana Cloud Icon
    Full-stack observability with actually useful AI | Grafana Cloud

    Our generous forever free tier includes the full platform, including the AI Assistant, for 3 users with 10k metrics, 50GB logs, and 50GB traces.

    Built on open standards like Prometheus and OpenTelemetry, Grafana Cloud includes Kubernetes Monitoring, Application Observability, Incident Response, plus the AI-powered Grafana Assistant. Get started with our generous free tier today.
    Create free account
  • 5
    CUDA Containers for Edge AI & Robotics

    CUDA Containers for Edge AI & Robotics

    Machine Learning Containers for NVIDIA Jetson and JetPack-L4T

    ...These containers simplify the deployment of complex machine learning environments by bundling libraries such as CUDA, TensorRT, and deep learning frameworks into reproducible container images. The project is particularly useful for developers building edge AI and robotics systems that rely on GPU-accelerated inference and real-time computer vision. By using containerized environments, developers can ensure that their applications run consistently across different Jetson platforms and JetPack versions. The repository also includes build tools and package management utilities that help automate the process of assembling machine learning environments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    local-llm

    local-llm

    Run LLMs locally on Cloud Workstations

    local-llm is a development framework that enables developers to run large language models locally within Google Cloud Workstations or standard environments without requiring GPU hardware. It focuses on making generative AI development more accessible by leveraging quantized models and CPU-based execution, eliminating the dependency on expensive GPU infrastructure. The repository includes tools, Docker configurations, and command-line utilities that simplify the process of downloading, running, and interacting with language models directly on local or cloud-based workstations. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 7
    Petastorm

    Petastorm

    Petastorm library enables single machine or distributed training

    Petastorm library enables single machine or distributed training and evaluation of deep learning models from datasets in Apache Parquet format. It supports ML frameworks such as Tensorflow, Pytorch, and PySpark and can be used from pure Python code. Petastorm is an open-source data access library developed at Uber ATG. This library enables single machine or distributed training and evaluation of deep learning models directly from datasets in Apache Parquet format. Petastorm supports popular...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    DocArray

    DocArray

    The data structure for multimodal data

    DocArray is a library for nested, unstructured, multimodal data in transit, including text, image, audio, video, 3D mesh, etc. It allows deep-learning engineers to efficiently process, embed, search, recommend, store, and transfer multimodal data with a Pythonic API. Door to multimodal world: super-expressive data structure for representing complicated/mixed/nested text, image, video, audio, 3D mesh data. The foundation data structure of Jina, CLIP-as-service, DALL·E Flow, DiscoArt etc. Data science powerhouse: greatly accelerate data scientists’ work on embedding, k-NN matching, querying, visualizing, evaluating via Torch/TensorFlow/ONNX/PaddlePaddle on CPU/GPU. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Qwen2.5-Omni

    Qwen2.5-Omni

    Capable of understanding text, audio, vision, video

    Qwen2.5-Omni is an end-to-end multimodal flagship model in the Qwen series by Alibaba Cloud, designed to process multiple modalities (text, images, audio, video) and generate responses both as text and natural speech in streaming real-time. It supports “Thinker-Talker” architecture, and introduces innovations for aligning modalities over time (for example synchronizing video/audio), robust speech generation, and low-VRAM/quantized versions to make usage more accessible. It holds...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 10
    Parallel and Distributed Process System

    Parallel and Distributed Process System

    NOTICE OF CONSOLIDATION & PARTNERSHIP PENDING As of April 2026, the 20

    NOTICE OF CONSOLIDATION & PARTNERSHIP PENDING As of April 2026, the 20 pipelines of the QCAUS/PDPBioGen suites are undergoing consolidation for high-scale institutional research. Core 'Ford 2026' algorithms remain the proprietary IP of the Ford Peace and Justice Foundation. Academic users at partner institutions are currently performing validation; all other commercial inquiries must contact the author Computational Neuroscience: Large-scale neural population dynamics, brain-inspired...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    GitHub Actions for DigitalOcean

    GitHub Actions for DigitalOcean

    GitHub Actions for DigitalOcean - doctl

    ...Powerful and production-ready, our cloud platform has the solutions that devs like you need to succeed, whether you're building world-changing AI apps, running a side project, or building a business. GPU solutions for everyone—novice to expert. Run training and inference, process large data sets and complex neural networks, and deploy high-performance computing clusters.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 12
    Firefly LLM

    Firefly LLM

    A large model training tool that supports training large models

    Firefly is an open-source framework designed to simplify the training and fine-tuning of large language models through a unified and configurable workflow. The project provides a comprehensive environment where developers can perform tasks such as model pre-training, instruction tuning, and preference optimization using widely adopted machine learning techniques. Its architecture supports both full-parameter training and parameter-efficient strategies like LoRA and QLoRA, making it suitable...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    text-generation-webui-colab

    text-generation-webui-colab

    A colab gradio web UI for running Large Language Models

    text-generation-webui-colab is a repository that provides Google Colab notebooks designed to simplify the process of running large language models through the popular text-generation-webui interface. The project automates the setup and deployment of AI models in cloud-based notebook environments, allowing users to experiment with text generation systems without configuring complex local environments. By leveraging Google Colab, the repository enables users to run open-source models such as LLaMA-based systems and other instruction-tuned models using accessible GPU resources. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Neural Tangents

    Neural Tangents

    Fast and Easy Infinite Neural Networks in Python

    Neural Tangents is a high-level neural network API for specifying complex, hierarchical models at both finite and infinite width, built in Python on top of JAX and XLA. It lets researchers define architectures from familiar building blocks—convolutions, pooling, residual connections, and nonlinearities—and obtain not only the finite network but also the corresponding Gaussian Process (GP) kernel of its infinite-width limit. With a single specification, you can compute NNGP and NTK kernels,...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 15
    EvaDB

    EvaDB

    Database system for building simpler and faster AI-powered application

    ...They are accurate on various tasks ranging from question answering to object tracking in videos. To use an AI model, the user needs to program against multiple low-level libraries, like PyTorch, Hugging Face, Open AI, etc. This tedious process often leads to a complex AI app that glues together these libraries to accomplish the given task. This programming complexity prevents people who are experts in other domains from benefiting from these models. Running these deep learning models on large document or video datasets is costly and time-consuming. For example, the state-of-the-art object detection model takes multiple GPU years to process just a week’s videos from a single traffic monitoring camera. ...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 16
    PyTorch Implementation of SDE Solvers

    PyTorch Implementation of SDE Solvers

    Differentiable SDE solvers with GPU support and efficient sensitivity

    This library provides stochastic differential equation (SDE) solvers with GPU support and efficient backpropagation. examples/demo.ipynb gives a short guide on how to solve SDEs, including subtle points such as fixing the randomness in the solver and the choice of noise types. examples/latent_sde.py learns a latent stochastic differential equation, as in Section 5 of [1]. The example fits an SDE to data, whilst regularizing it to be like an Ornstein-Uhlenbeck prior process.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Demucs

    Demucs

    Code for the paper Hybrid Spectrogram and Waveform Source Separation

    ...The repository includes pretrained models for common tasks such as isolating vocals, drums, bass, and accompaniment from stereo music, achieving state-of-the-art results in benchmarks like MUSDB18. Demucs supports GPU-accelerated inference and can process multi-channel audio with chunked streaming for real-time or batch operation. It also provides training scripts and utilities to fine-tune on custom datasets, along with remixing and enhancement tools.
    Downloads: 100 This Week
    Last Update:
    See Project
  • 18
    Super Easy AI Installer Tool

    Super Easy AI Installer Tool

    Application that simplifies the installation of AI-related projects

    "Super Easy AI Installer Tool" is a user-friendly application that simplifies the installation process of AI-related repositories for users. The tool is designed to provide an easy-to-use solution for accessing and installing AI repositories with minimal technical hassle to none the tool will automatically handle the installation process, making it easier for users to access and use AI tools. "Super Easy AI Installer Tool" is currently in early development phase and may have a few bugs. But...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 19
    Rio

    Rio

    A hardware-accelerated GPU terminal emulator powered by WebGPU.

    Rio is a terminal application that’s built with Rust, WebGPU, Tokio runtime. It targets to have the best frame per second experience as long you want, but is also configurable to use as minimal from GPU. It also relies on Rust memory behavior, since Rust is a memory-safe language that employs The terminal renderer is based on redux state machine, lines that has not updated will not suffer a redraw. Looking for the minimal rendering process in most of the time. Rio is also designed to support WebAssembly runtime so in the future you will be able to define how a tab system will work with a WASM plugin written in your favorite language. ...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 20
    Stable Diffusion

    Stable Diffusion

    A latent text-to-image diffusion model

    Stable Diffusion is a widely used open-source latent text-to-image diffusion model developed by the CompVis group for generating high-quality images from natural language prompts. The model operates by conditioning a diffusion process on text embeddings produced by a CLIP text encoder, enabling detailed and controllable image synthesis. It was trained on large-scale image datasets and later fine-tuned to produce 512×512 images with strong visual fidelity. Because the system runs efficiently...
    Downloads: 36 This Week
    Last Update:
    See Project
  • 21
    Diffusers-Interpret

    Diffusers-Interpret

    Model explainability for Diffusers

    ...To analyze how a token in the input prompt influenced the generation, you can study the token attribution scores. You can also check all the images that the diffusion process generated at the end of each step. Gradient checkpointing also reduces GPU usage, but makes computations a bit slower.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    What happens when

    What happens when

    What happens when you type google into your browser and press enter?

    What happens when is a large collaborative documentation-style project that aims to answer in exhaustive detail the canonical interview/thought experiment question, “What happens when you type google into your browser and press Enter?” Rather than giving a high-level overview, the repository tries to break down every step in the process, from low-level events (keyboard press, OS events, keyboard interrupts), through OS-level handling (keyboard scan codes, key events), parsing, DNS lookup, networking (ARP, socket creation, TCP/TLS handshake), HTTP requests, browser behavior, HTML/CSS/JS parsing, rendering engine, GPU rendering, layout, to final drawing and user-visible output. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Robust Video Matting (RVM)

    Robust Video Matting (RVM)

    Robust Video Matting in PyTorch, TensorFlow, TensorFlow.js, ONNX

    We introduce a robust, real-time, high-resolution human video matting method that achieves new state-of-the-art performance. Our method is much lighter than previous approaches and can process 4K at 76 FPS and HD at 104 FPS on an Nvidia GTX 1080Ti GPU. Unlike most existing methods that perform video matting frame-by-frame as independent images, our method uses a recurrent architecture to exploit temporal information in videos and achieves significant improvements in temporal coherence and matting quality. Furthermore, we propose a novel training strategy that enforces our network on both matting and segmentation objectives. ...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 24
    macOS Simple KVM

    macOS Simple KVM

    Tools to set up a quick macOS VM in QEMU, accelerated by KVM

    ...The project also supports GPU passthrough and other advanced configurations for users who want a more optimized macOS VM environment. While primarily intended for educational and testing purposes, it demonstrates how macOS can be virtualized outside of Apple hardware.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 25
    YouTube-8M

    YouTube-8M

    Starter code for working with the YouTube-8M dataset

    youtube-8m is Google’s open source starter code and reference implementation for training and evaluating machine learning models on the YouTube-8M dataset, one of the largest video understanding datasets publicly released. The repository provides a complete pipeline for video-level and frame-level modeling using TensorFlow, including data reading, model training, evaluation, and inference. It was developed to support the YouTube-8M Video Understanding Challenge (hosted on Kaggle and featured...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB