Search Results for "gpu max performance" - Page 12

Showing 457 open source projects for "gpu max performance"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • Auth0 B2B Essentials: SSO, MFA, and RBAC Built In Icon
    Auth0 B2B Essentials: SSO, MFA, and RBAC Built In

    Unlimited organizations, 3 enterprise SSO connections, role-based access control, and pro MFA included. Dev and prod tenants out of the box.

    Auth0's B2B Essentials plan gives you everything you need to ship secure multi-tenant apps. Unlimited orgs, enterprise SSO, RBAC, audit log streaming, and higher auth and API limits included. Add on M2M tokens, enterprise MFA, or additional SSO connections as you scale.
    Sign Up Free
  • 1
    SSD in PyTorch 1.0

    SSD in PyTorch 1.0

    High quality, fast, modular reference implementation of SSD in PyTorch

    This repository implements SSD (Single Shot MultiBox Detector). The implementation is heavily influenced by the projects ssd.pytorch, pytorch-ssd and maskrcnn-benchmark. This repository aims to be the code base for research based on SSD. Multi-GPU training and inference: We use DistributedDataParallel, you can train or test with arbitrary GPU(s), the training schema will change accordingly. Add your own modules without pain. We abstract backbone, Detector, BoxHead, BoxPredictor, etc. You can...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Aviary

    Aviary

    Ray Aviary - evaluate multiple LLMs easily

    Aviary is an LLM serving solution that makes it easy to deploy and manage a variety of open source LLMs. Providing an extensive suite of pre-configured open source LLMs, with defaults that work out of the box. Supporting Transformer models hosted on Hugging Face Hub or present on local disk. Aviary has native support for autoscaling and multi-node deployments thanks to Ray and Ray Serve. Aviary can scale to zero and create new model replicas (each composed of multiple GPU workers) in...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Parallel and Distributed Process System

    Parallel and Distributed Process System

    NOTICE OF CONSOLIDATION & PARTNERSHIP PENDING As of April 2026, the 20

    NOTICE OF CONSOLIDATION & PARTNERSHIP PENDING As of April 2026, the 20 pipelines of the QCAUS/PDPBioGen suites are undergoing consolidation for high-scale institutional research. Core 'Ford 2026' algorithms remain the proprietary IP of the Ford Peace and Justice Foundation. Academic users at partner institutions are currently performing validation; all other commercial inquiries must contact the author Computational Neuroscience: Large-scale neural population dynamics, brain-inspired...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    Chinese-LLaMA-Alpaca 2

    Chinese-LLaMA-Alpaca 2

    Chinese LLaMA-2 & Alpaca-2 Large Model Phase II Project

    This project is developed based on the commercially available large model Llama-2 released by Meta. It is the second phase of the Chinese LLaMA&Alpaca large model project. The Chinese LLaMA-2 base model and the Alpaca-2 instruction fine-tuning large model are open-sourced. These models expand and optimize the Chinese vocabulary on the basis of the original Llama-2, use large-scale Chinese data for incremental pre-training, and further improve the basic semantics and command understanding of...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 5
    Glumpy

    Glumpy

    Python+Numpy+OpenGL, scalable and beautiful scientific visualization

    Glumpy is a Python library that simplifies the development of high-performance, interactive OpenGL visualizations. It abstracts complex OpenGL tasks into Pythonic constructs, making it easier for scientists, artists, and developers to harness the power of the GPU for real-time rendering and data visualization. Glumpy is particularly well-suited for rapid prototyping of graphical applications, and its integration with NumPy and shader programming makes it a powerful tool for both research and creative exploration.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    SCIPIO ERP

    SCIPIO ERP

    eCommerce framework for multinational omnichannel installations

    High-performance DNA for your custom ONLINE SHOP. A flexible composable commerce framework for international and complex multichannel webshops! Scipio ERP is a perfect fit for large-scale, multinational, multichannel operations. Our modern eCommerce store is optimized for a quick time-to-market. You can run multiple at the same time. Highly configurable products and prices, custom promotions and region-based offerings – the store can handle it all.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    pipeless

    pipeless

    A computer vision framework to create and deploy apps in minutes

    ...You can easily use industry-standard models, such as YOLO, or load your custom model in one of the supported inference runtimes. Pipeless ships some of the most popular inference runtimes, such as the ONNX Runtime, allowing you to run inference with high performance on CPU or GPU out-of-the-box. You can deploy your Pipeless application with a single command to edge and IoT devices or the cloud.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    FEDML Open Source

    FEDML Open Source

    The unified and scalable ML library for large-scale training

    ...Highly integrated with TensorOpera open source library, TensorOpera AI provides holistic support of three interconnected AI infrastructure layers: user-friendly MLOps, a well-managed scheduler, and high-performance ML libraries for running any AI jobs across GPU Clouds. A typical workflow is shown in the figure above. When a developer wants to run a pre-built job in Studio or Job Store, TensorOperaLaunch swiftly pairs AI jobs with the most economical GPU resources, and auto-provisions, and effortlessly runs the job, eliminating complex environment setup and management.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    VirtualGL

    VirtualGL

    3D Without Boundaries

    VirtualGL redirects 3D commands from a Unix/Linux OpenGL application onto a server-side GPU and converts the rendered 3D images into a video stream with which remote clients can interact to view and control the 3D application in real time.
    Leader badge
    Downloads: 70,112 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build, govern, and optimize agents and models with Gemini Enterprise Agent Platform.
    Start Free
  • 10
    armorpaint

    armorpaint

    3D PBR Texture Painting Software

    ArmorPaint is stand-alone software designed for physically-based texture painting. Drag & drop your 3D models and start painting. Receive instant visual feedback in the viewport as you paint. Work fast with the convenience of nodes. Paint with fully procedural materials. Build fill layers with material nodes. Use brush nodes to create patterns and procedural brushes. ArmorPaint is designed from scratch to run completely on the GPU. This results in a smooth 4K painting experience on...
    Downloads: 48 This Week
    Last Update:
    See Project
  • 11
    ReplitLM

    ReplitLM

    Inference code and configs for the ReplitLM model family

    ...Developers can fine-tune the models using instruction-tuning techniques to adapt them for specific programming tasks or domains. The models were trained using modern deep learning techniques and large-scale GPU infrastructure to achieve strong performance in code completion and generation tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    text-dedup

    text-dedup

    All-in-one text de-duplication

    text-dedup is a Python library that enables efficient deduplication of large text corpora by using MinHash and other probabilistic techniques to detect near-duplicate content. This is especially useful for NLP tasks where duplicated training data can skew model performance. text-dedup scales to billions of documents and offers tools for chunking, hashing, and comparing text efficiently with low memory usage. It supports Jaccard similarity thresholding, parallel execution, and flexible...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Louvre

    Louvre

    High performance C++ library for building Wayland compositors

    Louvre is a high-performance C++ library designed for building Wayland compositors with a strong emphasis on ease of development. It provides a default way for managing protocols, enabling you to have a basic but functional compositor from day one and progressively explore and customize its functionality to precisely match your requirements. Within Louvre, you have the flexibility to either employ your own OpenGL ES 2.0 shaders/programs, use the LPainter class for fundamental 2D...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    AMD APU Tuning Utility

    AMD APU Tuning Utility

    Unlock the full potential of your Zen-based APU with AMD APU Tuning

    ...Universal x86 Tuning Utility serves as a lightweight alternative to well-known software like Ryzen Master and Intel XTU (Extreme Tuning Utility). One of the main features of the utility is the fine-grained control it offers over the device's processor and GPU(s). Users can access a wide range of settings, allowing them to adjust clock speeds, power limits, voltages, and other performance-related parameters.
    Downloads: 42 This Week
    Last Update:
    See Project
  • 15
    SIG Rust

    SIG Rust

    Rust language bindings for TensorFlow

    SIG Rust provides idiomatic Rust bindings for TensorFlow, making it possible for developers to work with TensorFlow functionality from within the Rust programming language. Rather than replacing TensorFlow itself, it acts as an integration layer that connects Rust applications to the TensorFlow C API. The repository is designed for developers who want Rust’s performance, safety, and systems programming strengths while still accessing TensorFlow’s machine learning capabilities. It includes...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Chinese-LLaMA-Alpaca-2 v2.0

    Chinese-LLaMA-Alpaca-2 v2.0

    Chinese LLaMA & Alpaca large language model + local CPU/GPU training

    This project has open-sourced the Chinese LLaMA model and the Alpaca large model with instruction fine-tuning to further promote the open research of large models in the Chinese NLP community. Based on the original LLaMA , these models expand the Chinese vocabulary and use Chinese data for secondary pre-training, which further improves the basic semantic understanding of Chinese. At the same time, the Chinese Alpaca model further uses Chinese instruction data for fine-tuning, which...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    SRM

    SRM

    C library for the development of Linux OpenGL DRM/KMS apps

    SRM is a C library that simplifies the development of Linux DRM/KMS API applications. With SRM, you can focus on the OpenGL ES 2.0 logic of your application. For each available display, you can start a rendering thread that triggers common events like initializeGL(), paintGL(), resizeGL(), pageFlipped() and uninitializeGL(). SRM allows you to use multiple GPUs simultaneously and automatically finds the most efficient configuration. It also offers functions for creating OpenGL textures,...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    OptiMate

    OptiMate

    Libraries for optimizing AI models, inference speed, and GPU usage

    ...One of the core components, Speedster, focuses on accelerating model inference by applying state of the art optimization techniques to increase performance while lowering operational costs. Another component, Nos, targets infrastructure optimization by improving GPU utilization in Kubernetes clusters through dynamic partitioning and elastic resource quotas.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 19
    LLaMA.go

    LLaMA.go

    llama.go is like llama.cpp in pure Golang

    llama.go is like llama.cpp in pure Golang. The code of the project is based on the legendary ggml.cpp framework of Georgi Gerganov written in C++ with the same attitude to performance and elegance. Both models store FP32 weights, so you'll needs at least 32Gb of RAM (not VRAM or GPU RAM) for LLaMA-7B. Double to 64Gb for LLaMA-13B.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    MAX StressTester

    MAX StressTester

    MAX StressTester: Python stress testing tool for max CPU, RAM and Disk

    ...Also, windows might flag the exe file for a virus as it a python to exe file (might fix it later) this explains why it could be flagged as a virus on some virus detectors: https://s.id/1Cuk1 MAX StressTester is an open-source stress testing tool developed by ssit, available on the popular software development platform, SourceForge. The tool is written in Python, making it easy to use and highly customizable for stress testing purposes. The MAX StressTester tool allows users to simulate high loads of traffic on their websites or servers to test their performance and reliability under heavy usage. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Ambient

    Ambient

    The multiplayer game engine

    Ambient is an open-source, cross-platform runtime and engine for building and deploying high-performance multiplayer games and 3D applications, using a modern stack built on Rust, WebAssembly (WASM), and WebGPU. It aims to make multiplayer game development accessible and flexible, providing an entity-component-system (ECS) at its core that doubles as a real-time in-game database; everything in the game — from world objects to runtime data — is represented as entities + components, which can...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    Alphafold

    Alphafold

    Open source code for AlphaFold

    This package provides an implementation of the inference pipeline of AlphaFold v2.0. This is a completely new model that was entered in CASP14 and published in Nature. For simplicity, we refer to this model as AlphaFold throughout the rest of this document. Any publication that discloses findings arising from using this source code or the model parameters should cite the AlphaFold paper. Please also refer to the Supplementary Information for a detailed description of the method. You can use...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    MLPACK is a C++ machine learning library with emphasis on scalability, speed, and ease-of-use. Its aim is to make machine learning possible for novice users by means of a simple, consistent API, while simultaneously exploiting C++ language features to provide maximum performance and flexibility for expert users. * More info + downloads: https://mlpack.org * Git repo: https://github.com/mlpack/mlpack
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    MetalPetal

    MetalPetal

    A GPU accelerated image and video processing framework built on Metal

    MetalPetal is an image processing framework based on Metal designed to provide real-time processing for still images and video with easy-to-use programming interfaces. This chapter covers the key concepts of MetalPetal, and will help you to get a better understanding of its design, implementation, performance implications, and best practices. A MTIImage object is a representation of an image to be processed or produced. It does directly represent image bitmap data instead it has all the...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Veldrid

    Veldrid

    A low-level, portable graphics library for .NET

    Veldrid is a low-level, portable graphics library for .NET, providing a unified API over multiple graphics backends such as Direct3D, Vulkan, OpenGL, and Metal. It enables developers to write high-performance, cross-platform graphics applications without being tied to a specific graphics API. Veldrid is suitable for game development, simulations, and other applications requiring advanced graphics capabilities.
    Downloads: 1 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB