Serve, optimize and scale PyTorch models in production
The Triton Inference Server provides an optimized cloud
Deep Learning API and Server in C++14 support for Caffe, PyTorch
A library for accelerating Transformer models on NVIDIA GPUs
Data manipulation and transformation for audio signal processing
Library for OCR-related tasks powered by Deep Learning
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Unified Model Serving Framework
A GPU-accelerated library containing highly optimized building blocks
High-level Deep Learning Framework written in Kotlin