A lightweight vision library for performing large object detection
Unified Model Serving Framework
Trainable models and NN optimization tools
MII makes low-latency and high-throughput inference possible
Libraries for applying sparsification recipes to neural networks
A high-performance ML model serving framework, offers dynamic batching
Uncover insights, surface problems, monitor, and fine tune your LLM
The Triton Inference Server provides an optimized cloud
Pytorch domain library for recommendation systems
Powering Amazon custom machine learning chips
The official Python client for the Huggingface Hub
Neural Network Compression Framework for enhanced OpenVINO
Framework that is dedicated to making neural data processing
Easy-to-use Speech Toolkit including Self-Supervised Learning model
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Gaussian processes in TensorFlow
Serve machine learning models within a Docker container
Everything you need to build state-of-the-art foundation models
Probabilistic reasoning and statistical analysis in TensorFlow
Efficient few-shot learning with Sentence Transformers
Images to inference with no labeling
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
Open-source tool designed to enhance the efficiency of workloads
A toolkit to optimize ML models for deployment for Keras & TensorFlow
Library for serving Transformers models on Amazon SageMaker