MII makes low-latency and high-throughput inference possible
PyTorch extensions for fast R&D prototyping and Kaggle farming
Probabilistic reasoning and statistical analysis in TensorFlow
Low-latency REST API for serving text-embeddings
A library for accelerating Transformer models on NVIDIA GPUs
Multilingual Automatic Speech Recognition with word-level timestamps
Open platform for training, serving, and evaluating language models
Multi-Modal Neural Networks for Semantic Search, based on Mid-Fusion
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Tensor search for humans
Run 100B+ language models at home, BitTorrent-style
An MLOps framework to package, deploy, monitor and manage models
A toolkit to optimize ML models for deployment for Keras & TensorFlow
High quality, fast, modular reference implementation of SSD in PyTorch
Create HTML profiling reports from pandas DataFrame objects
Library for serving Transformers models on Amazon SageMaker
Serve machine learning models within a Docker container
A set of Docker images for training and serving models in TensorFlow
OpenMLDB is an open-source machine learning database
A GPU-accelerated library containing highly optimized building blocks
Implementation of "Tree of Thoughts
Toolbox of models, callbacks, and datasets for AI/ML researchers
Lightweight anchor-free object detection model
Implementation of model parallel autoregressive transformers on GPUs
Sequence-to-sequence framework, focused on Neural Machine Translation