A Unified Library for Parameter-Efficient Learning
Large Language Model Text Generation Inference
Images to inference with no labeling
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
GPU environment management and cluster orchestration
PyTorch library of curated Transformer models and their components
State-of-the-art Parameter-Efficient Fine-Tuning
Superduper: Integrate AI models and machine learning workflows
A high-performance ML model serving framework, offers dynamic batching
Framework that is dedicated to making neural data processing
MII makes low-latency and high-throughput inference possible
PyTorch extensions for fast R&D prototyping and Kaggle farming
Probabilistic reasoning and statistical analysis in TensorFlow
Open platform for training, serving, and evaluating language models
Low-latency REST API for serving text-embeddings
A library for accelerating Transformer models on NVIDIA GPUs
Multilingual Automatic Speech Recognition with word-level timestamps
Multi-Modal Neural Networks for Semantic Search, based on Mid-Fusion
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Tensor search for humans
Run 100B+ language models at home, BitTorrent-style
An MLOps framework to package, deploy, monitor and manage models
A toolkit to optimize ML models for deployment for Keras & TensorFlow
High quality, fast, modular reference implementation of SSD in PyTorch
Create HTML profiling reports from pandas DataFrame objects