Images to inference with no labeling
OpenMMLab Model Deployment Framework
State-of-the-art Parameter-Efficient Fine-Tuning
A Pythonic framework to simplify AI service building
Uncover insights, surface problems, monitor, and fine tune your LLM
Efficient few-shot learning with Sentence Transformers
Uplift modeling and causal inference with machine learning algorithms
Data manipulation and transformation for audio signal processing
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Operating LLMs in production
A library to communicate with ChatGPT, Claude, Copilot, Gemini
Trainable models and NN optimization tools
Probabilistic reasoning and statistical analysis in TensorFlow
State-of-the-art diffusion models for image and audio generation
Optimizing inference proxy for LLMs
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
A lightweight vision library for performing large object detection
A Unified Library for Parameter-Efficient Learning
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
PyTorch extensions for fast R&D prototyping and Kaggle farming
The Triton Inference Server provides an optimized cloud
Easy-to-use Speech Toolkit including Self-Supervised Learning model
Unified Model Serving Framework
A high-performance ML model serving framework, offers dynamic batching
Framework that is dedicated to making neural data processing