State-of-the-art Parameter-Efficient Fine-Tuning
A Pythonic framework to simplify AI service building
Uncover insights, surface problems, monitor, and fine tune your LLM
Efficient few-shot learning with Sentence Transformers
Uplift modeling and causal inference with machine learning algorithms
OpenMMLab Model Deployment Framework
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Operating LLMs in production
A library to communicate with ChatGPT, Claude, Copilot, Gemini
FlashInfer: Kernel Library for LLM Serving
Trainable models and NN optimization tools
Probabilistic reasoning and statistical analysis in TensorFlow
Data manipulation and transformation for audio signal processing
Optimizing inference proxy for LLMs
A lightweight vision library for performing large object detection
A Unified Library for Parameter-Efficient Learning
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
State-of-the-art diffusion models for image and audio generation
PyTorch extensions for fast R&D prototyping and Kaggle farming
A high-performance ML model serving framework, offers dynamic batching
Framework that is dedicated to making neural data processing
Official inference library for Mistral models
OpenVINO™ Toolkit repository
Open-Source AI Camera. Empower any camera/CCTV