State-of-the-art Parameter-Efficient Fine-Tuning
Easiest and laziest way for building multi-agent LLMs applications
Bring the notion of Model-as-a-Service to life
Tensor search for humans
Pytorch domain library for recommendation systems
MII makes low-latency and high-throughput inference possible
Official inference library for Mistral models
A set of Docker images for training and serving models in TensorFlow
Easy-to-use Speech Toolkit including Self-Supervised Learning model
A toolkit to optimize ML models for deployment for Keras & TensorFlow
A library for accelerating Transformer models on NVIDIA GPUs
Superduper: Integrate AI models and machine learning workflows
20+ high-performance LLMs with recipes to pretrain, finetune at scale
GPU environment management and cluster orchestration
Lightweight Python library for adding real-time multi-object tracking
Create HTML profiling reports from pandas DataFrame objects
Powering Amazon custom machine learning chips
Low-latency REST API for serving text-embeddings
Library for serving Transformers models on Amazon SageMaker
Replace OpenAI GPT with another LLM in your app
LLM training code for MosaicML foundation models
An MLOps framework to package, deploy, monitor and manage models
Standardized Serverless ML Inference Platform on Kubernetes
Integrate, train and manage any AI models and APIs with your database
Open platform for training, serving, and evaluating language models