Training and deploying machine learning models on Amazon SageMaker
Run Local LLMs on Any Device. Open-source
Ready-to-use OCR with 80+ supported languages
A high-throughput and memory-efficient inference and serving engine
The official Python client for the Huggingface Hub
Unified Model Serving Framework
Uncover insights, surface problems, monitor, and fine tune your LLM
Bring the notion of Model-as-a-Service to life
Efficient few-shot learning with Sentence Transformers
Single-cell analysis in Python
State-of-the-art diffusion models for image and audio generation
Trainable models and NN optimization tools
A library for accelerating Transformer models on NVIDIA GPUs
Easiest and laziest way for building multi-agent LLMs applications
Official inference library for Mistral models
Visual Instruction Tuning: Large Language-and-Vision Assistant
Easy-to-use Speech Toolkit including Self-Supervised Learning model
OpenMMLab Model Deployment Framework
Library for OCR-related tasks powered by Deep Learning
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
A Pythonic framework to simplify AI service building
Everything you need to build state-of-the-art foundation models
20+ high-performance LLMs with recipes to pretrain, finetune at scale
The Triton Inference Server provides an optimized cloud
Standardized Serverless ML Inference Platform on Kubernetes