Ready-to-use OCR with 80+ supported languages
Run Local LLMs on Any Device. Open-source
GPU environment management and cluster orchestration
A high-throughput and memory-efficient inference and serving engine
Deep learning optimization library: makes distributed training easy
Standardized Serverless ML Inference Platform on Kubernetes
A library for accelerating Transformer models on NVIDIA GPUs
LLM training code for MosaicML foundation models
AIMET is a library that provides advanced quantization and compression
Simplifies the local serving of AI models from any source
PyTorch library of curated Transformer models and their components
The Triton Inference Server provides an optimized cloud
Easy-to-use Speech Toolkit including Self-Supervised Learning model
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
Low-latency REST API for serving text-embeddings
Trainable models and NN optimization tools
Data manipulation and transformation for audio signal processing
Powering Amazon custom machine learning chips
Integrate, train and manage any AI models and APIs with your database
Replace OpenAI GPT with another LLM in your app
Open-source tool designed to enhance the efficiency of workloads
Tensor search for humans
State-of-the-art diffusion models for image and audio generation
Uncover insights, surface problems, monitor, and fine tune your LLM
An MLOps framework to package, deploy, monitor and manage models