Training and deploying machine learning models on Amazon SageMaker
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
A high-throughput and memory-efficient inference and serving engine
Everything you need to build state-of-the-art foundation models
Standardized Serverless ML Inference Platform on Kubernetes
Fast inference engine for Transformer models
Optimizing inference proxy for LLMs
Framework that is dedicated to making neural data processing
Create HTML profiling reports from pandas DataFrame objects
A set of Docker images for training and serving models in TensorFlow
Operating LLMs in production
Single-cell analysis in Python
Set of comprehensive computer vision & machine intelligence libraries
Open-Source AI Camera. Empower any camera/CCTV
Sparsity-aware deep learning inference runtime for CPUs
MII makes low-latency and high-throughput inference possible
Easy-to-use deep learning framework with 3 key features
OpenMMLab Model Deployment Framework
Deep learning optimization library: makes distributed training easy
DoWhy is a Python library for causal inference
An easy-to-use LLMs quantization package with user-friendly apis
A Pythonic framework to simplify AI service building
20+ high-performance LLMs with recipes to pretrain, finetune at scale
Easiest and laziest way for building multi-agent LLMs applications