Training and deploying machine learning models on Amazon SageMaker
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
A high-throughput and memory-efficient inference and serving engine
Deep learning optimization library: makes distributed training easy
Fast inference engine for Transformer models
A general-purpose probabilistic programming system
Everything you need to build state-of-the-art foundation models
20+ high-performance LLMs with recipes to pretrain, finetune at scale
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)
Large Language Model Text Generation Inference
Unified Model Serving Framework
A set of Docker images for training and serving models in TensorFlow
Open-Source AI Camera. Empower any camera/CCTV
A Pythonic framework to simplify AI service building
Integrate, train and manage any AI models and APIs with your database
GPU environment management and cluster orchestration
State-of-the-art Parameter-Efficient Fine-Tuning
Easy-to-use deep learning framework with 3 key features
Replace OpenAI GPT with another LLM in your app
State-of-the-art diffusion models for image and audio generation
An MLOps framework to package, deploy, monitor and manage models
The official Python client for the Huggingface Hub
Optimizing inference proxy for LLMs
Sparsity-aware deep learning inference runtime for CPUs