Training and deploying machine learning models on Amazon SageMaker
Port of OpenAI's Whisper model in C/C++
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
A high-throughput and memory-efficient inference and serving engine
An MLOps framework to package, deploy, monitor and manage models
Easy-to-use deep learning framework with 3 key features
Sparsity-aware deep learning inference runtime for CPUs
OpenMLDB is an open-source machine learning database
AIMET is a library that provides advanced quantization and compression
Superduper: Integrate AI models and machine learning workflows
Integrate, train and manage any AI models and APIs with your database
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
Operating LLMs in production
DoWhy is a Python library for causal inference
Uplift modeling and causal inference with machine learning algorithms
Easiest and laziest way for building multi-agent LLMs applications
Large Language Model Text Generation Inference
Multilingual Automatic Speech Recognition with word-level timestamps
OpenMMLab Model Deployment Framework
Deep learning optimization library: makes distributed training easy
A unified framework for scalable computing
Database system for building simpler and faster AI-powered application