High-performance neural network inference framework for mobile
ONNX Runtime: cross-platform, high performance ML inferencing
Protect and discover secrets using Gitleaks
Port of Facebook's LLaMA model in C/C++
Library for serving Transformers models on Amazon SageMaker
AIMET is a library that provides advanced quantization and compression
Unified Model Serving Framework
Everything you need to build state-of-the-art foundation models
AICI: Prompts as (Wasm) Programs
An MLOps framework to package, deploy, monitor and manage models
Official inference library for Mistral models
Set of comprehensive computer vision & machine intelligence libraries
PArallel Distributed Deep LEarning: Machine Learning Framework
Easy-to-use deep learning framework with 3 key features
Neural Network Compression Framework for enhanced OpenVINO
Standardized Serverless ML Inference Platform on Kubernetes
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
A general-purpose probabilistic programming system
Powering Amazon custom machine learning chips
Superduper: Integrate AI models and machine learning workflows
A GPU-accelerated library containing highly optimized building blocks
LLM.swift is a simple and readable library
Open-source tool designed to enhance the efficiency of workloads
A unified framework for scalable computing
A set of Docker images for training and serving models in TensorFlow