Everything you need to build state-of-the-art foundation models
Open standard for machine learning interoperability
LMDeploy is a toolkit for compressing, deploying, and serving LLMs
The AI-native (edge and LLM) proxy for agents
Framework which allows you transform your Vector Database
Build Production-ready Agentic Workflow with Natural Language
Run serverless GPU workloads with fast cold starts on bare-metal
Build your chatbot within minutes on your favorite device
C++ library for high performance inference on NVIDIA GPUs
AI interface for tinkerers (Ollama, Haystack RAG, Python)
An MLOps framework to package, deploy, monitor and manage models
Set of comprehensive computer vision & machine intelligence libraries
A scalable inference server for models optimized with OpenVINO
Serving system for machine learning models
MNN is a blazing fast, lightweight deep learning framework
Probabilistic reasoning and statistical analysis in TensorFlow
A GPU-accelerated library containing highly optimized building blocks
Trainable models and NN optimization tools
Uncover insights, surface problems, monitor, and fine tune your LLM
Adversarial Robustness Toolbox (ART) - Python Library for ML security
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Powering Amazon custom machine learning chips
A library to communicate with ChatGPT, Claude, Copilot, Gemini
A toolkit to optimize ML models for deployment for Keras & TensorFlow