Run Local LLMs on Any Device. Open-source
LLMs as Copilots for Theorem Proving in Lean
User-friendly AI Interface
Open standard for machine learning interoperability
ONNX Runtime: cross-platform, high performance ML inferencing
An MLOps framework to package, deploy, monitor and manage models
FlashInfer: Kernel Library for LLM Serving
The free, Open Source alternative to OpenAI, Claude and others
A GPU-accelerated library containing highly optimized building blocks
Pytorch domain library for recommendation systems
C++ library for high performance inference on NVIDIA GPUs
Low-latency REST API for serving text-embeddings
Operating LLMs in production
Large Language Model Text Generation Inference
Replace OpenAI GPT with another LLM in your app
Standardized Serverless ML Inference Platform on Kubernetes
Single-cell analysis in Python
Training and deploying machine learning models on Amazon SageMaker
Uncover insights, surface problems, monitor, and fine tune your LLM
A set of Docker images for training and serving models in TensorFlow
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
Set of comprehensive computer vision & machine intelligence libraries
Tensor search for humans
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
A high-performance ML model serving framework, offers dynamic batching