Run Local LLMs on Any Device. Open-source
A high-throughput and memory-efficient inference and serving engine
Single-cell analysis in Python
A Pythonic framework to simplify AI service building
Everything you need to build state-of-the-art foundation models
Easiest and laziest way for building multi-agent LLMs applications
Official inference library for Mistral models
The official Python client for the Huggingface Hub
Simplifies the local serving of AI models from any source
Uncover insights, surface problems, monitor, and fine tune your LLM
Low-latency REST API for serving text-embeddings
Standardized Serverless ML Inference Platform on Kubernetes
Gaussian processes in TensorFlow
Data manipulation and transformation for audio signal processing
State-of-the-art diffusion models for image and audio generation
An MLOps framework to package, deploy, monitor and manage models
Operating LLMs in production
DoWhy is a Python library for causal inference
State-of-the-art Parameter-Efficient Fine-Tuning
A set of Docker images for training and serving models in TensorFlow
Python Package for ML-Based Heterogeneous Treatment Effects Estimation
A library for accelerating Transformer models on NVIDIA GPUs
Training and deploying machine learning models on Amazon SageMaker
A toolkit to optimize ML models for deployment for Keras & TensorFlow
Probabilistic reasoning and statistical analysis in TensorFlow