Run Local LLMs on Any Device. Open-source
The Triton Inference Server provides an optimized cloud
MII makes low-latency and high-throughput inference possible
DoWhy is a Python library for causal inference
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
Database system for building simpler and faster AI-powered application
Easy-to-use Speech Toolkit including Self-Supervised Learning model
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
Trainable models and NN optimization tools
The unofficial python package that returns response of Google Bard
A toolkit to optimize ML models for deployment for Keras & TensorFlow
Superduper: Integrate AI models and machine learning workflows
Tensor search for humans
LLMFlows - Simple, Explicit and Transparent LLM Apps
A computer vision framework to create and deploy apps in minutes