Run Local LLMs on Any Device. Open-source
Easy-to-use Speech Toolkit including Self-Supervised Learning model
A toolkit to optimize ML models for deployment for Keras & TensorFlow
The Triton Inference Server provides an optimized cloud
Visual Instruction Tuning: Large Language-and-Vision Assistant
DoWhy is a Python library for causal inference
Superduper: Integrate AI models and machine learning workflows
MII makes low-latency and high-throughput inference possible
The unofficial python package that returns response of Google Bard
Open platform for training, serving, and evaluating language models
A high-performance ML model serving framework, offers dynamic batching
Tensor search for humans
Database system for building simpler and faster AI-powered application
LLMFlows - Simple, Explicit and Transparent LLM Apps
A computer vision framework to create and deploy apps in minutes
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere