Optimizing inference proxy for LLMs
Openai style api for open large language models
Images to inference with no labeling
LLM training code for MosaicML foundation models
Easiest and laziest way for building multi-agent LLMs applications
Standardized Serverless ML Inference Platform on Kubernetes
Probabilistic reasoning and statistical analysis in TensorFlow
A set of Docker images for training and serving models in TensorFlow
An MLOps framework to package, deploy, monitor and manage models
A library for accelerating Transformer models on NVIDIA GPUs
A Unified Library for Parameter-Efficient Learning
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
Deep learning optimization library: makes distributed training easy
A lightweight vision library for performing large object detection
Open platform for training, serving, and evaluating language models
Tensor search for humans
Simplifies the local serving of AI models from any source
A toolkit to optimize ML models for deployment for Keras & TensorFlow
Multi-Modal Neural Networks for Semantic Search, based on Mid-Fusion
Powering Amazon custom machine learning chips
High quality, fast, modular reference implementation of SSD in PyTorch
OpenMMLab Model Deployment Framework
Framework that is dedicated to making neural data processing
Database system for building simpler and faster AI-powered application
A graphical manager for ollama that can manage your LLMs