Easiest and laziest way for building multi-agent LLMs applications
Integrate, train and manage any AI models and APIs with your database
Pytorch domain library for recommendation systems
Unified Model Serving Framework
Visual Instruction Tuning: Large Language-and-Vision Assistant
AIMET is a library that provides advanced quantization and compression
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Standardized Serverless ML Inference Platform on Kubernetes
Data manipulation and transformation for audio signal processing
Uncover insights, surface problems, monitor, and fine tune your LLM
Superduper: Integrate AI models and machine learning workflows
A high-performance ML model serving framework, offers dynamic batching
Libraries for applying sparsification recipes to neural networks
An easy-to-use LLMs quantization package with user-friendly apis
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT method
State-of-the-art Parameter-Efficient Fine-Tuning
Optimizing inference proxy for LLMs
Neural Network Compression Framework for enhanced OpenVINO
Openai style api for open large language models
Sparsity-aware deep learning inference runtime for CPUs
Large Language Model Text Generation Inference
Images to inference with no labeling
Trainable models and NN optimization tools
Efficient few-shot learning with Sentence Transformers
Simplifies the local serving of AI models from any source