MII makes low-latency and high-throughput inference possible
Neural Network Compression Framework for enhanced OpenVINO
Openai style api for open large language models
Superduper: Integrate AI models and machine learning workflows
Images to inference with no labeling
Easy-to-use deep learning framework with 3 key features
State-of-the-art diffusion models for image and audio generation
Uncover insights, surface problems, monitor, and fine tune your LLM
Pytorch domain library for recommendation systems
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Integrate, train and manage any AI models and APIs with your database
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
Libraries for applying sparsification recipes to neural networks
An easy-to-use LLMs quantization package with user-friendly apis
A set of Docker images for training and serving models in TensorFlow
Multilingual Automatic Speech Recognition with word-level timestamps
Unified Model Serving Framework
A Unified Library for Parameter-Efficient Learning
Optimizing inference proxy for LLMs
Large Language Model Text Generation Inference
Replace OpenAI GPT with another LLM in your app
Easiest and laziest way for building multi-agent LLMs applications
Efficient few-shot learning with Sentence Transformers
Framework that is dedicated to making neural data processing
Trainable models and NN optimization tools