The Triton Inference Server provides an optimized cloud
Library for OCR-related tasks powered by Deep Learning
Data manipulation and transformation for audio signal processing
Integrate, train and manage any AI models and APIs with your database
A unified framework for scalable computing
Optimizing inference proxy for LLMs
Visual Instruction Tuning: Large Language-and-Vision Assistant
Adversarial Robustness Toolbox (ART) - Python Library for ML security
Efficient few-shot learning with Sentence Transformers
Bring the notion of Model-as-a-Service to life
Pytorch domain library for recommendation systems
Official inference library for Mistral models
Neural Network Compression Framework for enhanced OpenVINO
Sparsity-aware deep learning inference runtime for CPUs
Large Language Model Text Generation Inference
OpenMMLab Model Deployment Framework
A library to communicate with ChatGPT, Claude, Copilot, Gemini
20+ high-performance LLMs with recipes to pretrain, finetune at scale
PyTorch library of curated Transformer models and their components
Multilingual Automatic Speech Recognition with word-level timestamps
AIMET is a library that provides advanced quantization and compression
Superduper: Integrate AI models and machine learning workflows
A high-performance ML model serving framework, offers dynamic batching
Phi-3.5 for Mac: Locally-run Vision and Language Models
State-of-the-art diffusion models for image and audio generation