A scalable inference server for models optimized with OpenVINO
Superduper: Integrate AI models and machine learning workflows
Self-contained Machine Learning and Natural Language Processing lib
The AI-native (edge and LLM) proxy for agents
AICI: Prompts as (Wasm) Programs
Images to inference with no labeling
Bolt is a deep learning library with high performance
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Library for OCR-related tasks powered by Deep Learning
Private Open AI on Kubernetes
Official inference library for Mistral models
A high-performance ML model serving framework, offers dynamic batching
Trainable, memory-efficient, and GPU-friendly PyTorch reproduction
PArallel Distributed Deep LEarning: Machine Learning Framework
Run 100B+ language models at home, BitTorrent-style
Phi-3.5 for Mac: Locally-run Vision and Language Models
Prem provides a unified environment to develop AI applications
A RWKV management and startup tool, full automation, only 8MB
Simplifies the local serving of AI models from any source
Training and deploying machine learning models on Amazon SageMaker
Data manipulation and transformation for audio signal processing
Serve, optimize and scale PyTorch models in production
lightweight, standalone C++ inference engine for Google's Gemma models
Optimizing inference proxy for LLMs
A graphical manager for ollama that can manage your LLMs