A scalable inference server for models optimized with OpenVINO
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
The free, Open Source alternative to OpenAI, Claude and others
Easiest and laziest way for building multi-agent LLMs applications
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Open Source and Lightweight Local LLM Platform
Visual Instruction Tuning: Large Language-and-Vision Assistant
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere