A scalable inference server for models optimized with OpenVINO
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
Easiest and laziest way for building multi-agent LLMs applications
The free, Open Source alternative to OpenAI, Claude and others
Deep Learning API and Server in C++14 support for Caffe, PyTorch
Visual Instruction Tuning: Large Language-and-Vision Assistant
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere