Deep learning optimization library: makes distributed training easy
Private Open AI on Kubernetes
Framework which allows you transform your Vector Database
lightweight, standalone C++ inference engine for Google's Gemma models
Lightweight inference library for ONNX files, written in C++
Bolt is a deep learning library with high performance
A RWKV management and startup tool, full automation, only 8MB
An innovative library for efficient LLM inference
LLM.swift is a simple and readable library
OpenAI swift async text to image for SwiftUI app using OpenAI
Turn your existing data infrastructure into a feature store
Run serverless GPU workloads with fast cold starts on bare-metal
Pure C++ implementation of several models for real-time chatting
Connect home devices into a powerful cluster to accelerate LLM
User-friendly AI Interface
Protect and discover secrets using Gitleaks
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
Set of comprehensive computer vision & machine intelligence libraries
A general-purpose probabilistic programming system
Serving system for machine learning models
LLMFlows - Simple, Explicit and Transparent LLM Apps
Serve machine learning models within a Docker container
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
Self-contained Machine Learning and Natural Language Processing lib
Framework for Accelerating LLM Generation with Multiple Decoding Heads