A high-performance inference system for large language models
LLMs as Copilots for Theorem Proving in Lean
Framework which allows you transform your Vector Database
lightweight, standalone C++ inference engine for Google's Gemma models
Bolt is a deep learning library with high performance
A RWKV management and startup tool, full automation, only 8MB
Lightweight inference library for ONNX files, written in C++
An innovative library for efficient LLM inference
LLM.swift is a simple and readable library
Turn your existing data infrastructure into a feature store
OpenAI swift async text to image for SwiftUI app using OpenAI
MNN is a blazing fast, lightweight deep learning framework
Run serverless GPU workloads with fast cold starts on bare-metal
Pure C++ implementation of several models for real-time chatting
Connect home devices into a powerful cluster to accelerate LLM
Protect and discover secrets using Gitleaks
Run local LLMs like llama, deepseek, kokoro etc. inside your browser
User-friendly AI Interface
Set of comprehensive computer vision & machine intelligence libraries
A general-purpose probabilistic programming system
Serving system for machine learning models
Prem provides a unified environment to develop AI applications
LLMFlows - Simple, Explicit and Transparent LLM Apps
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere
Serve machine learning models within a Docker container