Distribute and run LLMs with a single file
lightweight, standalone C++ inference engine for Google's Gemma models
On-device AI across mobile, embedded and edge for PyTorch
Open standard for machine learning interoperability
C++ library for high performance inference on NVIDIA GPUs
High-performance neural network inference framework for mobile
CodeGeeX: An Open Multilingual Code Generation Model (KDD 2023)
A GPU-accelerated library containing highly optimized building blocks
Inference framework for 1-bit LLMs
Open Source Computer Vision Library
Open Source Computer Vision Library
Run GGUF models easily with a UI or API. One File. Zero Install.
The world's only naturally intelligent knowledge technology
Award-winning modern data processing SDK in C++20
Tool to parse the command line and configuration files.