Port of Facebook's LLaMA model in C/C++
Run Local LLMs on Any Device. Open-source
Superduper: Integrate AI models and machine learning workflows
Deep learning optimization library: makes distributed training easy
Phi-3.5 for Mac: Locally-run Vision and Language Models
Official inference library for Mistral models
Framework that is dedicated to making neural data processing
Implementation of model parallel autoregressive transformers on GPUs
A real time inference engine for temporal logical specifications