Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
Emscripten: An LLVM-to-WebAssembly Compiler
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)
Implements a reference architecture for creating information systems