Port of Facebook's LLaMA model in C/C++
Run Local LLMs on Any Device. Open-source
A high-throughput and memory-efficient inference and serving engine
Central interface to connect your LLM's with external data
Easy-to-use LLM fine-tuning framework (LLaMA-2, BLOOM, Falcon
Structured outputs for llms
LLM abstractions that aren't obstructions
Open-source observability for your LLM application
ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat
An AI personal assistant for your digital brain
SimpleMem: Efficient Lifelong Memory for LLM Agents
Simple, Pythonic building blocks to evaluate LLM applications
Curated list of datasets and tools for post-training
One API for plugins and datasets, one interface for prompt engineering
An elegent pytorch implement of transformers
PandasAI is a Python library that integrates generative AI
Inference Llama 2 in one file of pure C
A RWKV management and startup tool, full automation, only 8MB
Multilingual sentence & image embeddings with BERT
LLM training in simple, raw C/CUDA
Vector database plugin for Postgres, written in Rust
Gorilla: An API store for LLMs
File Parser optimised for LLM Ingestion with no loss
Framework and no-code GUI for fine-tuning LLMs
Low-latency REST API for serving text-embeddings