Port of Facebook's LLaMA model in C/C++
Run Local LLMs on Any Device. Open-source
A high-throughput and memory-efficient inference and serving engine
SimpleMem: Efficient Lifelong Memory for LLM Agents
Central interface to connect your LLM's with external data
Easy-to-use LLM fine-tuning framework (LLaMA-2, BLOOM, Falcon
Structured outputs for llms
LLM abstractions that aren't obstructions
Open-source observability for your LLM application
An AI personal assistant for your digital brain
A New Axis of Sparsity for Large Language Models
ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat
Simple, Pythonic building blocks to evaluate LLM applications
One API for plugins and datasets, one interface for prompt engineering
Curated list of datasets and tools for post-training
Inference Llama 2 in one file of pure C
An elegent pytorch implement of transformers
PandasAI is a Python library that integrates generative AI
A RWKV management and startup tool, full automation, only 8MB
LLM training in simple, raw C/CUDA
Vector database plugin for Postgres, written in Rust
File Parser optimised for LLM Ingestion with no loss
Framework and no-code GUI for fine-tuning LLMs
Integrating LLMs into structured NLP pipelines
Toolkit for conversational AI