Python bindings for llama.cpp
Structured outputs for llms
Run Local LLMs on Any Device. Open-source
Port of Facebook's LLaMA model in C/C++
INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & GLM4(V)
Phi-3.5 for Mac: Locally-run Vision and Language Models
Low-code app builder for RAG and multi-agent AI applications
A high-throughput and memory-efficient inference and serving engine
A guidance language for controlling large language models
Database system for building simpler and faster AI-powered application
Framework and no-code GUI for fine-tuning LLMs
Revolutionizing Database Interactions with Private LLM Technology
Application that simplifies the installation of AI-related projects
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training
Easy-to-use LLM fine-tuning framework (LLaMA-2, BLOOM, Falcon
PyTorch library of curated Transformer models and their components
Framework to easily create LLM powered bots over any dataset
⚡ Building applications with LLMs through composability ⚡
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
Adding guardrails to large language models
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
Integrate cutting-edge LLM technology quickly and easily into your app
Operating LLMs in production
Central interface to connect your LLM's with external data