RETRO (Retrieval-Enhanced Transformer) is a large language model architecture developed by OpenAI that augments transformer models with a retrieval mechanism. Instead of relying solely on learned parameters, RETRO retrieves relevant documents from a large external database during inference, allowing it to ground responses in external knowledge. This design improves factual accuracy, reduces hallucinations, and enables smaller models to perform comparably to much larger ones by leveraging retrieval. The repository provides code and resources for training and evaluating RETRO models, along with infrastructure for integrating retrieval into the transformer pipeline. It includes example configurations, datasets, and utilities for building retrieval-augmented generation systems. RETRO represents an important step toward combining large-scale language modeling with information retrieval, offering researchers a foundation to study hybrid approaches for scaling AI responsibly.
Features
- Implements the Retrieval-Enhanced Transformer (RETRO) architecture
- Integrates external document retrieval into model inference
- Improves factual grounding and reduces hallucinations
- Enables smaller models to match performance of larger ones
- Provides code and configurations for training and evaluation
- Supports research into retrieval-augmented generation systems