ModernBERT is an open-source research project that modernizes the classic BERT encoder architecture by incorporating recent advances in transformer design, training techniques, and efficiency improvements. The goal of the project is to bring BERT-style models up to date with the capabilities of modern large language models while preserving the strengths of bidirectional encoder architectures used for tasks such as classification, retrieval, and semantic search. ModernBERT introduces architectural improvements that enhance both training efficiency and inference performance, making the model more suitable for modern large-scale machine learning pipelines. The repository also includes FlexBERT, a modular framework that allows developers to experiment with different encoder building blocks and configurations when constructing new models.
Features
- Modernized BERT-style encoder architecture with updated transformer components
- FlexBERT modular framework for customizable encoder building blocks
- Support for long-context inputs and large document processing
- Configuration-driven model design using structured YAML files
- Training pipelines for large-scale language and code datasets
- Optimized attention mechanisms and inference performance improvements