Megatron-LM is a GPU-optimized deep learning framework from NVIDIA designed to train extremely large transformer-based language models efficiently at scale. The repository provides both a reference training implementation and Megatron Core, a composable library of high-performance building blocks for custom large-model pipelines. It supports advanced parallelism strategies including tensor, pipeline, data, expert, and context parallelism, enabling training across massive multi-GPU and multi-node clusters. The framework includes mixed-precision training options such as FP16, BF16, FP8, and FP4 to maximize performance and memory efficiency on modern hardware. Megatron-LM is widely used in research and industry for pretraining GPT-, BERT-, T5-, and multimodal-style models, with tooling for checkpoint conversion and interoperability with Hugging Face. Overall, it is a production-grade system for organizations pushing the limits of large-scale language model training.

Features

  • GPU-optimized transformer training
  • Advanced parallelism strategies
  • Mixed precision training support
  • Composable Megatron Core library
  • Hugging Face checkpoint conversion
  • Multi-node scalable training pipelines

Project Samples

Project Activity

See All Activity >

Categories

Research

License

MIT License

Follow Megatron-LM

Megatron-LM Web Site

Other Useful Business Software
MongoDB Atlas runs apps anywhere Icon
MongoDB Atlas runs apps anywhere

Deploy in 115+ regions with the modern database for every enterprise.

MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
Start Free
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of Megatron-LM!

Additional Project Details

Programming Language

Python

Related Categories

Python Research Software

Registered

2026-02-25