You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This roadmap outlines the key features, enhancements, and improvements planned for Megatron Core. This is a tentative roadmap and subject to change based on community feedback and priorities.
For detailed information on past releases, see the Changelog.For MoE-specific roadmap, see MoE Roadmap #1729.
Future Releases
Parallelism
Megatron FSDP Enhancements
Fully Sharded Hybrid Sharded Data Parallel (FSHSDP) refinements
Non-MCore offload support
PyTorch Expert Parallel Pipeline Parallel research
MiMo is the canonical early-fusion multimodal architecture in Megatron Core, enabling modular vision, audio, and video encoders to plug into a shared LLM with out-of-the-box distributed training support
Enable independent nD parallelism (TP/DP/CP/EP/PP) for each module (encoders and LLM)
Colocated training (encoder and LLM share GPUs)
Non-colocated training (encoders and LLM on disjoint GPU sets)
Configurable Megatron FSDP communication double buffering - Improved FSDP communication efficiency and throughput with persistent param/grad collective buffers
This roadmap outlines the key features, enhancements, and improvements planned for Megatron Core. This is a tentative roadmap and subject to change based on community feedback and priorities.
For detailed information on past releases, see the Changelog. For MoE-specific roadmap, see MoE Roadmap #1729.
Future Releases
Parallelism
Performance
Model Support
Inference
Ease of Use
Precision
Multimodal
Infrastructure & Ecosystem
Finetuning
megatron.core.datasetsSFTDataset andget_batch*functions refactor #3542v0.16 Highlights (Released February 2026)
Parallelism
Performance & Memory
Inference
Model Support
Megatron FSDP
Ease of Use
Precision
RL
rl_offload_kv_cache_during_trainingto offload KV cache to CPU while retaining fixed virtual address #3048v0.15 Highlights (Released December 2025)
Parallelism
Performance
Model Support & Training
Inference
RL
Ease of Use
How to Provide Feedback
We welcome community input on prioritization! Please:
enhancementlabelCredits
This roadmap reflects the collective efforts of NVIDIA and our collaborators.
Last updated: March 2026