From 2c63219cfef972c3846c14a2444c1d75d913f8fd Mon Sep 17 00:00:00 2001 From: Santosh Bhavani Date: Fri, 9 Jan 2026 10:04:17 -0800 Subject: [PATCH 1/2] Move older news to Previous Signed-off-by: Santosh Bhavani --- README.rst | 24 ++++++++++++------------ 1 file changed, 12 insertions(+), 12 deletions(-) diff --git a/README.rst b/README.rst index 9241e26cdd..24118e0165 100644 --- a/README.rst +++ b/README.rst @@ -18,18 +18,6 @@ Latest News * [09/2025] `Faster Training Throughput in FP8 Precision with NVIDIA NeMo `_ * [08/2025] `How we built DeepL's next-generation LLMs with FP8 for training and inference `_ * [08/2025] `NVFP4 Trains with Precision of 16-bit and Speed and Efficiency of 4-bit `_ -* [06/2025] `Floating Point 8: An Introduction to Efficient, Lower-Precision AI Training `_ -* [05/2025] `Advanced Optimization Strategies for LLM Training on NVIDIA Grace Hopper `_ -* [03/2025] `Stable and Scalable FP8 Deep Learning Training on Blackwell | GTC 2025 `_ -* [03/2025] `Measure and Improve AI Workload Performance with NVIDIA DGX Cloud Benchmarking `_ - -.. image:: docs/examples/comparison-fp8-bf16-training-nvidia-dgx-cloud-benchmarking-performance-explorer.jpg - :width: 600 - :alt: Comparison of FP8 versus BF16 training, as seen in NVIDIA DGX Cloud Benchmarking Performance Explorer - -* [02/2025] `Understanding the Language of Life's Biomolecules Across Evolution at a New Scale with Evo 2 `_ -* [02/2025] `NVIDIA DGX Cloud Introduces Ready-To-Use Templates to Benchmark AI Platform Performance `_ -* [01/2025] `Continued Pretraining of State-of-the-Art LLMs for Sovereign AI and Regulated Industries with iGenius and NVIDIA DGX Cloud `_ `Previous News <#previous-news>`_ @@ -425,6 +413,18 @@ Videos Previous News ============= +* [06/2025] `Floating Point 8: An Introduction to Efficient, Lower-Precision AI Training `_ +* [05/2025] `Advanced Optimization Strategies for LLM Training on NVIDIA Grace Hopper `_ +* [03/2025] `Stable and Scalable FP8 Deep Learning Training on Blackwell | GTC 2025 `_ +* [03/2025] `Measure and Improve AI Workload Performance with NVIDIA DGX Cloud Benchmarking `_ + +.. image:: docs/examples/comparison-fp8-bf16-training-nvidia-dgx-cloud-benchmarking-performance-explorer.jpg + :width: 600 + :alt: Comparison of FP8 versus BF16 training, as seen in NVIDIA DGX Cloud Benchmarking Performance Explorer + +* [02/2025] `Understanding the Language of Life's Biomolecules Across Evolution at a New Scale with Evo 2 `_ +* [02/2025] `NVIDIA DGX Cloud Introduces Ready-To-Use Templates to Benchmark AI Platform Performance `_ +* [01/2025] `Continued Pretraining of State-of-the-Art LLMs for Sovereign AI and Regulated Industries with iGenius and NVIDIA DGX Cloud `_ * [11/2024] `Developing a 172B LLM with Strong Japanese Capabilities Using NVIDIA Megatron-LM `_ * [11/2024] `How FP8 boosts LLM training by 18% on Amazon SageMaker P5 instances `_ * [11/2024] `Efficiently train models with large sequence lengths using Amazon SageMaker model parallel `_ From 1d4e7f0ec6149bbe749de37e12e59610dbfd6648 Mon Sep 17 00:00:00 2001 From: Santosh Bhavani Date: Fri, 9 Jan 2026 10:30:46 -0800 Subject: [PATCH 2/2] Add Nov 2025 news entries Signed-off-by: Santosh Bhavani --- README.rst | 3 +++ 1 file changed, 3 insertions(+) diff --git a/README.rst b/README.rst index 24118e0165..55be0e583f 100644 --- a/README.rst +++ b/README.rst @@ -13,6 +13,9 @@ Transformer Engine Latest News =========== +* [11/2025] `NVIDIA Blackwell Architecture Sweeps MLPerf Training v5.1 Benchmarks `_ +* [11/2025] `Scale Biology Transformer Models with PyTorch and NVIDIA BioNeMo Recipes `_ +* [11/2025] `FP8 Training of Large-Scale RL Models `_ * [09/2025] `Pretraining Large Language Models with NVFP4 `_ * [09/2025] `Native FP8 Mixed Precision Training for Ling 2.0, Open Sourced! `_ * [09/2025] `Faster Training Throughput in FP8 Precision with NVIDIA NeMo `_