diff --git a/README.rst b/README.rst
index 9241e26cdd..55be0e583f 100644
--- a/README.rst
+++ b/README.rst
@@ -13,23 +13,14 @@ Transformer Engine
Latest News
===========
+* [11/2025] `NVIDIA Blackwell Architecture Sweeps MLPerf Training v5.1 Benchmarks `_
+* [11/2025] `Scale Biology Transformer Models with PyTorch and NVIDIA BioNeMo Recipes `_
+* [11/2025] `FP8 Training of Large-Scale RL Models `_
* [09/2025] `Pretraining Large Language Models with NVFP4 `_
* [09/2025] `Native FP8 Mixed Precision Training for Ling 2.0, Open Sourced! `_
* [09/2025] `Faster Training Throughput in FP8 Precision with NVIDIA NeMo `_
* [08/2025] `How we built DeepL's next-generation LLMs with FP8 for training and inference `_
* [08/2025] `NVFP4 Trains with Precision of 16-bit and Speed and Efficiency of 4-bit `_
-* [06/2025] `Floating Point 8: An Introduction to Efficient, Lower-Precision AI Training `_
-* [05/2025] `Advanced Optimization Strategies for LLM Training on NVIDIA Grace Hopper `_
-* [03/2025] `Stable and Scalable FP8 Deep Learning Training on Blackwell | GTC 2025 `_
-* [03/2025] `Measure and Improve AI Workload Performance with NVIDIA DGX Cloud Benchmarking `_
-
-.. image:: docs/examples/comparison-fp8-bf16-training-nvidia-dgx-cloud-benchmarking-performance-explorer.jpg
- :width: 600
- :alt: Comparison of FP8 versus BF16 training, as seen in NVIDIA DGX Cloud Benchmarking Performance Explorer
-
-* [02/2025] `Understanding the Language of Life's Biomolecules Across Evolution at a New Scale with Evo 2 `_
-* [02/2025] `NVIDIA DGX Cloud Introduces Ready-To-Use Templates to Benchmark AI Platform Performance `_
-* [01/2025] `Continued Pretraining of State-of-the-Art LLMs for Sovereign AI and Regulated Industries with iGenius and NVIDIA DGX Cloud `_
`Previous News <#previous-news>`_
@@ -425,6 +416,18 @@ Videos
Previous News
=============
+* [06/2025] `Floating Point 8: An Introduction to Efficient, Lower-Precision AI Training `_
+* [05/2025] `Advanced Optimization Strategies for LLM Training on NVIDIA Grace Hopper `_
+* [03/2025] `Stable and Scalable FP8 Deep Learning Training on Blackwell | GTC 2025 `_
+* [03/2025] `Measure and Improve AI Workload Performance with NVIDIA DGX Cloud Benchmarking `_
+
+.. image:: docs/examples/comparison-fp8-bf16-training-nvidia-dgx-cloud-benchmarking-performance-explorer.jpg
+ :width: 600
+ :alt: Comparison of FP8 versus BF16 training, as seen in NVIDIA DGX Cloud Benchmarking Performance Explorer
+
+* [02/2025] `Understanding the Language of Life's Biomolecules Across Evolution at a New Scale with Evo 2 `_
+* [02/2025] `NVIDIA DGX Cloud Introduces Ready-To-Use Templates to Benchmark AI Platform Performance `_
+* [01/2025] `Continued Pretraining of State-of-the-Art LLMs for Sovereign AI and Regulated Industries with iGenius and NVIDIA DGX Cloud `_
* [11/2024] `Developing a 172B LLM with Strong Japanese Capabilities Using NVIDIA Megatron-LM `_
* [11/2024] `How FP8 boosts LLM training by 18% on Amazon SageMaker P5 instances `_
* [11/2024] `Efficiently train models with large sequence lengths using Amazon SageMaker model parallel `_