From abe0cb7d8307922c67e0989e02035d9b2d7e8846 Mon Sep 17 00:00:00 2001 From: chunhanyao-stable <156005023+chunhanyao-stable@users.noreply.github.com> Date: Tue, 20 May 2025 07:44:04 -0700 Subject: [PATCH] Update README.md --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 0a937213..86aa302c 100755 --- a/README.md +++ b/README.md @@ -5,7 +5,7 @@ ## News -**April 4, 2025** +**May 20, 2025** - We are releasing **[Stable Video 4D 2.0 (SV4D 2.0)](https://huggingface.co/stabilityai/sv4d2.0)**, an enhanced video-to-4D diffusion model for high-fidelity novel-view video synthesis and 4D asset generation. For research purposes: - **SV4D 2.0** was trained to generate 48 frames (12 video frames x 4 camera views) at 576x576 resolution, given a 12-frame input video of the same size, ideally consisting of white-background images of a moving object. - Compared to our previous 4D model [SV4D](https://huggingface.co/stabilityai/sv4d), **SV4D 2.0** can generate videos with higher fidelity, sharper details during motion, and better spatio-temporal consistency. It also generalizes much better to real-world videos. Moreover, it does not rely on refernce multi-view of the first frame generated by SV3D, making it more robust to self-occlusions.