aipixel / GPS-Gaussian

[CVPR 2024 Highlight] The official repo for “GPS-Gaussian: Generalizable Pixel-wise 3D Gaussian Splatting for Real-time Human Novel View Synthesis”

Home Page:https://shunyuanzheng.github.io/GPS-Gaussian

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

3DoF or 6DoF?

leviome opened this issue · comments

thanks for the nice work!

In my understanding, GPS-Gaussian leverage two adjacent views to synthesize novel view, if we want a view that is far from all source views, will GPS-Gaussian still work?

Thank you for your interest!

Our method still works if the mentioned novel viewpoint refers to a closer or a more distant one. However, the performance will degrade when synthesizing pitching viewpoints. This drawback is mainly caused by self-occlusion, which means it remains high-quality rendering in unoccluded regions.

thanks, got it !

The occluded regions may be reduced by involving more (>2) referring viewpoints, will GPS-Gaussian support it ?

By the way, I find that GPS-Gaussian doesn't involve any human-related priors, so it looks feasible in non-human-object NVS. Why did you title GPS-Gaussian as Human NVS?

thanks, got it !

The occluded regions may be reduced by involving more (>2) referring viewpoints, will GPS-Gaussian support it ?

Directly aggregating more source views is not trivial. You can try to replace the binocular depth estimator with an MVS module.

By the way, I find that GPS-Gaussian doesn't involve any human-related priors, so it looks feasible in non-human-object NVS. Why did you title GPS-Gaussian as Human NVS?

Yes, GPS-Gaussian is not limited to Human NVS. However, the real-time efficiency is highlighted in our method. Validation on static objects cannot showcase the main highlight of GPS-Gaussian while NVS on moving objects are not general applications. So our method can be widely used in Human-centered scenarios such as holographic communication systems.

Got it, thanks for your kind reply!