Boese0601 / RC-MVSNet

[ECCV 2022] RC-MVSNet: Unsupervised Multi-View Stereo with Neural Rendering

Home Page:https://boese0601.github.io/rc-mvsnet/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Question about code

DongyangHuLi opened this issue · comments

Hi~ Thanks for your open code!
I have read your paper and code, and one key step is to do ray marching in implicit neural volume for volume rendering, where the sampling points on the ray are converted to ndc space; It is well known that in Nerf, the ray marching is completed in the world coordinate system. Because it's converted to the world coordinate system by the camera pose.
I am confused about why can the ray marching of the implicit neural volume be done in the ndc space. Is the implicit neural volume in the ndc space after regularization? What's the physical meaning behind it? Since I can't seem to find the code to convert the final result from the ndc coordinate system to the world coordinate system, can you help me point it out?
BTW, If the implicit neural volume is indexed in NDC space, shouldn't it be trilinear interpolation? Why bilinear interpolation? Or is this a representation of pytorch :)

features = F.grid_sample(volume_feature, grid, align_corners=True, mode='bilinear')[:,:,0].permute(2,3,0,1).squeeze()#, padding_mode="border"

Looking forward to your reply!