-
Notifications
You must be signed in to change notification settings - Fork 42
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Epipolar cost volume and depth prediction #42
Comments
Same concerns. |
Hi, @chenjiajie9811, thanks for your interest in our work. Our project indeed assumes that there are significant overlaps between the input views, and this is actually how the data is structured during testing (see the index-generating code at here). For parts that have no overlap, MVSplat relies on the following UNet to help propagate the matching information (see Cost volume refinement in Sec. 3.1 of the paper). However, this solution is intuitive and may only somewhat ease the issue. It is a promising future direction to consider improving accuracy in those non-overlap regions. Hi, @fangchuan. Your findings are pretty interesting. May I know how do you confirm that "the point cloud scale is almost the same as the real world one"? I remember there is no ground-truth 3D data for the RE10K or ACID dataset. In fact, MVSplat is not intended to predict real-scale depth (aka metric depth), which should be quite difficult to achieve without additional regularization. Instead, MVSplat merely aims to predict a relative depth bounded by the predefined near and far planes. |
Hi there,
thank you for your great work! I have a question regarding the concept of using epipolar cost volume for the depth prediction.
It is reasonable to use this method when we consider the region where two images have overlap, we can search along the epipolar line and find the position with the highest similarity as the depth prediction.
What if we are considering the pixels on the region without overlap, we won't find any truth correspondences on the epipolar line, and the cost volume we constructed might have equally low similarities, how is it possible for the network to learn the depth in this case?
I am a little bit confused and looking forward to your reply.
Regards
The text was updated successfully, but these errors were encountered: