Our objective is to leverage a differentiable radiance field e.g. NeRF to reconstruct detailed 3D surfaces in addition to producing the standard novel view renderings. RaNeuS adaptively adjusts the regularization on the signed distance field so that unsatisfying rendering rays won’t enforce strong Eikonal regularization which is ineffective, and allow the gradients from regions with well-learned radiance to effectively back-propagated to the SDF. Consequently, balancing the two objectives in order to generate accurate and detailed surfaces.
Applications
Urban area reconstruction
Given a set of images shot by cameras mounted on drones, an urban area is represented by the mesh extracted by marching cube from a leared signed distance field (SDF) which is optimized by RaNeuS. | |
---|---|
Trained with 200 images shot from about 400m |
Validations
Advantages against NeuS
Comparison of our mesh to Neus 2, focusing on some important details on the bonsai dataset that our method was able to reconstruct while NeuS 2 missed. | |
---|---|
Bonsai in MipNeRF360 |
Boosting other approaches
Geometric reconstruction comparison evaluated on the Mip-NeRF 360 dataset
Cite
If you find this work useful in your research, please cite:
@inproceedings{wang2023raneus,
title={RaNeuS: Ray-adaptive Neural Surface Reconstruction},
author={Wang, Yida and Tan, David and Tombari, Federico and Navab, Nassir},
booktitle={Proceedings of the IEEE/CVF International Conference on 3D Vision},
year={2023}
}