TY - GEN
T1 - S2NeRF
T2 - 17th International Conference on Intelligent Robotics and Applications, ICIRA 2024
AU - Zhang, Zhihong
AU - Wang, Wenjun
AU - Qi, Dexin
AU - Mei, Xuesong
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
PY - 2025
Y1 - 2025
N2 - Neural volume rendering methods, especially NeRF, have demonstrated remarkable performance in novel view synthesis. However, NeRF relies solely on image data and lacks explicit geometric information, necessitating a large number of posed images and a computationally intensive ray sampling strategy to learn accurate scene representations. This poses challenges and may result in incomplete or locally optimal scene geometry when views are sparse or incomplete, as the limited views may not provide sufficient constraints to determine a unique geometry solution for complex scenes. Meanwhile, sparse point clouds provide an attractive source of scene information, especially for geometry, to complement images in neural scene representations, particularly when input views are sparse. To overcome these limitations, we propose S2NeRF, a novel Neural Radiance Field that simultaneously incorporates features from both point clouds and images for volume rendering. Specifically, S2NeRF extracts patch-wise point features from point clouds and ray-wise image features from adjacent views. Then the scene feature volume is constructed by implicitly fusing these point and image features through self-attention. Finally, the volume feature is utilized to render novel views of the scene. Experimental results on the challenging TartanAir dataset demonstrate that, thanks to the integration of feature volume from point clouds and images, S2NeRF achieves state-of-the-art performance in novel view synthesis.
AB - Neural volume rendering methods, especially NeRF, have demonstrated remarkable performance in novel view synthesis. However, NeRF relies solely on image data and lacks explicit geometric information, necessitating a large number of posed images and a computationally intensive ray sampling strategy to learn accurate scene representations. This poses challenges and may result in incomplete or locally optimal scene geometry when views are sparse or incomplete, as the limited views may not provide sufficient constraints to determine a unique geometry solution for complex scenes. Meanwhile, sparse point clouds provide an attractive source of scene information, especially for geometry, to complement images in neural scene representations, particularly when input views are sparse. To overcome these limitations, we propose S2NeRF, a novel Neural Radiance Field that simultaneously incorporates features from both point clouds and images for volume rendering. Specifically, S2NeRF extracts patch-wise point features from point clouds and ray-wise image features from adjacent views. Then the scene feature volume is constructed by implicitly fusing these point and image features through self-attention. Finally, the volume feature is utilized to render novel views of the scene. Experimental results on the challenging TartanAir dataset demonstrate that, thanks to the integration of feature volume from point clouds and images, S2NeRF achieves state-of-the-art performance in novel view synthesis.
KW - Computer Vision
KW - Machine Learning
KW - Volume Rendering
UR - https://www.scopus.com/pages/publications/85218471501
U2 - 10.1007/978-981-96-0774-7_8
DO - 10.1007/978-981-96-0774-7_8
M3 - 会议稿件
AN - SCOPUS:85218471501
SN - 9789819607730
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 101
EP - 116
BT - Intelligent Robotics and Applications - 17th International Conference, ICIRA 2024, Proceedings
A2 - Lan, Xuguang
A2 - Mei, Xuesong
A2 - Jiang, Caigui
A2 - Zhao, Fei
A2 - Tian, Zhiqiang
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 31 July 2024 through 2 August 2024
ER -