SRes-NeRF: Improved Neural Radiance Fields for Realism and Accuracy of Specular Reflections

  • Shufan Dai
  • , Yangjie Cao*
  • , Pengsong Duan
  • , Xianfu Chen
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference article in proceedingsScientificpeer-review

Abstract

The Neural Radiance Fields (NeRF) is a popular view synthesis technique that represents a scene using a multilayer perceptron (MLP) combined with classic volume rendering and uses positional encoding techniques to increase image resolution. Although it can effectively represent the appearance of a scene, they often fail to accurately capture and reproduce the specular details of surfaces and require a lengthy training time ranging from hours to days for a single scene. We address this limitation by introducing a representation consisting of a density voxel grid and an enhanced MLP for a complex view-dependent appearance and model acceleration. Modeling with explicit and discretized volume representations is not new, but we propose Swish Residual MLP (SResMLP). Compared with the standard MLP+ReLU network, the introduction of layer scale module allows the shallow information of the network to be transmitted to the deep layer more accurately, maintaining the consistency of features. Introduce affine layers to stabilize training, accelerate convergence and use the Swish activation function instead of ReLU. Finally, an evaluation of four inward-facing benchmarks shows that our method surpasses NeRF’s quality, it only takes about 18 min to train from scratch for a new scene and accuracy capture the specular details of the scene surface. Excellent performance even without positional encoding.

Original languageEnglish
Title of host publicationMultiMedia Modeling: 29th International Conference, MMM 2023
Subtitle of host publicationProceedings
EditorsDuc-Tien Dang-Nguyen, Cathal Gurrin, Alan F. Smeaton, Martha Larson, Stevan Rudinac, Minh-Son Dao, Christoph Trattner, Phoebe Chen
PublisherSpringer
Pages306-317
ISBN (Electronic)978-3-031-27077-2
ISBN (Print)978-3-031-27076-5
DOIs
Publication statusPublished - 2023
MoE publication typeA4 Article in a conference publication
Event29th International Conference on MultiMedia Modeling, MMM 2023 - Bergen, Norway
Duration: 9 Jan 202312 Jan 2023

Publication series

SeriesLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume13833 LNCS
ISSN0302-9743

Conference

Conference29th International Conference on MultiMedia Modeling, MMM 2023
Country/TerritoryNorway
CityBergen
Period9/01/2312/01/23

Funding

Acknowledgements. The authors would like to thank the School of Cyber Science and Engineering, Zhengzhou University, the Zhengzhou City Collaborative Innovation Major Project, and the GPU support provided by the 3DCV lab.

Keywords

  • 3D deep learning
  • Image-based rendering
  • Scene representation
  • Spectral bias
  • View synthesis
  • Volume rendering

Fingerprint

Dive into the research topics of 'SRes-NeRF: Improved Neural Radiance Fields for Realism and Accuracy of Specular Reflections'. Together they form a unique fingerprint.

Cite this