Creating 3D models is a way to make scenes more realistic and immersive compared to 2D images. It allows viewers to explore and interact with the scene from different angles, giving them a better sense of the spatial layout and depth of information.
These 3D models are crucial for virtual reality (VR) and augmented reality (AR) applications. They can overlay digital information onto the real world (AR) or create entirely virtual environments (VR), enhancing user experiences in gaming, education, training, and various industries.
Neural Radiance Fields (NeRFs) is a computer vision technique used for 3D scene reconstruction and rendering. It treats a scene as a 3D volume where each point in the volume has a corresponding color (radiance) and density. A neural network learns to predict the color and density of each point based on 2D images taken from different viewpoints.
NeRFs have several applications, such as view synthesis and depth estimation. However, learning from multiview images comes with inherent uncertainties. Current methods to quantify these uncertainties are either heuristic or computationally demanding. Researchers from Google DeepMind, Adobe Research, and the University of Toronto have introduced a new technique called BayesRays.
BayesRays is a framework that evaluates uncertainty in pretrained NeRFs without modifying the training process. By adding a volumetric uncertainty field using spatial perturbations and a Bayesian Laplace approximation, they were able to overcome the limitations of NeRFs. The Bayesian Laplace approximation is a mathematical method that approximates complex probability distributions with simpler multivariate Gaussian distributions.
The calculated uncertainties from BayesRays are statistically meaningful and can be rendered as additional color channels. This method outperforms previous works on key metrics like correlation to reconstructed depth errors. It can be used to quantify the uncertainty of any pretrained NeRFs, regardless of their architecture, in real time.
The researchers explain that their method was inspired by using volumetric fields to model 3D scenes. Similarly, volumetric deformation fields are often used to manipulate implicitly represented objects. Their approach is also similar to photogrammetry, where reconstructing uncertainty is modeled by placing Gaussian distributions on spatial positions.
However, it’s important to note that their algorithm is specific to quantifying the uncertainty of NeRFs and cannot be easily applied to other frameworks. They plan to further develop a deformation-based Laplace approximation for more recent spatial representations like 3D Gaussian splatting.
The researchers have introduced a new technique called BayesRays to evaluate uncertainty in pretrained NeRFs without modifying the training process. This method overcomes the limitations of NeRFs in quantifying uncertainties and outperforms previous works on key metrics. While the algorithm is currently limited to NeRFs, the researchers plan to expand its application to other frameworks in the future.
You can read the paper and learn more about the project here. All credit for this research goes to the researchers involved in the project.
If you’re interested in staying up to date with the latest AI research news, cool AI projects, and more, don’t forget to join our ML subreddithere, our Facebook communityhere, our Discord channel here, and subscribe to our email newsletterhere.
If you enjoy our work, you’ll love our newsletter. Subscribe to stay updated on the latest AI news and research.