Hackster is hosting Hackster Holidays, Ep. 7: Livestream & Giveaway Drawing. Watch previous episodes or stream live on Friday!Stream Hackster Holidays, Ep. 7 on Friday!

SMERF Renders 3D Scenes on Smurf-Sized Hardware

SMERF can perform real-time, photorealistic view synthesis for interactive exploration on resource-constrained hardware like smartphones.

Nick Bild
12 months agoMachine Learning & AI
Real-time rendering of scenes on resource-constrained hardware (📷: D. Duckworth et al.)

Neural Radiance Fields (NRFs) are a major breakthrough in computer graphics and computer vision. They offer a novel approach to reconstructing 3D scenes that is unlike traditional methods that rely on explicit geometric representations. NRFs leverage deep neural networks to model the volumetric scene appearance directly, which allows for the generation of high-quality, photo-realistic renderings with high levels of detail and complex lighting effects.

At the heart of NRFs is the idea of learning a function that maps 3D spatial coordinates to radiance values, capturing the appearance of a scene from various viewpoints. The neural network is trained on a dataset of images and corresponding 3D structures, enabling it to generalize and reconstruct novel scenes. This approach is particularly advantageous for scenes with intricate geometry and complex lighting conditions, providing a more accurate representation compared to traditional methods.

The applications of NRFs are very diverse. In the realm of virtual reality and augmented reality, they can enhance the realism of virtual environments, creating immersive and visually stunning experiences. In the field of gaming, this technology enables the creation of highly detailed and dynamic game worlds. Additionally, NRF finds applications in medical imaging, allowing for the reconstruction of detailed 3D models from medical scans, facilitating diagnosis and treatment planning.

Despite its transformative potential, the implementation of NRFs comes with many computational challenges. The algorithms demand substantial processing power and memory, making real-time interaction a significant hurdle. Developers often face a trade-off between rendering quality and performance, as pushing for higher quality requires more computational resources. But recently a team led by researchers at Google DeepMind has released what they call Streamable Memory Efficient Radiance Fields (SMERF). The techniques that they described make it possible to perform real-time, photorealistic view synthesis for the exploration of large scenes. Interestingly, the algorithm can run on even highly resource-constrained platforms, like smartphones.

The tool builds upon an existing view synthesis system called Memory-Efficient Radiance Fields (MERF). The SMERF architecture is hierarchical, consisting of a number of MERF submodels. The submodels are specialized, with each rendering a region of the viewpoints in the scene. Since only a single submodel is required to render the view from a given camera angle, the computational load is much lower than was the case with previous approaches.

While independence between the submodels does greatly increase the algorithm’s efficiency, it also has the unfortunate effect of removing the inductive biases that exist in the current best models that help them to produce such realistic results. To overcome this limitation, the team developed a novel distillation training procedure that provides plenty of extra supervision to the models in the area of color and geometry. This enables SMERF to produce plausible renderings, and maintain stable results even while the camera is in motion.

With SMERF, six degrees of freedom can be explored in a rendered environment in real-time from a web browser running on a commodity laptop or smartphone. Experiments revealed that SMERF renderings are three orders of magnitude faster than those generated by the present state of the art systems. It was also observed that the renderings were of higher quality than other models could produce.

In spite of the many successes achieved by SMERF, there are still some trade-offs to consider. While the runtime operation is quite swift, the model training process is extensive, shifting much of the processing to earlier in the process. Moreover, SMERF requires large amounts of stored data, which can be a problem for the small compute platforms it is intended to run on. But in any case, the advances presented by the researchers are sure to move the field forward in the near future.

Nick Bild
R&D, creativity, and building the next big thing you never knew you wanted are my specialties.
Latest articles
Sponsored articles
Related articles
Latest articles
Read more
Related articles