Leverage light field technology and create visual digital twins in minutes!
Under terms like Neural Radiance Fields (NeRFs), Light-Fields, Neural Graphics Primitives, and Integral Photography, a new set of algorithms and technologies are revolutionizing graphics.
Automated photograph capture and Deep-Learning optimization are enabling the creation of visual digital twins in minutes, with no manual modelling of products. This gives designers, marketers, and visual merchandisers, the power of “3D” without the painstaking process of creating it.
Photorealistic renderings
NeRF uses a neural network to learn the relationships between the appearance of a scene and the 3D structure of the objects within it, allowing it to generate highly realistic lifelike “3D images” from a single 2D image or sequence of images.
Highly efficient and scalable
NeRF can be trained on large datasets, allowing it to process large amounts of information quickly and accurately. As it can process data in parallel, NeRF is faster than traditional rendering techniques.
Interactive product experiences
Just like 3D renderings, Light field display enables an engaging, immersive, and innovative 360° product experience (3D for e-commerce, AR features).
Step 1:
Light Field capturing gear
SmartPixels has developed several robotic photography machines, based on turn tables, motorized cameras, controllable lighting, and background. Once placed at the center of the machine, an object or product is photographed thousands of times, from every angle.
The next step is detouring the photos. As the lighting and background are accurately controlled, this process is fully automated using regular computer vision algorithms, or deep learning-based object recognition and detouring for more complex cases (transparency, high shininess).
This step, common between our traditional computer graphics pipeline and our neural pipeline, is highly optimized.
Step 2:
NVIDIA Instant NeRF - Optimization and Training
In 2022, NVIDIA has introduced a highly optimized Nerf-like model to learn and render computational light fields: Instant-NGP. As a member of the NVIDIA Inception program for start-ups, SmartPixels has signed a licensing agreement on this technology.
Instant-NGP allows us to train the neural primitive in minutes and render it at 60 images per second. Instant-NGP leverages the latest advances in NVIDIA GPUs for optimal performance. While optimal, Instant-NGP is a research-oriented project and does not provide a full 3D production pipeline for at-scale content delivery.
Step 3:
Neural Rendering setup and editability
SmartPixels has leveraged Unity to create a cloud rendering and experience platform tailored for luxury fashion. To benefit from years of existing development and integration into the fashion ecosystem, we have integrated the ability to render an NVIDIA NGP inside the Unity engine.
By integrating Light Fields inside Unity, we can create hybrid renderings, mixing 3D backgrounds, light-field products, or vice-versa.
The key features of this integration are:
• Fully automatic capture of the light field, per above, optimization and training thanks to NVIDIA NGP.
• One-click import into Unity editor.
• Occlusion (the light field hides objects behind it and is hidden by the foreground).
• Shadows: The light field casts shadows.
• Transparency.
• Reflections: The Light-field casts reflections.
1) Improve the visual appeal and interactivity of your e-commerce catalog
Light-fields give you the “omnidirectional viewpoint” (free camera motion) of “3D”. It’s an integral photography: the photo of all photos, you can spin the object captured, zoom, and pan.
In the context of e-commerce, light field rendering can be used to create more realistic and interactive product images for online catalogues. For example, a retailer could use light field rendering to capture detailed 3D images of a product from multiple angles, allowing customers to view the product from different perspectives and see subtle details that may not be visible in a traditional 2D image.
2) Enhanced realism and interactivity of AR experiences
Light field rendering can be used to create more realistic and immersive Augmented Reality experiences. For example, light field rendering can be used to display virtual objects in AR with realistic lighting and reflections, making them look like they are part of the real world. Light field rendering can also allow users to interact with virtual objects in a more natural way, such as by changing the focus of an image to see different parts of an object more clearly.
People also ask about...
When using a computer to render a light field, you can spin the object captured, zoom, and pan: “everything feels like 3D”, so the comparison/confusion with traditional 3D graphics is easy.
Let’s look at a few similarities and differences between “traditional computer-generated images” and “neural graphics primitives”:
Traditional | Neural Graphics Primitives | |
Arbitrary viewpoint and camera: look around in space freely. | Yes | Yes |
Compose with other objects, add a background, compute transparency, occlusion, shadows, reflections | Yes | Yes (Advanced!) |
The shape, material attributes, textures and light positions must be modelled individually and manually by graphics artists using tools like 3D scanners and material scanners. | Yes | No. Fully automatic. |
The shape of the object can be edited. | Yes | No |
The colors and materials can be edited. | Yes | No |
The lighting can be edited. | Yes | Partially. |
The “asset” is usable in 3D worlds such as video games, metaverse, CAD, etc. | Yes, increasingly. | Not yet. |