The new tools and programs revealed by the company promise to enable easier and faster 3D content creation.
During the SIGGRAPH 2022 conference, NVIDIA introduced a new research and several new tools that ease and speed up the 3D content creation process. According to the team, the newly-revealed tools are powered by neural graphics, a new technology that combines AI and graphics to create an accelerated graphics pipeline that learns from data. These SDKs and research contribute to each stage of the content creation pipeline, including 3D content creation, physics and animation, and experience.
The list of new SDKs revealed by the company includes NeuralVDB, a ground-breaking update to industry standard OpenVDB, Kaolin Wisp, a Pytorch library establishing a framework for neural fields research, 3D MoMa, a new inverse rendering pipeline that allows users to quickly import a 2D object into a graphics engine, and more.
"Neural graphics is a new field intertwining AI and graphics to create an accelerated graphics pipeline that learns from data. Integrating AI enhances results, helps automate design choices, and provides new, yet to be imagined opportunities for artists and creators. Neural graphics will redefine how virtual worlds are created, simulated, and experienced by users," comments the team.
And here's the list of NVIDIA's new SDKs and research:
- Kaolin Wisp: An addition to Kaolin, a PyTorch library enabling faster 3D deep learning research by reducing the time needed to test and implement new techniques from weeks to days. Kaolin Wisp is a research-oriented library for neural fields, establishing a common suite of tools and a framework to accelerate new research in neural fields.
- Instant Neural Graphics Primitives: A new approach to capturing the shape of real-world objects, and the inspiration behind NVIDIA Instant NeRF, an inverse rendering model that turns a collection of still images into a digital 3D scene. This technique and associated GitHub code accelerate the process by up to 1,000x.
- 3D MoMa: A new inverse rendering pipeline that allows users to quickly import a 2D object into a graphics engine to create a 3D object that can be modified with realistic materials, lighting, and physics.
- GauGAN360: The next evolution of NVIDIA GauGAN, an AI model that turns rough doodles into photorealistic masterpieces. GauGAN360 generates 8K, 360-degree panoramas that can be ported into Omniverse scenes.
- Omniverse Avatar Cloud Engine (ACE): A new collection of cloud APIs, microservices, and tools to create, customize and deploy digital human applications. ACE is built on NVIDIA’s Unified Compute Framework, allowing developers to seamlessly integrate core NVIDIA AI technologies into their avatar applications.
- NeuralVDB: A groundbreaking improvement on OpenVDB, the current industry standard for volumetric data storage. Using machine learning, NeuralVDB introduces compact neural representations, dramatically reducing memory footprint to allow for higher-resolution 3D data.
- Omniverse Audio2Face: An AI technology that generates expressive facial animation from a single audio source. It’s useful for interactive real-time applications and as a traditional facial animation authoring tool.
- ASE: Animation Skills Embedding: An approach enabling physically simulated characters to act in a more responsive and life-like manner in unfamiliar situations. It uses deep learning to teach characters how to respond to new tasks and actions.
- TAO Toolkit: A framework to enable users to create an accurate, high-performance pose estimation model, which can evaluate what a person might be doing in a scene using computer vision much more quickly than current methods.
- Image Features Eye Tracking: A research model linking the quality of pixel rendering to a user’s reaction time. By predicting the best combination of rendering quality, display properties, and viewing conditions for the least latency, it will allow for better performance in fast-paced, interactive computer graphics applications such as competitive gaming.
- Holographic Glasses for Virtual Reality: A collaboration with Stanford University on a new VR glasses design that delivers full-color 3D holographic images in a groundbreaking 2.5-mm-thick optical stack.