Building the Next Wave of Autonomous Facilities
Industrial, physical AI-based systems are being accelerated across training, simulation and inference.
Robotic facilities result from a culmination of all of these technologies.
Manufacturers like Foxconn or logistics companies like Amazon Robotics can orchestrate teams of autonomous robots to work alongside human workers and monitor factory operations through hundreds or thousands of sensors.
These autonomous warehouses, plants and factories will have digital twins. The digital twins are used for layout planning and optimization, operations simulation and, most importantly, robot fleet software-in-the-loop testing.
Built on Omniverse, “Mega” is a blueprint for factory digital twins that enables industrial enterprises to test and optimize their robot fleets in simulation before deploying them to physical factories. This helps ensure seamless integration, optimal performance and minimal disruption.
Mega lets developers populate their factory digital twins with virtual robots and their AI models, or the brains of the robots. Robots in the digital twin execute tasks by perceiving their environment, reasoning, planning their next motion and, finally, completing planned actions.
These actions are simulated in the digital environment by the world simulator in Omniverse, and the results are perceived by the robot brains through Omniverse sensor simulation.
With sensor simulations, the robot brains decide the next action, and the loop continues, all while Mega meticulously tracks the state and position of every element within the factory digital twin.
This advanced software-in-the-loop testing methodology enables industrial enterprises to simulate and validate changes within the safe confines of the Omniverse digital twin, helping them anticipate and mitigate potential issues to reduce risk and costs during real-world deployment.
SOFTWARE 2.0
The Three Computer Solution: Powering the Next Wave of AI Robotics - Industrial, physical AI-based systems — from humanoids to factories — are being accelerated across training, simulation and inference.
Like chatbots and image generators, this robotics technology learns its skills by analysing enormous amounts of digital data.
ChatGPT marked the big bang moment of generative AI. Answers can be generated in response to nearly any query, helping transform digital work such as content creation, customer service, software development and business operations for knowledge workers.
Physical AI, the embodiment of artificial intelligence in humanoids, factories and other devices within industrial systems, has yet to experience its breakthrough moment.
This has held back industries such as transportation and mobility, manufacturing, logistics and robotics. But that’s about to change thanks to three computers bringing together advanced training, simulation and inference.
Today, software writes software. The world’s computing workloads are shifting from general-purpose computing on CPUs to accelerated computing on GPUs, leaving Moore’s law far behind.
With generative AI, multimodal transformer and diffusion models have been trained to generate responses.
Large language models are one-dimensional, able to predict the next token, in modes like letters or words. Image- and video-generation models are two-dimensional, able to predict the next pixel.
None of these models can understand or interpret the three-dimensional world. And that’s where physical AI comes in.
Physical AI models can perceive, understand, interact with and navigate the physical world with generative AI. With accelerated computing, multimodal physical AI breakthroughs and large-scale physically based simulations are allowing the world to realize the value of physical AI through robots.
A robot is a system that can perceive, reason, plan, act and learn. Robots are often thought of as autonomous mobile robots (AMRs), manipulator arms or humanoids. But there are many more types of robotic embodiments.
In the near future, everything that moves, or that monitors things that move, will be autonomous robotic systems. These systems will be capable of sensing and responding to their environments.
Everything from surgical rooms to data centers, warehouses to factories, even traffic control systems or entire smart cities will transform from static, manually operated systems to autonomous, interactive systems embodied by physical AI.
The Rising Wave of Physical AI
“Physical AIs are models that can understand instructions and autonomously perform complex tasks in the real world,” said Huang, who is extremely optimistic about the extent to which robots will become a part of every industry.
“Everything is going to be robotic,” he said. Huang believes that there would be an entire ecosystem of robots, where all factories will orchestrate robots and those robots will build robotic products. NVIDIA is banking on Omniverse to make this happen
NVIDIA’s Omniverse, a platform designed for real-time 3D design collaboration and simulation, form the basis for digital twins which is crucial for simulation. Digital twins are the virtual replicas of physical objects or systems where robots can be tested to fit into the real world.
Showcasing a wide range of scenarios where robots have been trained on NVIDIA’s Omniverse, Huang spoke about how companies are building robotic warehouses around it.
In digital twins, factory planners optimise floor layout and line configurations and locate optimal camera placements to monitor future operations. Also referred to as a ‘robot gym’, Foxconn developers train and test NVIDIA ISAC AI applications for robotic perception and manipulation in Omniverse digital twins.
Multimodal LLMs have only accelerated the process of robotic training. “Multimodal LLMs are breakthroughs that enable robots to learn, perceive and understand the world around them, and plan how they’ll act,” said Huang.
Clubbing this technique with human demonstrations, robots can acquire the skills needed to interact with the world using gross and fine motor skills.
Robotics Race Continues
The robotics race is only gaining steam with all the recent advancements. Big tech companies have aggressively invested in robotics companies over the last few years. Figure 01, the humanoid built by deep tech robotics company Figure AI, is backed by some of the biggest players like NVIDIA, Microsoft, Jeff Bezos, and others.