Powered by Nvidia and AI, humanoid ambitions take shape

In This Article:

When the Department of Defense commenced its robotics challenge in 2015, the stated goal was to develop ground robots that can aid in disaster recovery with the help of human operators. Each android was given an hour to complete eight tasks, which included driving a car and climbing some stairs.

Nearly a decade later, generative AI is accelerating that learning curve, pushing human-like machines to pick up new tasks in real time.

“The holy grail for us is what we call zero-shot learning, or the ability to show the robot what to do, and it can do it the same way that you do that task,” said Jeff Cardenas, co-founder of Apptronik, an Austin-based robotics maker.

That vision is slowly becoming reality. Last week, OpenAI-backed Figure unveiled the latest iteration of its humanoid robot. The robot is equipped with a vision language model that allows the machine to reason visually and self-correct learned behavior, according to the company’s claims.

And in June, Tesla (TSLA) presented an updated version of its Optimus robot at Tesla’s Investor Day and showed it roaming a factory floor. CEO Elon Musk touted the robot’s potential, saying it had the ability to push the company’s market cap to $25 trillion.

Tesla's humanoid robot is displayed at the World Artificial Intelligence Conference (WAIC) in Shanghai, China, on July 6, 2023. (REUTERS/Aly Song/File Photo)
Tesla's humanoid robot is displayed at the World Artificial Intelligence Conference (WAIC) in Shanghai, China, on July 6, 2023. (REUTERS/Aly Song/File Photo) · REUTERS / Reuters

Robotics have been integrated into factory floors and warehouses to improve efficiencies for years. But current machines in use have largely been limited to moving from point A to point B and tackling a handful of tasks.

Humanoids that can adapt to existing environments have long been seen as the ultimate test if they can work alongside humans in spaces built for them.

“If you want a versatile robot, then having a robot that can retrofit into our environment so you don’t have to change anything seems important,” said Cardenas. “Today, three to six [times] the price of the robot is spent just integrating that robot into a new workflow.”

Nvidia's humanoid ambition

Nvidia (NVDA) is driving rapid development through an ecosystem built specifically for humanoids. It combines high-powered GPUs that process data at high speeds with Nvidia Omniverse, a digital world that allows users to train robots on skills applied in the real world.

The company announced the development of physical AI foundation models earlier this year. Just last month, Nvidia unveiled “NIM Microservices,” software technology that offers generative AI models for different applications, with some models that are able to visually interpret their surroundings in 3D.

“Before, when we used to do an AI model for doing a particular task, we had to train it with specific data for that task … and if you needed that AI model to do something different, you needed to retrain that model,” said Deepu Talla, Nvidia's vice president of Robotics and Edge Computing.