Skip to content

Commit 9a4d8d8

Browse files
Vaibhavs10pcuenca
andauthored
fix typo (#2749)
* add nvidia physical AI blogpost * up * Apply suggestions from code review * Update _blog.yml * Update _blog.yml * Update nvidia-physical-ai.md --------- Co-authored-by: Pedro Cuenca <[email protected]>
1 parent 17e9e63 commit 9a4d8d8

File tree

3 files changed

+133
-1
lines changed

3 files changed

+133
-1
lines changed

_blog.yml

Lines changed: 15 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5693,4 +5693,18 @@
56935693
tags:
56945694
- storage
56955695
- infrastructure
5696-
- content addressed store
5696+
- content addressed store
5697+
5698+
- local: nvidia-physical-ai
5699+
title: "NVIDIA's GTC 2025 Announcement for Physical AI Developers: New Open Models and Datasets"
5700+
author: mingyuliutw
5701+
guest: true
5702+
org: nvidia
5703+
thumbnail: /blog/assets/nvidia-physical-ai/thumbnail.png
5704+
date: March 18, 2025
5705+
tags:
5706+
- robotics
5707+
- ai
5708+
- datasets
5709+
- community
5710+
- research
79.5 KB
Loading

nvidia-physical-ai.md

Lines changed: 118 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,118 @@
1+
---
2+
title: "NVIDIA's GTC 2025 Announcement for Physical AI Developers: New Open Models and Datasets"
3+
thumbnail: /blog/assets/nvidia-physical-ai/thumbnail.png
4+
authors:
5+
- user: mingyuliutw
6+
guest: true
7+
org: nvidia
8+
- user: hannamao
9+
guest: true
10+
org: nvidia
11+
- user: jwgu
12+
guest: true
13+
org: nvidia
14+
- user: PranjaliJoshi
15+
guest: true
16+
org: nvidia
17+
- user: asawareeb
18+
guest: true
19+
org: nvidia
20+
---
21+
22+
| ![Humanoid pick and place](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/humanoid.gif) |
23+
| :--: |
24+
| *NVIDIA Isaac GR00T N1 used in object manipulation.* |
25+
26+
At its annual GTC conference, NVIDIA has unveiled a trio of groundbreaking open-source releases aimed at accelerating physical AI development. Release of a new suite of world foundation models (WFMs) with multicontrols called **Cosmos Transfer**, a highly curated **Physical AI Dataset**, and the first open model for general humanoid reasoning called **NVIDIA Isaac GR00T N1** - represent a significant leap forward in physical AI technology, offering developers powerful tools and resources to advance robotics systems, and enhance autonomous vehicle technology.
27+
28+
## New World Foundation Model - Cosmos Transfer
29+
30+
[Cosmos Transfer](https://huggingface.co/collections/nvidia/cosmos-transfer1-67c9d328196453be6e568d3e), the latest addition to NVIDIA's Cosmos™ world foundation models (WFMs), introduces a new level of control and accuracy in generating virtual world scenes.
31+
32+
Available in 7 billion parameter size, the model utilizes multicontrols to guide the generation of high-fidelity world scenes from structural inputs, ensuring precise spatial alignment and scene composition.
33+
34+
### How it works
35+
36+
The model is built by training individual ControlNets separately for each sensor modality used to capture the simulated world.
37+
38+
<div style="display: grid; grid-template-columns: repeat(2, 1fr); gap: 10px;">
39+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/3D_bounding_box_map.gif" alt="3D bounding box map input" style="width: 100%;">
40+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/trajectory_map.gif" alt="Trajectory map input" style="width: 100%;">
41+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/depth_map.gif" alt="Depth map input" style="width: 100%;">
42+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/segmentation_map.gif" alt="Segmentation map input" style="width: 100%;">
43+
</div>
44+
<p style="text-align: center; font-style: italic;">Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.</p>
45+
46+
| ![Cosmos Transfer input types](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/3D_bounding_box_map.gif) |
47+
| :--: |
48+
| *Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.* |
49+
50+
| ![Cosmos Transfer input types](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/trajectory_map.gif) |
51+
| :--: |
52+
| *Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.* |
53+
54+
| ![Cosmos Transfer input types](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/depth_map.gif) |
55+
| :--: |
56+
| *Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.* |
57+
58+
| ![Cosmos Transfer input types](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/segmentation_map.gif) |
59+
| :--: |
60+
| *Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.* |
61+
62+
63+
- At inference time, developers can use various input types, including structured visual or geometric data such as segmentation maps, depth maps, edge maps, human motion keypoints, LiDAR scans, trajectories, HD maps, and 3D bounding boxes to guide the output.
64+
- The control signals from each control branch are multiplied by their corresponding adaptive spatiotemporal control maps and then summed before being added to the transformer blocks of the base model.
65+
- The generated output is photorealistic video sequences with controlled layout, object placement, and motion. Developers can control the output in multiple ways, such as preserving structure and appearance or allowing appearance variations while maintaining structure.
66+
67+
<div style="display: flex; justify-content: space-between;">
68+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/output_1.gif" alt="Output 1" style="width: 30%;">
69+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/output_2.gif" alt="Output 2" style="width: 30%;">
70+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/output_3.gif" alt="Output 3" style="width: 30%;">
71+
</div>
72+
<p style="text-align: center; font-style: italic;">Outputs from Cosmos Transfer varying environments and weather conditions.</p>
73+
74+
| ![Cosmos Transfer output](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/output_1.gif) |
75+
| :--: |
76+
| *Outputs from Cosmos Transfer varying environments and weather conditions.* |
77+
78+
| ![Cosmos Transfer output](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/output_2.gif) |
79+
| :--: |
80+
| *Outputs from Cosmos Transfer varying environments and weather conditions.* |
81+
82+
| ![Cosmos Transfer output](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/physical-ai/output_3.gif) |
83+
| :--: |
84+
| *Outputs from Cosmos Transfer varying environments and weather conditions.* |
85+
86+
87+
Cosmos Transfer coupled with the NVIDIA Omniverse platform is driving controllable synthetic data generation for robotics and autonomous vehicle development at scale. Find more Cosmos Transfer [Examples](https://github.com/nvidia-cosmos/cosmos-transfer1#examples) on GitHub.
88+
89+
Cosmos Transfer [samples](https://huggingface.co/nvidia/Cosmos-Transfer1-7B-Sample-AV) built using post-training base model are also available for autonomous vehicles.
90+
91+
## Open Physical AI Dataset
92+
93+
NVIDIA has also released **Physical AI Dataset**, an open-source dataset on [Hugging Face](https://huggingface.co/collections/nvidia/physical-ai-67c643edbb024053dcbcd6d8) for developing physical AI. This commercial-grade, pre-validated dataset consists of 15 terabytes of data representing more than 320,000 trajectories for robotics training, plus up to 1,000 Universal Scene Description (OpenUSD) assets, including a SimReady collection.
94+
95+
The dataset is designed for post-training foundation models like Cosmos Predict world foundation models, providing developers with high-quality, diverse data to enhance their AI models.
96+
97+
## Purpose Built Model for Humanoids - NVIDIA Isaac GR00T N1
98+
99+
Another exciting announcement is the release of [NVIDIA Isaac GR00T N1](https://developer.nvidia.com/blog/accelerate-generalist-humanoid-robot-development-with-nvidia-isaac-gr00t-n1/), the world's first open foundation model for generalized humanoid robot reasoning and skills. This cross-embodiment model takes multimodal input, including language and images, to perform manipulation tasks in diverse environments. The NVIDIA Isaac GR00T-N1-2B model is available on [Hugging Face](https://huggingface.co/nvidia/GR00T-N1-2B).
100+
101+
Isaac GR00T N1 was trained on an expansive humanoid dataset, consisting of real captured data, synthetic data generated using components of the [NVIDIA Isaac GR00T Blueprint](https://developer.nvidia.com/blog/building-a-synthetic-motion-generation-pipeline-for-humanoid-robot-learning/), and internet-scale video data. It is adaptable through post-training for specific embodiments, tasks and environments.
102+
103+
Isaac GR00T N1 uses a single model and set of weights to enable manipulation behaviors on various humanoid robots, such as the [Fourier GR-1](https://www.fftai.com/products-gr1) and [1X Neo](https://www.1x.tech/neo). It demonstrates robust generalization across a range of tasks, including grasping and manipulating objects with one or both arms, as well as transferring items between arms. It can also execute complex, multi-step tasks that require sustained contextual understanding and the integration of diverse skills. These capabilities make it well-suited for applications in material handling, packaging, and inspection.
104+
105+
Isaac GR00T N1 features a dual-system architecture inspired by human cognition, consisting of the following complementary components:
106+
107+
- **Vision-Language Model (System 2)**: This methodical thinking system is based on [NVIDIA-Eagle](https://huggingface.co/NVEagle) with [SmolLM-1.7B](https://huggingface.co/HuggingFaceTB/SmolLM-1.7B). It interprets the environment through vision and language instructions, enabling robots to reason about their environment and instructions, and plan the right actions.
108+
- **Diffusion Transformer (System 1)**: This action model generates continuous actions to control the robot's movements, translating the action plan made by System 2 into precise, continuous robot movements.
109+
110+
## Path Forward
111+
112+
Post-training is the path forward to advancing autonomous systems, creating specialized models for downstream physical AI tasks.
113+
114+
Check out GitHub for [Cosmos Predict](https://github.com/nvidia-cosmos/cosmos-predict1) and [Cosmos Transfer](https://github.com/nvidia-cosmos/cosmos-transfer1) inference scripts. Explore the Cosmos Transfer [research paper](https://research.nvidia.com/labs/dir/cosmos-transfer1/) for more details.
115+
116+
The NVIDIA Isaac GR00T-N1-2B model is available on [Hugging Face](https://huggingface.co/nvidia/GR00T-N1-2B). Sample datasets and PyTorch scripts for post-training using custom user datasets, which is compatible with the Hugging Face LeRobot format are available on [GitHub](http://github.com/NVIDIA/Isaac-GR00T). For more information about the Isaac GR00T N1 model, see the [research paper](https://research.nvidia.com/publication/2025-03_nvidia-isaac-gr00t-n1-open-foundation-model-humanoid-robots).
117+
118+
Follow [NVIDIA on Hugging Face](https://huggingface.co/nvidia) for more updates.

0 commit comments

Comments
 (0)