You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: nvidia-physical-ai.md
+1-31Lines changed: 1 addition & 31 deletions
Original file line number
Diff line number
Diff line change
@@ -43,23 +43,6 @@ The model is built by training individual ControlNets separately for each sensor
43
43
</div>
44
44
<pstyle="text-align: center; font-style: italic;">Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.</p>
45
45
46
-
||
47
-
| :--: |
48
-
|*Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.*|
49
-
50
-
||
51
-
| :--: |
52
-
|*Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.*|
53
-
54
-
||
55
-
| :--: |
56
-
|*Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.*|
57
-
58
-
||
59
-
| :--: |
60
-
|*Input types include 3D bounding box map, Trajectory map, Depth map, Segmentation map.*|
61
-
62
-
63
46
- At inference time, developers can use various input types, including structured visual or geometric data such as segmentation maps, depth maps, edge maps, human motion keypoints, LiDAR scans, trajectories, HD maps, and 3D bounding boxes to guide the output.
64
47
- The control signals from each control branch are multiplied by their corresponding adaptive spatiotemporal control maps and then summed before being added to the transformer blocks of the base model.
65
48
- The generated output is photorealistic video sequences with controlled layout, object placement, and motion. Developers can control the output in multiple ways, such as preserving structure and appearance or allowing appearance variations while maintaining structure.
@@ -71,19 +54,6 @@ The model is built by training individual ControlNets separately for each sensor
71
54
</div>
72
55
<pstyle="text-align: center; font-style: italic;">Outputs from Cosmos Transfer varying environments and weather conditions.</p>
73
56
74
-
||
75
-
| :--: |
76
-
|*Outputs from Cosmos Transfer varying environments and weather conditions.*|
77
-
78
-
||
79
-
| :--: |
80
-
|*Outputs from Cosmos Transfer varying environments and weather conditions.*|
81
-
82
-
||
83
-
| :--: |
84
-
|*Outputs from Cosmos Transfer varying environments and weather conditions.*|
85
-
86
-
87
57
Cosmos Transfer coupled with the NVIDIA Omniverse platform is driving controllable synthetic data generation for robotics and autonomous vehicle development at scale. Find more Cosmos Transfer [Examples](https://github.com/nvidia-cosmos/cosmos-transfer1#examples) on GitHub.
88
58
89
59
Cosmos Transfer [samples](https://huggingface.co/nvidia/Cosmos-Transfer1-7B-Sample-AV) built using post-training base model are also available for autonomous vehicles.
@@ -115,4 +85,4 @@ Check out GitHub for [Cosmos Predict](https://github.com/nvidia-cosmos/cosmos-pr
115
85
116
86
The NVIDIA Isaac GR00T-N1-2B model is available on [Hugging Face](https://huggingface.co/nvidia/GR00T-N1-2B). Sample datasets and PyTorch scripts for post-training using custom user datasets, which is compatible with the Hugging Face LeRobot format are available on [GitHub](http://github.com/NVIDIA/Isaac-GR00T). For more information about the Isaac GR00T N1 model, see the [research paper](https://research.nvidia.com/publication/2025-03_nvidia-isaac-gr00t-n1-open-foundation-model-humanoid-robots).
117
87
118
-
Follow [NVIDIA on Hugging Face](https://huggingface.co/nvidia) for more updates.
88
+
Follow [NVIDIA on Hugging Face](https://huggingface.co/nvidia) for more updates.
0 commit comments