Cosmos-Transfer1: Conditional World Generation with Adaptive Multimodal Control
Abstract
We introduce Cosmos-Transfer, a conditional world generation model that can generate world simulations based on multiple spatial control inputs of various modalities such as segmentation, depth, and edge. In the design, the spatial conditional scheme is adaptive and customizable. It allows weighting different conditional inputs differently at different spatial locations. This enables highly controllable world generation and finds use in various world-to-world transfer use cases, including Sim2Real. We conduct extensive evaluations to analyze the proposed model and demonstrate its applications for Physical AI, including robotics Sim2Real and autonomous vehicle data enrichment. We further demonstrate an inference scaling strategy to achieve real-time world generation with an NVIDIA GB200 NVL72 rack. To help accelerate research development in the field, we open-source our models and code at https://github.com/nvidia-cosmos/cosmos-transfer1.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Learning Real-World Action-Video Dynamics with Heterogeneous Masked Autoregression (2025)
- Proc4Gem: Foundation models for physical agency through procedural generation (2025)
- Hybrid Rendering for Multimodal Autonomous Driving: Merging Neural and Physics-Based Simulation (2025)
- CAML: Collaborative Auxiliary Modality Learning for Multi-Agent Systems (2025)
- The Role of World Models in Shaping Autonomous Driving: A Comprehensive Survey (2025)
- SimWorld: A Unified Benchmark for Simulator-Conditioned Scene Generation via World Model (2025)
- From Gaming to Research: GTA V for Synthetic Data Generation for Robotics and Navigations (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on HF中国镜像站 checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 2
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper