D2E: Scaling Vision-Action Pretraining on Desktop Data for Transfer to Embodied AI

Suhwan Choi†1, Jaeyoon Jung†1, Haebin Seong†1, Minchan Kim1, Minyeong Kim2,
Yongjun Cho1, Yoonshik Kim1, Yubeen Park1, Youngjae Yu‡3, Yunsung Lee‡1,
Equal contribution, Co-corresponding author, 1 MAUM.AI, 2 Stanford University, 3 Seoul National University

Under Review

d2e teaser

Overview of D2E (Desktop to Embodied AI) framework. (1) The OWA Toolkit captures 335.6 hours of rich desktop demonstrations across 31 games with 152× compression. (2) The Generalist-IDM uses nextevent prediction with temporal offset (NEP-τ) to achieve OOD generalization, enabling pseudolabeling of 1K+ hours of YouTube gameplay. (3) Vision-Action Pretraining transfers desktoppretrained representations to embodied AI, achieving 96.6% success on LIBERO manipulation and 83.3% on CANVAS navigation benchmarks which demonstrates desktop-to-robotics transfer.

Pseudo-Label result on YouTube dataset

Brotato

CSGO2

Stardew Valley

Minecraft

Slime Rancher

Raft

Barony

Dinkum

Generalist-IDM uses a single model to label actions on video-only data, across 2D/3D games and visual navigation/UI interactions without separate processing.
Remarkably, in Counter-Strike videos where spectator mode begins around 10 seconds, it can distinguish between active gameplay and spectator mode by recognizing subtle UI elements, avoiding action predictions during spectator phases.

Abstract

Large language models leverage internet-scale text data, yet embodied AI remains constrained by the prohibitive costs of physical trajectory collection. Desktop environments---particularly gaming---offer a compelling alternative: they provide rich sensorimotor interactions at scale while maintaining the structured observation-action coupling essential for embodied learning. We present D2E (Desktop to Embodied AI), a framework that demonstrates desktop interactions can serve as an effective pretraining substrate for robotics embodied AI tasks. Unlike prior work that remained domain-specific (e.g., VPT for Minecraft) or kept data proprietary (e.g., SIMA), D2E establishes a complete pipeline from scalable desktop data collection to verified transfer in embodied domains. Our framework comprises three components: (1) the OWA Toolkit that unifies diverse desktop interactions into a standardized format with 152× compression, (2) the Generalist-IDM that achieves strong zero-shot generalization across unseen games through timestamp-based event prediction, enabling internet-scale pseudo-labeling, and (3) VAPT that transfers desktop-pretrained representations to physical manipulation and navigation. Using 1.3K+ hours of data (259 hours of human demonstrations, and 1K+ hours of pseudo-labeled gameplay), we achieve a total of 96.6% success rate on LIBERO manipulation and 83.3% on CANVAS navigation benchmarks. This validates that sensorimotor primitives in digital interactions exhibit sufficient invariance to transfer meaningfully to physical embodied tasks, establishing desktop pretraining as a practical paradigm for robotics. We will make all our work public, including the OWA toolkit, datasets of human-collected and pseudo-labeled, and VAPT-trained models.

Generalist Inverse Dynamics Model (G-IDM)

gidm_id

The Generalist Inverse Dynamics Model (G-IDM) learns to predict actions from observation transitions across diverse desktop environments. This enables zero-shot generalization to unseen games and provides a foundation for pseudo-labeling large-scale gameplay data.

Ground Truth

IDM

G-IDM

Out-of-Distribution Generalization

gidm_ood

G-IDM demonstrates strong out-of-distribution generalization capabilities, successfully transferring learned sensorimotor patterns from desktop gaming environments to completely unseen game domains and interaction modalities.

Ground Truth

IDM (Fine Tune)

G-IDM (Zero Shot)

G-IDM (Few Shot)

G-IDM (Fine Tune)

In Battlefield 6, for G-IDM (Zero Shot) we can observe that scale of mouse movement is different with GT, but we can observe that scale of movement remain nearly same for G-IDM (Few Shot). This improvement occurs because providing context examples helps the model calibrate the appropriate movement scale.

Desktop-to-Embodied Transfer

To validate the effectiveness of desktop pretraining for embodied AI, we evaluate our approach on two challenging downstream tasks: LIBERO manipulation and CANVAS navigation. These benchmarks represent diverse embodied scenarios requiring different sensorimotor skills - precise object manipulation and spatial navigation respectively. Our Vision-Action Pretraining (VAPT) framework transfers desktop-pretrained representations to these physical domains, demonstrating that sensorimotor patterns learned from gaming environments can generalize to real-world robotic tasks.

LIBERO Manipulation Results

libero_results

D2E achieves impressive results on LIBERO manipulation benchmarks, demonstrating that desktop-pretrained models can effectively transfer to physical robotic manipulation tasks with a 96.6% success rate.

Baseline (42%)

+ VAPT w/o pseudo (98%)

+ VAPT w/ pseudo (100%)

CANVAS Navigation Results

canvas_results

The framework also excels in navigation tasks, achieving 83.3% success rate on CANVAS navigation benchmarks, validating the transferability of desktop sensorimotor patterns to embodied navigation scenarios.

Baseline (fail)

+ VAPT w/ pseudo (success)

BibTeX

@article{choi2025d2e,
    title={D2E: Scaling Vision-Action Pretraining on Desktop Data for Transfer to Embodied AI},
    author={Choi, Suwhan and Jung, Jaeyoon and Seong, Haebin and Kim, Minchan and Kim, Minyeong and Cho, Yongjun and Kim, Yoonshik and Park, Yubeen and Yu, Youngjae and Lee, Yunsung},
    journal={arXiv preprint arXiv:2510.05684},
    year={2025}
}

The website design was based on general-navigation-models adapted from Nerfies.