Fast-WAM: Do World Action Models Need Test-time Future Imagination?

1IIIS, Tsinghua University 2Galaxea AI

TL;DR

Q

Do WAMs benefit mainly from explicit future imagination at inference, or from video modeling during training?

A

The main benefit comes from video co-training during training. Fast-WAM removes explicit future generation at test time, stays competitive, and runs much faster.

Three representative WAM paradigms and the Fast-WAM design.

Fast-WAM keeps video co-training during training but removes explicit future generation at inference time, enabling direct action generation from latent world representations in a single forward pass.

Abstract

World Action Models (WAMs) have emerged as a promising alternative to Vision-Language-Action (VLA) models for embodied control because they explicitly model how visual observations may evolve under action. Most existing WAMs follow an imagine-then-execute paradigm, incurring substantial test-time latency from iterative video denoising, yet it remains unclear whether explicit future imagination is actually necessary for strong action performance.

We ask whether WAMs need explicit future imagination at test time, or whether their benefit comes primarily from video modeling during training. To answer this, we propose Fast-WAM, a WAM architecture that retains video co-training during training but skips future prediction at test time. Across controlled variants, Fast-WAM remains competitive with imagine-then-execute variants, while removing video co-training causes a much larger performance drop.

Empirically, Fast-WAM achieves competitive results on LIBERO, RoboTwin, and real-world towel folding without embodied pretraining. It runs in real time with 190 ms latency, more than 4x faster than existing imagine-then-execute WAM designs.

Core Question

Do WAMs benefit mainly from explicit future imagination at inference, or from video modeling during training?

Design

Fast-WAM preserves world-model supervision during training while using a direct-policy interface at test time.

Finding

Video co-training matters much more than explicit test-time future generation for final control performance.

A WAM That Drops Future Imagination at Test Time

Fast-WAM model architecture.

Fast-WAM architecture with a shared-attention Mixture-of-Transformer design.

Training and inference attention masks.

Structured attention masks disentangle video co-training from action generation.

Fast-WAM is built on a pretrained video Diffusion Transformer backbone and an action expert DiT. During training, it jointly learns action prediction and video modeling so the shared visual backbone acquires stronger world-grounded representations.

At inference time, Fast-WAM keeps only the clean latent tokens of the current observation, processes them with the video backbone once, and directly generates actions without explicit future video denoising. This removes the main runtime bottleneck of imagine-then-execute WAMs.

Backbone: Wan2.2-5B video DiT with a 1B action expert
Runtime: 190 ms on a single RTX 5090D V2 32GB GPU

Results on Simulation and Real-World Tasks

We evaluate Fast-WAM on LIBERO, RoboTwin 2.0, and a real-world towel-folding task. Across all settings, the central pattern is consistent: Fast-WAM remains close to imagine-then-execute variants, while removing video co-training causes a much larger performance drop.

Real-world towel-folding benchmark on Galaxea R1 Lite.

Real-world towel-folding benchmark. Folding a deformable object requires long-horizon planning and precise closed-loop manipulation, making it a useful testbed for both success and execution efficiency.

RoboTwin 2.0

Method Embodied PT. Clean Rand. Average
π065.9258.4062.2
π0.582.7476.7679.8
Motus88.6687.0287.8
Motus from WAN2.277.5677.0077.3
LingBot-VA92.9091.5092.2
LingBot-VA from WAN2.280.60--80.6
Fast-WAM (Ours)91.8891.7891.8
Fast-WAM Variants
Fast-WAM91.8891.7891.8
Fast-WAM-Joint90.8490.3290.6
Fast-WAM-IDM91.1691.3491.3
Fast-WAM w.o. video co-train82.7684.8083.8

LIBERO

Method Embodied PT. Spatial Object Goal Long Average
OpenVLA84.788.479.253.776.5
π096.898.895.885.294.1
π0.598.898.298.092.496.9
LingBot-VA98.599.697.298.598.5
Motus96.899.896.697.697.7
Fast-WAM (Ours)98.2100.097.095.297.6
Fast-WAM Variants
Fast-WAM98.2100.097.095.297.6
Fast-WAM-Joint99.699.498.296.898.5
Fast-WAM-IDM98.897.897.897.698.0
Fast-WAM w.o. video co-train89.299.295.490.093.5
Fast-WAM real-world performance and latency.

Fast-WAM achieves strong real-world performance with substantially lower latency than imagine-then-execute baselines. Fast-WAM-IDM is slower at 810 ms, while Fast-WAM keeps latency at 190 ms.

Removing video co-training causes a dramatic degradation in both success and completion time, again reinforcing that co-training matters more than explicit future imagination at inference.

Video Co-Training Matters More Than Test-Time Imagination in WAMs

Fast-WAM revisits a basic question in World Action Models: do their gains come from explicit future imagination at test time, or from video modeling during training? Across simulation and real-world results, the answer appears consistent: Fast-WAM remains close to imagine-then-execute variants, while removing video co-training causes a much larger degradation.

This suggests the main value of video prediction in WAMs may lie in learning better world representations during training rather than generating future observations at test time.

BibTeX

@misc{yuan2026fastwam,
  title={Fast-WAM: Do World Action Models Need Test-time Future Imagination?},
  author={Tianyuan Yuan and Zibin Dong and Yicheng Liu and Hang Zhao},
  year={2026},
  note={arXiv preprint arXiv:2603.16666}
}