EgoForge: Goal-Directed Egocentric World Simulator

Yifan Shen1, Jiateng Liu1, Xinzhuo Li1, Yuanzhe Liu1, Bingxuan Li1, Houze Yang1, Wenqi Jia1, Yijiang Li2, Tianjiao Yu1, James Matthew Rehg1, Xu Cao1,†, Ismini Lourentzou1,†
1 University of Illinois Urbana-Champaign    2University of California San Diego
Figure 1: Egocentric video rollouts produced by EgoForge in real-world smart-glasses experiments.

Figure 1: Egocentric video rollouts produced by EgoForge in real-world smart-glasses experiments. Given a single smart-glasses egocentric image, a high-level goal instruction, and an auxiliary exocentric view, EgoForge generates egocentric rollouts that follow user intent and preserve scene structure, without requiring dense supervision such as camera trajectories, pose, video, or synchronized multi-view capture streams.

Abstract

Generative world models have shown promise for simulating dynamic environments, yet egocentric video remains challenging due to rapid viewpoint changes, frequent hand–object interactions, and goal-directed procedures whose evolution depends on latent human intent. Existing approaches either focus on hand-centric instructional synthesis with limited scene evolution, perform static view translation without modeling action dynamics, or rely on dense supervision such as camera trajectories, long video prefixes, or synchronized multi-camera capture.

In this work, we introduce EgoForge, an egocentric goal-directed world simulator that generates coherent, first-person video rollouts from minimal static inputs: a single egocentric image, a high-level instruction, and an optional auxiliary exocentric view. To improve intent alignment and temporal consistency, we propose VideoDiffusionNFT, a trajectory-level reward-guided refinement that optimizes goal completion, temporal causality, scene consistency, and perceptual fidelity during diffusion sampling. Extensive experiments show EgoForge achieves consistent gains in semantic alignment, geometric stability, and motion fidelity over strong baselines, and robust performance in real-world smart-glasses experiments.

Method

Figure 2: EgoForge Overview.

Figure 2: EgoForge Overview. Given a single egocentric observation, a high-level (instruction) intent, and an auxiliary exo-view reference, EgoForge fuses encoded visual features with noisy video latents at each DiT block to guide generation. Geometry alignment weakly supervises intermediate features using angular and scale consistency to encourage spatially stable rollouts. The resulting rollout videos are further refined via a novel VideoDiffusionNFT alignment policy that optimizes goal completion, scene consistency, temporal causality, and perceptual fidelity.

Diffusion-Based Egocentric Generator

EgoForge models egocentric video generation in the latent space of a pretrained video autoencoder, conditioned on context 𝒞 = {ego image, instruction y, exo reference} through adaptive normalization and cross-attention. Unlike prior work, EgoForge does not require camera trajectories, pose signals, or synchronized multi-view streams at inference time.

Geometry Weak Supervision

To inject 3D reasoning into the diffusion backbone, EgoForge aligns intermediate DiT representations with geometry features from a pretrained VGGT encoder via two losses:

  • Angular Alignment (ℒang): Cosine similarity loss that encourages diffusion features to match the direction of VGGT geometry features at each spatial–temporal token.
  • Scale Alignment (ℒsca): MSE loss between projected diffusion features and VGGT features to prevent scale collapse.

VideoDiffusionNFT Alignment

VideoDiffusionNFT extends DiffusionNFT to the video domain and performs negative-aware finetuning with four fine-grained reward functions evaluated on entire video trajectories:

  • Goal Completion (ℛgoal): Evaluates whether the trajectory successfully achieves the task outcome, measured by similarity of the final state to the target reference.
  • Scene Consistency (ℛenv): Measures consistency with the initial scene, penalizing drift, misplaced objects, or transitions into unrelated environments.
  • Temporal Causality (ℛtemp): Assesses whether motion evolves in a physically plausible, coherent, and causal manner without temporal artifacts.
  • Perceptual Fidelity (ℛper): Captures overall visual clarity, stability, and absence of distortions or artifacts (PSNR + FVD + LPIPS).

Quantitative Results

Table 3: Quantitative comparisons on the X-Ego benchmark. EgoForge outperforms all baselines across semantic, perceptual, and temporal metrics.

Model DINO-Score↑ CLIP-Score↑ SSIM↑ LPIPS↓ FVD↓ Flow MSE↓ PSNR↑
EgoDreamer42.3525.400.580.35580.458.1515.20
Handi31.1218.250.420.52912.3014.5012.85
Cosmos49.4229.770.700.26448.126.4018.73
HunyuanVideo53.5429.430.710.26384.316.1018.88
WAN2.253.9935.690.720.23322.175.7820.44
EgoForge (Ours) 61.2539.30 0.790.15 182.252.83 24.08

Table 2: Quantitative comparisons with enhanced baseline variants (+EV: exo-view image; +TT: text-only domain adaptation; +CI: our conditioning inputs with Geometry Weak Supervision). EgoForge still achieves best results across all seven metrics.

Model DINO-Score↑ CLIP-Score↑ SSIM↑ LPIPS↓ FVD↓ Flow MSE↓ PSNR↑
Cosmos+EV48.6029.600.670.28485.756.8218.30
Cosmos+TT50.8030.400.710.25433.906.3118.88
HunyuanVideo+EV52.8029.200.700.27405.876.3018.61
HunyuanVideo+TT54.1029.860.720.24365.805.9519.10
WAN2.2+EV52.9135.110.710.27352.416.2520.05
WAN2.2+TT54.8036.200.730.25310.575.6020.64
WAN2.2+CI58.9238.050.760.18218.723.9222.87
EgoForge (Ours) 61.2539.30 0.790.15 182.252.83 24.08

Table 4: User study (1–5 scale, 20 annotators, 25 video groups). EgoForge achieves substantial gains in Alignment (4.75) and Fidelity (4.71). * best model variants from Table 2.

Model Quality↑ Fidelity↑ Smooth Motion↑ Smooth Env.↑ Alignment↑
Cosmos*3.292.543.072.472.19
Hunyuan*3.462.863.723.163.08
WAN2.2*3.223.483.824.073.15
EgoForge (Ours) 4.584.71 4.254.48 4.75

Table 5: Ablation on EgoForge modules (FT = Denoising Fine-Tuning, GWS = Geometry Weak Supervision). Each component consistently improves performance, with the full model achieving best results.

FTGWSVideoDiffusionNFT DINO↑CLIP↑SSIM↑ LPIPS↓FVD↓Flow MSE↓PSNR↑
56.8137.100.740.21260.894.8221.92
58.9238.050.760.18218.723.9222.87
61.2539.300.79 0.15182.252.8324.08

Qualitative Results: Comparison with Baselines

Figure 3: Qualitative Comparison between EgoForge and baselines.

Figure 3: Qualitative comparison between EgoForge and baselines. EgoForge accurately reconstructs multi-step, causally ordered actions, preserving hand–object geometry, temporal consistency, and goal alignment. In the first example (coffee pouring), Cosmos erroneously generates a third hand, Hunyuan depicts a disconnected arm, and Wan2.2 fails to complete the task. In the second example (basketball), Cosmos generates multiple balls, while Hunyuan and Wan2.2 generate an incorrect person. EgoForge accurately completes both tasks.

Figure 4: Qualitative Comparison between EgoForge and baselines (hand-washing and soccer).

Figure 4: Qualitative comparison: hand-washing task (top) and soccer task (bottom). In the hand-washing task, baselines struggle with object consistency (Cosmos hallucinates the soap source; Wan2.2 bypasses the on-table soap), while EgoForge successfully executes the action using existing objects. In the soccer task, baselines exhibit ghosting (Cosmos) or fail to follow precise instructions (Hunyuan, Wan2.2). EgoForge accurately traps with the left leg and shoots with the right.

Qualitative Results: With vs. Without Exocentric Input

Figure 6: EgoForge with vs. without exocentric input.

Figure 5: Qualitative comparison of EgoForge with vs. without exocentric input. Rows 1 & 3 ("Without Exo-View") use text prompt alone; Rows 2 & 4 ("With Exo-View") additionally provide an auxiliary exocentric image. The kitchen scene (Row 2) correctly incorporates the potted plants on the windowsill from the reference image; the basketball court scene (Row 4) adopts the red and green rubberized surface. EgoForge can be successfully steered toward simulations that inherit key semantic and stylistic properties from the reference exo-view.

Qualitative Results: Long-Duration Sequences

Figure 5: Qualitative egocentric video rollouts.

Figure 7: Dense visualization of long-duration sequences.

Figure 6: Dense visualization of long-duration sequences. We present 26 frames from each of 8 videos generated by EgoForge — cracking an egg, folding a blanket, tearing adhesive tape, rock climbing, placing a package, shooting a basketball, chopping onions, and installing a drawer handle — to highlight seamless transitions and stable dynamics in complex egocentric tasks.

BibTeX

@article{shen2025egoforge,
  title={EgoForge: Goal-Directed Egocentric World Simulator},
  author={Shen, Yifan and Liu, Jiateng and Li, Xinzhuo and Liu, Yuanzhe and Li, Bingxuan and Yang, Houze and Jia, Wenqi and Li, Yijiang and Yu, Tianjiao and Rehg, James Matthew and Cao, Xu and Lourentzou, Ismini},
  journal={arXiv preprint},
  year={2025}
}