LFG: Learning to Drive is a Free Gift

Large-Scale Label-Free Autonomy Pretraining from Unposed In-The-Wild Videos

CVPR 2026

1Applied Intuition 2Stanford University 3UC Berkeley
LFG Teaser Figure

Figure 1. LFG learns a unified pseudo-4D representation of geometry, semantics, motion, and short-term future evolution directly from unposed, unlabeled single-view driving videos. A single feedforward encoder processes observed frames and produces temporally consistent predictions of 3D point maps, camera poses, semantic layouts, confidence, and motion masks for both current and future frames.

Abstract

Ego-centric driving videos available online provide an abundant source of visual data for autonomous driving, yet their lack of annotations makes it difficult to learn representations that capture both semantic structure and 3D geometry. Recent advances in large feedforward spatial models demonstrate that point maps and ego-motion can be inferred in a single forward pass, suggesting a promising direction for scalable driving perception. We therefore propose a label-free, teacher-guided framework for learning autonomous driving representations directly from unposed videos.

Unlike prior self-supervised approaches that focus primarily on frame-to-frame consistency, we posit that safe and reactive driving depends critically on temporal context. To this end, we leverage a feedforward architecture equipped with a lightweight autoregressive module, trained using multi-modal supervisory signals that guide the model to jointly predict current and future point maps, camera poses, semantic segmentation, and motion masks.

Multi-modal teachers provide sequence-level pseudo-supervision, enabling LFG to learn a unified pseudo-4D representation from raw YouTube videos without poses, labels, or LiDAR. The resulting encoder not only transfers effectively to downstream autonomous driving planning on the NAVSIM benchmark, surpassing multi-camera and LiDAR baselines with only a single monocular camera, but also yields strong performance when evaluated on a range of semantic, geometric, and qualitative motion prediction tasks.

We emphasize that the novelty of LFG lies more within the pretraining paradigm than the model architecture itself.

Method at a Glance

LFG Method

LFG Architecture. Starting from unposed single-view driving clips, a pretrained π3 backbone encodes N observed frames into latent scene tokens. A lightweight causal autoregressive transformer rolls out M future tokens, which a shared decoder maps to point maps, camera poses, semantic segmentation, confidence maps, and motion masks for all N+M frames. Multi-modal teachers provide pseudo-supervision, enabling LFG to learn a unified pseudo-4D representation that transfers effectively to downstream planning.

Depth Estimation

Click on the image to cycle through: RGB Input → π3 Depth → LFG Depth

Depth comparison
RGB Input
Current: Future:

LFG predicts depth for both current (observed) and future (predicted) frames, while π3 requires all frames as input.

Semantic Segmentation

Click on the image to cycle through: RGB Input → SegFormer → LFG Semantics

Semantics comparison
RGB Input
Current: Future:

LFG outperforms its SegFormer teacher on both current (observed) and future (predicted) frame predictions.

Motion Prediction

Click on the image to cycle through: RGB Input → LFG Motion → Pseudo GT

Motion comparison
RGB Input

LFG correctly identifies dynamic objects (vehicles, pedestrians) and separates them from static scene elements.

Point Cloud Reconstruction

Click on the image to cycle through: LFG → π3

Point cloud comparison
LFG

LFG preserves geometric structure and camera motion even when predicting future frames (shown in red poses).

Quantitative Results

NAVSIM Planning Benchmark

Single-camera LFG vs BEV-based baselines. Higher is better for all metrics.

Method Input NC DAC TTC C. EP PDMS
UniAD 6Cam 97.8 91.9 92.9 100.0 78.8 83.4
TransFuser 3Cam+L 97.7 92.8 92.0 100.0 79.2 84.0
Hydra-MDP 3Cam+L 96.9 94.0 94.0 100.0 78.7 84.7
DiffusionDrive 3Cam+L 96.8 95.4 94.7 100.0 82.0 88.1
LFG (Ours) 1Cam 98.2 93.7 94.4 100.0 79.1 85.2

L = LiDAR. 1Cam uses only the front-view camera with past temporal frames (3-frame input).

Data Efficiency Comparison

PDMS scores on NAVSIM with varying amounts of labeled training data.

Method Input 1% 10% 100%
DiffusionDrive 3Cam+L 64.9 72.6 88.1
DINOv3 1Cam 60.0 75.8 81.4
PPGeo 1Cam 61.5 65.6 74.6
π3 1Cam 56.2 77.5 82.8
LFG (Ours) 1Cam 66.3 81.4 85.2

LFG demonstrates strong data efficiency, matching full-data DINOv3 performance with only 10% labeled data.

Ablation Studies

Component and scaling ablations on NAVSIM (PDMS). Higher is better.

Setting 1% 10% 100%
Original setting 66.3 81.4 85.2
+ 2× pretraining data 76.6 82.3 84.8
+ Longer prediction horizon 80.5 84.4 84.8
- Seg, Motion 64.8 77.1 84.6
- Autoregressive head 66.3 77.7 84.2

Scaling pretraining data and extending prediction horizon improve performance, while removing segmentation/motion supervision or the autoregressive head degrades results.

BibTeX

@inproceedings{strong2026lfg,
  title={Learning to Drive is a Free Gift: Large-Scale Label-Free Autonomy Pretraining from Unposed In-The-Wild Videos},
  author={Strong, Matthew and Chang, Wei-Jer and Herau, Quentin and Yang, Jiezhi and Hu, Yihan and Peng, Chensheng and Zhan, Wei},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2026}
}