NIL Overview: from single frame + prompt → generated video → policy learning.
Acquiring physically plausible motor skills across diverse and unconventional morphologies—from humanoids to ants—is crucial for robotics and simulation.
We introduce No-data Imitation Learning (NIL), which:
NIL matches or outperforms baselines trained on real motion‐capture data, effectively replacing expensive data collection with generative video priors.
We validate NIL on locomotion tasks for multiple morphologies (humanoids, quadrupeds, animals).
@article{albaba2025nil,
title = {NIL: No-data Imitation Learning by Leveraging Pre-trained Video Diffusion Models},
author = {Albaba, Mert and Li, Chenhao and Diomataris, Markos and Taheri, Omid and Krause, Andreas and Black, Michael},
journal = {arXiv},
year = {2025},
}