NIL Overview: from single frame + prompt → generated video → policy learning.Acquiring physically plausible motor skills across diverse and unconventional morphologies—from humanoids to ants—is crucial for robotics and simulation. We introduce No-data Imitation Learning (NIL), which:
NIL matches or outperforms baselines trained on real motion-capture data, effectively replacing expensive data collection with generative video priors.

We validate NIL on locomotion tasks for multiple morphologies (humanoids, quadrupeds, animals).
@inproceedings{albaba2026nil,
title = {{NIL}: No-data Imitation Learning by Leveraging Pre-trained Video Diffusion Models},
author = {Albaba, Mert and Li, Chenhao and Diomataris, Markos and Taheri, Omid and Krause, Andreas and Black, Michael},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2026},
}