Skip to content

showlab/Exo2Ego-V

Repository files navigation

[NeurIPS 2024] Exocentric-to-Egocentric Video Generation

This is the official repository of Exo2Ego-V Paper

Jia-Wei Liu*, Weijia Mao*, Zhongcong Xu, Jussi Keppo, Mike Zheng Shou

TL;DR: A novel exocentric-to-egocentric video generation method for challenging daily-life skilled human activities.

Exo2Ego-V

📝 Preparation

Installation

git clone https://github.com/showlab/Exo2Ego-V.git
cd Exo2Ego-V
pip install -r requirements.txt

Download Pre-Trained Weights

python tools/download_weights.py

Model weights should be placed under ./pretrained_weights.

🚀 Ego-Exo4D Dataset

Please refer to https://ego-exo4d-data.org/ for downloading the Ego-Exo4D dataset. Our experiments utilized the downscaled takes at 448px on the shortest side.

Data Processing -- Frames Extraction

Please modify the data and output directory in each script.

python scripts_preprocess/extract_frames_from_videos.py

Data Processing -- Camera Poses

Please modify the data, input, and output directory in each script. The "train_dict" and "test_dict" are the train and test split dictionary: link.

python scripts_preprocess/get_ego_pose.py
python scripts_preprocess/get_exo_pose.py
python scripts_preprocess/get_ego_intrinsics.py

🏋️‍️ Experiment

Training

Stage 1: Train Exo2Ego Spatial Appearance Generation. Please modify the data and pretrained model weights directory in configs/train/stage1.yaml

bash train_stage1.sh

Stage 2: Train Exo2Ego Temporal Motion Video Generation. Please modify the data, pretrained model weights, and stage 1 model weights directory in configs/train/stage2.yaml

bash train_stage2.sh

Checkpoints

We release the 5 Pretrained Exo2Ego View Translation Prior checkpoints on link.

🎓 Citation

If you find our work helps, please cite our paper.

@article{liu2024exocentric,
  title={Exocentric-to-egocentric video generation},
  author={Liu, Jia-Wei and Mao, Weijia and Xu, Zhongcong and Keppo, Jussi and Shou, Mike Zheng},
  journal={Advances in Neural Information Processing Systems},
  volume={37},
  pages={136149--136172},
  year={2024}
}

✉️ Contact

This repo is maintained by Jiawei Liu. Questions and discussions are welcome via [email protected].

🙏 Acknowledgements

This codebase is based on MagicAnimate, Moore-AnimateAnyone, and PixelNeRF. Thanks for open-sourcing!

LICENSE

Copyright (c) 2025 Show Lab, National University of Singapore. All Rights Reserved. Licensed under the Apache License, Version 2.0 (see LICENSE for details)

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages