The aim of this research is to estimate 3D human pose and shape in videos, which is a challenging task due to the complex nature of the human body and the wide range of possible pose and shape variations. This problem also poses difficulty in finding a satisfactory solution due to the trade-off between the accuracy and temporal consistency of the estimated 3D pose and shape. Thus previous researches have prioritized one objective over the other. In contrast, we propose a novel approach called the action-conditioned mesh recovery (ACMR) model, which improves accuracy without compromising temporal consistency by leveraging human action information. Our ACMR model outperforms existing methods that prioritize temporal consistency in terms of accuracy, while also achieving comparable temporal consistency with other state-of-the-art methods. Significantly, the action-conditioned learning process occurs only during training, requiring no additional resources at inference time, thereby enhancing performance without increasing computational demands.
This repository provides the PyTorch implementation of our paper Action-conditioned contrastive learning for 3D human pose and shape estimation in videos. The implementation builds upon the excellent work of TCMR.
Please follow the dataset preparation steps from TCMR. For action pseudo-labels, we used SlowFast through MMAction2. You can either:
- Use our pre-processed pt files that include action labels, or
- Set up MMAction2 and utilize the demo code in trainer to generate pseudo action labels
- Download the ACMR weights from here
- Run demo.py with your input frame folder
If you find this work useful, please consider citing:
@article{song2024action,
title={Action-conditioned contrastive learning for 3D human pose and shape estimation in videos},
author={Song, Inpyo and Ryu, Moonwook and Lee, Jangwon},
journal={Computer Vision and Image Understanding},
volume={249},
pages={104149},
year={2024},
publisher={Elsevier}
}
This work builds upon several excellent previous works. We sincerely thank the authors of: