- Install PyTorch (pytorch.org).
pip install -r requirements.txt
- Download the MEAD dataset from (here).
- Download the pre-trained weights (here) (" backbone.pth ") and place it under "./pretrain/backbone.pth"
The obtained MEAD dataset is first preprocessed with 'dataloader/align_face.py':
python align_face.py
Get audio feature corresponding to image:
python w2f.py
cd CCRL
python train.py
cd CDRL
python train.py
First, follow the data preparation and processing workflow of NED.
bash run_manipulator.sh
bash run_render.sh
cd metrics
bash eval_crossID_driven.sh