This repository includes the code to reconstruct facial sketches from voices.
mlsp_project_code
contains the old code from the course projectvoice2face
contains the refactored codespeaker_id_weights.pth
is the speaker ID weights file
Use the requirements.txt
file to install the dependencies. Preferably, use conda since PyTorch likes conda.
The dataloader expects a file containing the filepath of each sample along with labels. For example, take a look at vgg_voxceleb_edge_preserving.txt
. Then train the model as follows:
$ python train.py vgg_voxceleb_edge_preserving.txt
- Make sure you can understand and can run the code inside
voice2face
- Create a new directory in this repo, call it
face_id
orspeaker_id
and use the code insidevoice2face
as a template (makes it easier for everyone to read your code) - I used Weights&Biases to visualize the training metrics, feel free to use it, take it out or use anything else you like
Please reach out to me (Mahmoud) on Slack