Until now, Amphion Evaluation has supported the following objective metrics:
- F0 Modeling:
- F0 Pearson Coefficients (FPC)
- F0 Periodicity Root Mean Square Error (PeriodicityRMSE)
- F0 Root Mean Square Error (F0RMSE)
- Voiced/Unvoiced F1 Score (V/UV F1)
- Energy Modeling:
- Energy Root Mean Square Error (EnergyRMSE)
- Energy Pearson Coefficients (EnergyPC)
- Intelligibility:
- Spectrogram Distortion:
- Frechet Audio Distance (FAD)
- Mel Cepstral Distortion (MCD)
- Multi-Resolution STFT Distance (MSTFT)
- Perceptual Evaluation of Speech Quality (PESQ)
- Short Time Objective Intelligibility (STOI)
- Scale Invariant Signal to Distortion Ratio (SISDR)
- Scale Invariant Signal to Noise Ratio (SISNR)
- Speaker Similarity:
- Cosine similarity based on:
We provide a recipe to demonstrate how to objectively evaluate your generated audios. There are three steps in total:
- Pretrained Models Preparation
- Audio Data Preparation
- Evaluation
If you want to calculate RawNet3
based speaker similarity, you need to download the pretrained model first, as illustrated here.
Prepare reference audios and generated audios in two folders, the ref_dir
contains the reference audio and the gen_dir
contains the generated audio. Here is an example.
┣ {ref_dir}
┃ ┣ sample1.wav
┃ ┣ sample2.wav
┣ {gen_dir}
┃ ┣ sample1.wav
┃ ┣ sample2.wav
You have to make sure that the pairwise reference audio and generated audio are named the same, as illustrated above (sample1 to sample1, sample2 to sample2).
Run the run.sh
with specified refenrece folder, generated folder, dump folder and metrics.
cd Amphion
sh egs/metrics/run.sh \
--reference_folder [Your path to the reference audios] \
--generated_folder [Your path to the generated audios] \
--dump_folder [Your path to dump the objective results] \
--metrics [The metrics you need] \
--fs [Optional. To calculate all metrics in the specified sampling rate] \
--similarity_model [Optional. To choose the model for calculating the speaker similarity. Currently "rawnet", "wavlm" and "resemblyzer" are available. Default to "wavlm"] \
--similarity_mode [Optional. To choose the mode for calculating the speaker similarity. "pairwith" for calculating a series of ground truth / prediction audio pairs to obtain the speaker similarity, and "overall" for computing the average score with all possible pairs between the refernece folder and generated folder. Default to "pairwith"] \
--intelligibility_mode [Optionoal. To choose the mode for computing CER and WER. "gt_audio" means selecting the recognition content of the reference audio as the target, "gt_content" means using transcription as the target. Default to "gt_audio"] \
--ltr_path [Optional. Path to the transcription file] \
--language [Optional. Language for computing CER and WER. Default to "english"]
As for the metrics, an example is provided below:
--metrics "mcd pesq fad"
All currently available metrics keywords are listed below:
Keys | Description |
---|---|
fpc |
F0 Pearson Coefficients |
f0_periodicity_rmse |
F0 Periodicity Root Mean Square Error |
f0rmse |
F0 Root Mean Square Error |
v_uv_f1 |
Voiced/Unvoiced F1 Score |
energy_rmse |
Energy Root Mean Square Error |
energy_pc |
Energy Pearson Coefficients |
cer |
Character Error Rate |
wer |
Word Error Rate |
similarity |
Speaker Similarity |
fad |
Frechet Audio Distance |
mcd |
Mel Cepstral Distortion |
mstft |
Multi-Resolution STFT Distance |
pesq |
Perceptual Evaluation of Speech Quality |
si_sdr |
Scale Invariant Signal to Distortion Ratio |
si_snr |
Scale Invariant Signal to Noise Ratio |
stoi |
Short Time Objective Intelligibility |
For example, if want to calculate the speaker similarity between the synthesized audio and the reference audio with the same content, run:
sh egs/metrics/run.sh \
--reference_folder [Your path to the reference audios] \
--generated_folder [Your path to the generated audios] \
--dump_folder [Your path to dump the objective results] \
--metrics "similarity" \
--similarity_model [Optional. To choose the model for calculating the speaker similarity. Currently "rawnet", "wavlm" and "resemblyzer" are available. Default to "wavlm"] \
--similarity_mode "pairwith" \
If you don't have the reference audio with the same content, run the following to get the conteng-free similarity score:
sh egs/metrics/run.sh \
--reference_folder [Your path to the reference audios] \
--generated_folder [Your path to the generated audios] \
--dump_folder [Your path to dump the objective results] \
--metrics "similarity" \
--similarity_model [Optional. To choose the model for calculating the speaker similarity. Currently "rawnet", "wavlm" and "resemblyzer" are available. Default to "wavlm"] \
--similarity_mode "overall" \
If your system is unable to access huggingface.co from the terminal, you might run into an error like "OSError: Can't load tokenizer for ...". To work around this, follow these steps to use local models:
- Download the bert-base-uncased, roberta-base, and facebook/bart-base models from
huggingface.co
. Ensure that the models are complete and uncorrupted. Place these directories withinAmphion/pretrained
. For a detailed file structure reference, see This README underAmphion/pretrained
. - Inside the
Amphion/pretrained
directory, create a bash script with the content outlined below. This script will automatically update the tokenizer paths used by your system:
#!/bin/bash
BERT_DIR="bert-base-uncased"
ROBERTA_DIR="roberta-base"
BART_DIR="facebook/bart-base"
PYTHON_SCRIPT="[YOUR ENV PATH]/lib/python3.9/site-packages/laion_clap/training/data.py"
update_tokenizer_path() {
local dir_name=$1
local tokenizer_variable=$2
local full_path
if [ -d "$dir_name" ]; then
full_path=$(realpath "$dir_name")
if [ -f "$PYTHON_SCRIPT" ]; then
sed -i "s|${tokenizer_variable}.from_pretrained(\".*\")|${tokenizer_variable}.from_pretrained(\"$full_path\")|" "$PYTHON_SCRIPT"
echo "Updated ${tokenizer_variable} path to $full_path."
else
echo "Error: The specified Python script does not exist."
exit 1
fi
else
echo "Error: The directory $dir_name does not exist in the current directory."
exit 1
fi
}
update_tokenizer_path "$BERT_DIR" "BertTokenizer"
update_tokenizer_path "$ROBERTA_DIR" "RobertaTokenizer"
update_tokenizer_path "$BART_DIR" "BartTokenizer"
echo "BERT, BART and RoBERTa Python script paths have been updated."
- The script provided is intended to adjust the tokenizer paths in the
data.py
file, found under/lib/python3.9/site-packages/laion_clap/training/
, within your specific environment. For those utilizing conda, you can determine your environment path by runningconda info --envs
. Then, substitute[YOUR ENV PATH]
in the script with this path. If your environment is configured differently, you'll need to update thePYTHON_SCRIPT
variable to correctly point to thedata.py
file. - Run the script. If it executes successfully, the tokenizer paths will be updated, allowing them to be loaded locally.
If your system is unable to access huggingface.co from the terminal and you want to calculate WavLM
based speaker similarity, you need to download the pretrained model first, as illustrated here.