Spaces:
Running
on
Zero
Running
on
Zero
# Amphion Evaluation Recipe | |
## Supported Evaluation Metrics | |
Until now, Amphion Evaluation has supported the following objective metrics: | |
- **F0 Modeling**: | |
- F0 Pearson Coefficients (FPC) | |
- F0 Periodicity Root Mean Square Error (PeriodicityRMSE) | |
- F0 Root Mean Square Error (F0RMSE) | |
- Voiced/Unvoiced F1 Score (V/UV F1) | |
- **Energy Modeling**: | |
- Energy Root Mean Square Error (EnergyRMSE) | |
- Energy Pearson Coefficients (EnergyPC) | |
- **Intelligibility**: | |
- Character Error Rate (CER) based on [Whipser](https://github.com/openai/whisper) | |
- Word Error Rate (WER) based on [Whipser](https://github.com/openai/whisper) | |
- **Spectrogram Distortion**: | |
- Frechet Audio Distance (FAD) | |
- Mel Cepstral Distortion (MCD) | |
- Multi-Resolution STFT Distance (MSTFT) | |
- Perceptual Evaluation of Speech Quality (PESQ) | |
- Short Time Objective Intelligibility (STOI) | |
- Scale Invariant Signal to Distortion Ratio (SISDR) | |
- Scale Invariant Signal to Noise Ratio (SISNR) | |
- **Speaker Similarity**: | |
- Cosine similarity based on: | |
- [Rawnet3](https://github.com/Jungjee/RawNet) | |
- [Resemblyzer](https://github.com/resemble-ai/Resemblyzer) | |
- [WavLM](https://huggingface.co./microsoft/wavlm-base-plus-sv) | |
We provide a recipe to demonstrate how to objectively evaluate your generated audios. There are three steps in total: | |
1. Pretrained Models Preparation | |
2. Audio Data Preparation | |
3. Evaluation | |
## 1. Pretrained Models Preparation | |
If you want to calculate `RawNet3` based speaker similarity, you need to download the pretrained model first, as illustrated [here](../../pretrained/README.md). | |
## 2. Audio Data Preparation | |
Prepare reference audios and generated audios in two folders, the `ref_dir` contains the reference audio and the `gen_dir` contains the generated audio. Here is an example. | |
```plaintext | |
┣ {ref_dir} | |
┃ ┣ sample1.wav | |
┃ ┣ sample2.wav | |
┣ {gen_dir} | |
┃ ┣ sample1.wav | |
┃ ┣ sample2.wav | |
``` | |
You have to make sure that the pairwise **reference audio and generated audio are named the same**, as illustrated above (sample1 to sample1, sample2 to sample2). | |
## 3. Evaluation | |
Run the `run.sh` with specified refenrece folder, generated folder, dump folder and metrics. | |
```bash | |
cd Amphion | |
sh egs/metrics/run.sh \ | |
--reference_folder [Your path to the reference audios] \ | |
--generated_folder [Your path to the generated audios] \ | |
--dump_folder [Your path to dump the objective results] \ | |
--metrics [The metrics you need] \ | |
--fs [Optional. To calculate all metrics in the specified sampling rate] \ | |
--similarity_model [Optional. To choose the model for calculating the speaker similarity. Currently "rawnet", "wavlm" and "resemblyzer" are available. Default to "wavlm"] \ | |
--similarity_mode [Optional. To choose the mode for calculating the speaker similarity. "pairwith" for calculating a series of ground truth / prediction audio pairs to obtain the speaker similarity, and "overall" for computing the average score with all possible pairs between the refernece folder and generated folder. Default to "pairwith"] \ | |
--intelligibility_mode [Optionoal. To choose the mode for computing CER and WER. "gt_audio" means selecting the recognition content of the reference audio as the target, "gt_content" means using transcription as the target. Default to "gt_audio"] \ | |
--ltr_path [Optional. Path to the transcription file] \ | |
--language [Optional. Language for computing CER and WER. Default to "english"] | |
``` | |
As for the metrics, an example is provided below: | |
```bash | |
--metrics "mcd pesq fad" | |
``` | |
All currently available metrics keywords are listed below: | |
| Keys | Description | | |
| ------------------------- | ------------------------------------------ | | |
| `fpc` | F0 Pearson Coefficients | | |
| `f0_periodicity_rmse` | F0 Periodicity Root Mean Square Error | | |
| `f0rmse` | F0 Root Mean Square Error | | |
| `v_uv_f1` | Voiced/Unvoiced F1 Score | | |
| `energy_rmse` | Energy Root Mean Square Error | | |
| `energy_pc` | Energy Pearson Coefficients | | |
| `cer` | Character Error Rate | | |
| `wer` | Word Error Rate | | |
| `similarity` | Speaker Similarity | |
| `fad` | Frechet Audio Distance | | |
| `mcd` | Mel Cepstral Distortion | | |
| `mstft` | Multi-Resolution STFT Distance | | |
| `pesq` | Perceptual Evaluation of Speech Quality | | |
| `si_sdr` | Scale Invariant Signal to Distortion Ratio | | |
| `si_snr` | Scale Invariant Signal to Noise Ratio | | |
| `stoi` | Short Time Objective Intelligibility | | |
For example, if want to calculate the speaker similarity between the synthesized audio and the reference audio with the same content, run: | |
```bash | |
sh egs/metrics/run.sh \ | |
--reference_folder [Your path to the reference audios] \ | |
--generated_folder [Your path to the generated audios] \ | |
--dump_folder [Your path to dump the objective results] \ | |
--metrics "similarity" \ | |
--similarity_model [Optional. To choose the model for calculating the speaker similarity. Currently "rawnet", "wavlm" and "resemblyzer" are available. Default to "wavlm"] \ | |
--similarity_mode "pairwith" \ | |
``` | |
If you don't have the reference audio with the same content, run the following to get the conteng-free similarity score: | |
```bash | |
sh egs/metrics/run.sh \ | |
--reference_folder [Your path to the reference audios] \ | |
--generated_folder [Your path to the generated audios] \ | |
--dump_folder [Your path to dump the objective results] \ | |
--metrics "similarity" \ | |
--similarity_model [Optional. To choose the model for calculating the speaker similarity. Currently "rawnet", "wavlm" and "resemblyzer" are available. Default to "wavlm"] \ | |
--similarity_mode "overall" \ | |
``` | |
## Troubleshooting | |
### FAD (Using Offline Models) | |
If your system is unable to access huggingface.co from the terminal, you might run into an error like "OSError: Can't load tokenizer for ...". To work around this, follow these steps to use local models: | |
1. Download the [bert-base-uncased](https://huggingface.co./bert-base-uncased), [roberta-base](https://huggingface.co./roberta-base), and [facebook/bart-base](https://huggingface.co./facebook/bart-base) models from `huggingface.co`. Ensure that the models are complete and uncorrupted. Place these directories within `Amphion/pretrained`. For a detailed file structure reference, see [This README](../../pretrained/README.md#optional-model-dependencies-for-evaluation) under `Amphion/pretrained`. | |
2. Inside the `Amphion/pretrained` directory, create a bash script with the content outlined below. This script will automatically update the tokenizer paths used by your system: | |
```bash | |
#!/bin/bash | |
BERT_DIR="bert-base-uncased" | |
ROBERTA_DIR="roberta-base" | |
BART_DIR="facebook/bart-base" | |
PYTHON_SCRIPT="[YOUR ENV PATH]/lib/python3.9/site-packages/laion_clap/training/data.py" | |
update_tokenizer_path() { | |
local dir_name=$1 | |
local tokenizer_variable=$2 | |
local full_path | |
if [ -d "$dir_name" ]; then | |
full_path=$(realpath "$dir_name") | |
if [ -f "$PYTHON_SCRIPT" ]; then | |
sed -i "s|${tokenizer_variable}.from_pretrained(\".*\")|${tokenizer_variable}.from_pretrained(\"$full_path\")|" "$PYTHON_SCRIPT" | |
echo "Updated ${tokenizer_variable} path to $full_path." | |
else | |
echo "Error: The specified Python script does not exist." | |
exit 1 | |
fi | |
else | |
echo "Error: The directory $dir_name does not exist in the current directory." | |
exit 1 | |
fi | |
} | |
update_tokenizer_path "$BERT_DIR" "BertTokenizer" | |
update_tokenizer_path "$ROBERTA_DIR" "RobertaTokenizer" | |
update_tokenizer_path "$BART_DIR" "BartTokenizer" | |
echo "BERT, BART and RoBERTa Python script paths have been updated." | |
``` | |
3. The script provided is intended to adjust the tokenizer paths in the `data.py` file, found under `/lib/python3.9/site-packages/laion_clap/training/`, within your specific environment. For those utilizing conda, you can determine your environment path by running `conda info --envs`. Then, substitute `[YOUR ENV PATH]` in the script with this path. If your environment is configured differently, you'll need to update the `PYTHON_SCRIPT` variable to correctly point to the `data.py` file. | |
4. Run the script. If it executes successfully, the tokenizer paths will be updated, allowing them to be loaded locally. | |
### WavLM-based Speaker Similarity (Using Offline Models) | |
If your system is unable to access huggingface.co from the terminal and you want to calculate `WavLM` based speaker similarity, you need to download the pretrained model first, as illustrated [here](../../pretrained/README.md). |