# Evaluation Install packages for evaluation: ```bash pip install -e .[eval] ``` ## Generating Samples for Evaluation ### Prepare Test Datasets 1. *Seed-TTS testset*: Download from [seed-tts-eval](https://github.com/BytedanceSpeech/seed-tts-eval). 2. *LibriSpeech test-clean*: Download from [OpenSLR](http://www.openslr.org/12/). 3. Unzip the downloaded datasets and place them in the `data/` directory. 4. Update the path for *LibriSpeech test-clean* data in `src/f5_tts/eval/eval_infer_batch.py` 5. Our filtered LibriSpeech-PC 4-10s subset: `data/librispeech_pc_test_clean_cross_sentence.lst` ### Batch Inference for Test Set To run batch inference for evaluations, execute the following commands: ```bash # batch inference for evaluations accelerate config # if not set before bash src/f5_tts/eval/eval_infer_batch.sh ``` ## Objective Evaluation on Generated Results ### Download Evaluation Model Checkpoints 1. Chinese ASR Model: [Paraformer-zh](https://huggingface.co/funasr/paraformer-zh) 2. English ASR Model: [Faster-Whisper](https://huggingface.co/Systran/faster-whisper-large-v3) 3. WavLM Model: Download from [Google Drive](https://drive.google.com/file/d/1-aE1NfzpRCLxA4GUxX9ITI3F9LlbtEGP/view). Then update in the following scripts with the paths you put evaluation model ckpts to. ### Objective Evaluation Update the path with your batch-inferenced results, and carry out WER / SIM evaluations: ```bash # Evaluation for Seed-TTS test set python src/f5_tts/eval/eval_seedtts_testset.py # Evaluation for LibriSpeech-PC test-clean (cross-sentence) python src/f5_tts/eval/eval_librispeech_test_clean.py ```