shuaijiang
commited on
Commit
•
009de6d
1
Parent(s):
ce7c2ee
Update README.md
Browse files
README.md
CHANGED
@@ -38,7 +38,7 @@ transcription = transcriber("my_audio.wav")
|
|
38 |
## Fine-tuning
|
39 |
| Model | (Re)Sample Rate | Train Datasets | Fine-tuning (full or peft) |
|
40 |
|:----------------:|:-------:|:----------------------------------------------------------:|:-----------:|
|
41 |
-
| Belle-whisper-large-v3-turbo-zh | 16KHz | [AISHELL-1](https://openslr.magicdatatech.com/resources/33/) [AISHELL-2](https://www.aishelltech.com/aishell_2) [WenetSpeech](https://wenet.org.cn/WenetSpeech/) [HKUST](https://catalog.ldc.upenn.edu/LDC2005S15) | [
|
42 |
|
43 |
To incorporate punctuation marks without compromising performance, Lora fine-tuning was employed.
|
44 |
If you want to fine-thuning the model on your datasets, please reference to the [github repo](https://github.com/shuaijiang/Whisper-Finetune)
|
@@ -51,8 +51,7 @@ If you want to fine-thuning the model on your datasets, please reference to the
|
|
51 |
| whisper-large-v3-turbo | Chinese | 8.639 | 6.014 | 13.507 | 20.313 | 37.324 |
|
52 |
| Belle-whisper-large-v3-turbo-zh | Chinese | 2.x | 3.x | 8.x | 11.x | 16.x |
|
53 |
|
54 |
-
It is worth mentioning that compared to
|
55 |
-
And the punctation marks of Belle-whisper-large-v3-zh-punct are removed to compute the CER.
|
56 |
|
57 |
## Citation
|
58 |
|
|
|
38 |
## Fine-tuning
|
39 |
| Model | (Re)Sample Rate | Train Datasets | Fine-tuning (full or peft) |
|
40 |
|:----------------:|:-------:|:----------------------------------------------------------:|:-----------:|
|
41 |
+
| Belle-whisper-large-v3-turbo-zh | 16KHz | [AISHELL-1](https://openslr.magicdatatech.com/resources/33/) [AISHELL-2](https://www.aishelltech.com/aishell_2) [WenetSpeech](https://wenet.org.cn/WenetSpeech/) [HKUST](https://catalog.ldc.upenn.edu/LDC2005S15) | [full fine-tuning](https://github.com/shuaijiang/Whisper-Finetune) |
|
42 |
|
43 |
To incorporate punctuation marks without compromising performance, Lora fine-tuning was employed.
|
44 |
If you want to fine-thuning the model on your datasets, please reference to the [github repo](https://github.com/shuaijiang/Whisper-Finetune)
|
|
|
51 |
| whisper-large-v3-turbo | Chinese | 8.639 | 6.014 | 13.507 | 20.313 | 37.324 |
|
52 |
| Belle-whisper-large-v3-turbo-zh | Chinese | 2.x | 3.x | 8.x | 11.x | 16.x |
|
53 |
|
54 |
+
It is worth mentioning that compared to whisper-large-v3-turbo, Belle-whisper-large-v3-turbo-zh even has a significant improvement in complex acoustic scenes(such as wenetspeech_meeting).
|
|
|
55 |
|
56 |
## Citation
|
57 |
|