shuaijiang commited on
Commit
009de6d
1 Parent(s): ce7c2ee

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -38,7 +38,7 @@ transcription = transcriber("my_audio.wav")
38
  ## Fine-tuning
39
  | Model | (Re)Sample Rate | Train Datasets | Fine-tuning (full or peft) |
40
  |:----------------:|:-------:|:----------------------------------------------------------:|:-----------:|
41
- | Belle-whisper-large-v3-turbo-zh | 16KHz | [AISHELL-1](https://openslr.magicdatatech.com/resources/33/) [AISHELL-2](https://www.aishelltech.com/aishell_2) [WenetSpeech](https://wenet.org.cn/WenetSpeech/) [HKUST](https://catalog.ldc.upenn.edu/LDC2005S15) | [lora fine-tuning](https://github.com/shuaijiang/Whisper-Finetune) |
42
 
43
  To incorporate punctuation marks without compromising performance, Lora fine-tuning was employed.
44
  If you want to fine-thuning the model on your datasets, please reference to the [github repo](https://github.com/shuaijiang/Whisper-Finetune)
@@ -51,8 +51,7 @@ If you want to fine-thuning the model on your datasets, please reference to the
51
  | whisper-large-v3-turbo | Chinese | 8.639 | 6.014 | 13.507 | 20.313 | 37.324 |
52
  | Belle-whisper-large-v3-turbo-zh | Chinese | 2.x | 3.x | 8.x | 11.x | 16.x |
53
 
54
- It is worth mentioning that compared to Belle-whisper-large-v3-zh, Belle-whisper-large-v3-zh-punct even has a slight improvement in complex acoustic scenes(such as wenetspeech_meeting).
55
- And the punctation marks of Belle-whisper-large-v3-zh-punct are removed to compute the CER.
56
 
57
  ## Citation
58
 
 
38
  ## Fine-tuning
39
  | Model | (Re)Sample Rate | Train Datasets | Fine-tuning (full or peft) |
40
  |:----------------:|:-------:|:----------------------------------------------------------:|:-----------:|
41
+ | Belle-whisper-large-v3-turbo-zh | 16KHz | [AISHELL-1](https://openslr.magicdatatech.com/resources/33/) [AISHELL-2](https://www.aishelltech.com/aishell_2) [WenetSpeech](https://wenet.org.cn/WenetSpeech/) [HKUST](https://catalog.ldc.upenn.edu/LDC2005S15) | [full fine-tuning](https://github.com/shuaijiang/Whisper-Finetune) |
42
 
43
  To incorporate punctuation marks without compromising performance, Lora fine-tuning was employed.
44
  If you want to fine-thuning the model on your datasets, please reference to the [github repo](https://github.com/shuaijiang/Whisper-Finetune)
 
51
  | whisper-large-v3-turbo | Chinese | 8.639 | 6.014 | 13.507 | 20.313 | 37.324 |
52
  | Belle-whisper-large-v3-turbo-zh | Chinese | 2.x | 3.x | 8.x | 11.x | 16.x |
53
 
54
+ It is worth mentioning that compared to whisper-large-v3-turbo, Belle-whisper-large-v3-turbo-zh even has a significant improvement in complex acoustic scenes(such as wenetspeech_meeting).
 
55
 
56
  ## Citation
57