Update README.md
Browse files
README.md
CHANGED
@@ -56,11 +56,18 @@ model-index:
|
|
56 |
whisper tiny fine-tuned on a very big collection of vietnamese speech datasets
|
57 |
|
58 |
TODO:
|
59 |
-
- [x] training then publish checkpoint
|
60 |
-
- [x] evaluate WER on Common Voice & FLEURS
|
61 |
- [ ] convert to `openai-whisper`, `whisper.cpp`, `faster-whisper`
|
62 |
-
- [ ] convert to ONNX: to try
|
|
|
63 |
|
64 |
21k steps, warm-up 5%, batch size 16×2 (kaggle free T4×2)
|
65 |
|
66 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
56 |
whisper tiny fine-tuned on a very big collection of vietnamese speech datasets
|
57 |
|
58 |
TODO:
|
59 |
+
- [x] training then publish checkpoint
|
60 |
+
- [x] evaluate WER on Common Voice & FLEURS & VIVOS
|
61 |
- [ ] convert to `openai-whisper`, `whisper.cpp`, `faster-whisper`
|
62 |
+
- [ ] convert to ONNX: to try https://github.com/k2-fsa/sherpa-onnx & https://github.com/zhuzilin/whisper-openvino
|
63 |
+
- [ ] convert to TensorRT: https://github.com/openai/whisper/discussions/169
|
64 |
|
65 |
21k steps, warm-up 5%, batch size 16×2 (kaggle free T4×2)
|
66 |
|
67 |
+
manually evaluate WER on test set - vietnamese part:
|
68 |
+
| @ `float16` | `CommonVoice v16.1` | `FLEURS` | `VIVOS` |
|
69 |
+
|---|---|---|---|
|
70 |
+
| original `whisper-tiny` | >100% | 88.6% | 62.5% |
|
71 |
+
| this model | 26.6% | 37.1% | 18.7% |
|
72 |
+
|
73 |
+
all training + evaluation scripts are on my repo: https://github.com/phineas-pta/fine-tune-whisper-vi
|