edaiofficial's picture
initial commits
78aa4ee
|
raw
history blame
595 Bytes

English to Xitsonga

Author: Laura Martinus

Data

- The JW300 dataset.

Model

- Default Masakhane Transformer translation model.
- Link to google drive folder with model(https://drive.google.com/open?id=1onvLxPsRoem2KGnykp2vLEIe4GldDl84)

Analysis

 - TODO

Results

- BLEU dev: 35.07
- BLEU test: 44.15 

- Note: It is probably best to train this model for longer, as it timed out on Google Colab
- Note: Will likely benefit from optimising the number of BPE codes

Ran with BPE set to 40 000 instead of 4 000 and got the following:
- BLEU dev: 39.01
- BLEU test: 46.41