# English to Xitsonga Author: Laura Martinus ## Data - The JW300 dataset. ## Model - Default Masakhane Transformer translation model. - Link to google drive folder with model(https://drive.google.com/open?id=1onvLxPsRoem2KGnykp2vLEIe4GldDl84) ## Analysis - TODO # Results - BLEU dev: 35.07 - BLEU test: 44.15 - Note: It is probably best to train this model for longer, as it timed out on Google Colab - Note: Will likely benefit from optimising the number of BPE codes Ran with BPE set to 40 000 instead of 4 000 and got the following: - BLEU dev: 39.01 - BLEU test: 46.41