how to fine tune this model to get better polish translation
can someone guide me how i can fine tune this model to get better polish translations of paragraphs and large passages with better quality grammar
@Devdarshan much better is to find good translation dataset and fine-tuning for Mistral 7B model, check my repo
Anyone tried to generate with this one? The output is always empty (or rather containing only bos and eos tokens). Other Opus-mt-tc-en-... models generate just fine.
Hi
@Tomsterrus
, thanks for reporting this. Indeed there seem to be a problem with the model, probably introduced by newer versions of the transformers
library. I sadly won't have time to look into this myself, but maybe re-exporting the original model from the Tatoeba repository would work! I'll add a mention in the readme.