gpt2-medium-persian / README.md
m3hrdadfi's picture
Add normalization steps
a90e731
|
raw
history blame
1.23 kB

GPT2 - Persian

Scripts

Normalizer

from src.normalizer import normalize

input_text = "ὑ蕉Ұ제ṅ尘̲改座◦花芝秀黄天자埃澤ಿ ˈazbab اینجا ایران خانه‌شما است؟!۱۲۳۱۲۳۱۳۱۲ اَلْحُرُوفُ ٱلْعَرَبِیَّة"
print(normalize(input_text))

Output:

azbab اینجا ایران خانه‌شما است ؟ ! 1231231312 الحروف لعربیه

Training tokenizer

python train_tokenizer.py --dataset_name oscar --dataset_config_name unshuffled_deduplicated_als --vocab_size 42000

Configuration

python create_config.py --name_or_path gpt2-medium --params '{"vocab_size": 42000}'

Normalization steps

Steps:

  • Remove stretched words such as ســــــــــلام

  • Remove links, user-mentioning (such as @jane_doe)

  • Remove Telegram, Instagram advertisements, or posts (a whole record)

  • Remove advertisement records

  • Remove separated words (or the whole record) which are showing up as an individual record, while they are just the tags at the end of the post (such as بلاب ... بلاب ... ورزشی، خبری، سیاسی، اجتماعی، خانوده)