Datasets:
metadata
language:
- en
HuggingFaceFW/fineweb filter Malaysian context
What is it?
We filter the original 🍷 FineWeb dataset that consists more than 15T tokens on simple Malaysian keywords.
Total tokens for the filtered dataset is 174102784199 tokens, 174B tokens.
How we do it?
- We filter rows using
{'malay', 'malaysia', 'melayu', 'bursa', 'ringgit'}
keywords on r5.16xlarge EC2 instance for 7 days. - We calculate total tokens using
tiktoken.encoding_for_model("gpt2")
on c7a.24xlarge EC2 instance for 1 hour.
source code at https://github.com/mesolitica/malaysian-dataset/tree/master/corpus/fineweb
Why we do it?
So anybody can use this filtered corpus to pretrain, continue pretraining or generate synthetic dataset for their own use cases.