sagorsarker
commited on
Commit
•
0c58c9f
1
Parent(s):
8e25d45
Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,17 @@
|
|
1 |
---
|
2 |
language:
|
3 |
- bn
|
|
|
|
|
|
|
|
|
4 |
---
|
5 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
language:
|
3 |
- bn
|
4 |
+
license: apache-2.0
|
5 |
+
datasets:
|
6 |
+
- uonlp/CulturaX
|
7 |
+
pipeline_tag: text-generation
|
8 |
---
|
9 |
|
10 |
+
# TituLM-1B-BN-V1
|
11 |
+
|
12 |
+
TituLM-1B-BN-V1 is a large language model specifically trained for generating and understanding Bangla text. Utilizing a decoder-style transformer architecture, this model has been extensively trained on a dataset comprising 4.51 billion Bangla tokens. This model is the part of iterative train and release Bangla LLM from Hishab.
|
13 |
+
|
14 |
+
## Training
|
15 |
+
The training process was managed using the robust framework provided by MosaicML's [llm-foundry](https://github.com/mosaicml/llm-foundry) repository. Throughout the training phase, titulm-1b-bn-v1 underwent a total of 59 iterations, allowing for iterative refinements and optimization
|
16 |
+
|
17 |
+
## Datasets
|