Update README.md
Browse files
README.md
CHANGED
@@ -16,14 +16,14 @@ tags:
|
|
16 |
|
17 |
InfCapital LLama2-7b clone of [Meta's Llama 2 7B Chat](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf).
|
18 |
|
19 |
-
Adopt for Vietnamese continued pretraining or fine-tuning by extend vocab size from 32,000 to
|
20 |
|
21 |
|
22 |
## Model Architecture
|
23 |
```
|
24 |
LlamaForCausalLM(
|
25 |
(model): LlamaModel(
|
26 |
-
(embed_tokens): Embedding(
|
27 |
(layers): ModuleList(
|
28 |
(0-31): 32 x LlamaDecoderLayer(
|
29 |
(self_attn): LlamaAttention(
|
@@ -45,6 +45,6 @@ LlamaForCausalLM(
|
|
45 |
)
|
46 |
(norm): LlamaRMSNorm()
|
47 |
)
|
48 |
-
(lm_head): Linear(in_features=4096, out_features=
|
49 |
)
|
50 |
```
|
|
|
16 |
|
17 |
InfCapital LLama2-7b clone of [Meta's Llama 2 7B Chat](https://huggingface.co/meta-llama/Llama-2-7b-chat-hf).
|
18 |
|
19 |
+
Adopt for Vietnamese continued pretraining or fine-tuning by extend vocab size from 32,000 to 44,800. Vocabs added by training sentencepiece method from dataset vnnews-corpus.
|
20 |
|
21 |
|
22 |
## Model Architecture
|
23 |
```
|
24 |
LlamaForCausalLM(
|
25 |
(model): LlamaModel(
|
26 |
+
(embed_tokens): Embedding(44800, 4096)
|
27 |
(layers): ModuleList(
|
28 |
(0-31): 32 x LlamaDecoderLayer(
|
29 |
(self_attn): LlamaAttention(
|
|
|
45 |
)
|
46 |
(norm): LlamaRMSNorm()
|
47 |
)
|
48 |
+
(lm_head): Linear(in_features=4096, out_features=44800, bias=False)
|
49 |
)
|
50 |
```
|