mylesgoose
commited on
Commit
•
a80e91a
1
Parent(s):
b9dc56c
Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,12 @@
|
|
1 |
-
---
|
2 |
-
license: other
|
3 |
-
license_name: meta
|
4 |
-
license_link: https://ai.meta.com/llama/license
|
5 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
license_name: meta
|
4 |
+
license_link: https://ai.meta.com/llama/license
|
5 |
+
---
|
6 |
+
pretrain dont use will ouput garbage
|
7 |
+
|
8 |
+
adjusted tensors for the new tokenizer.
|
9 |
+
Some weights of LlamaForCausalLM were not initialized from the model checkpoint at mylesgoose/Meta-Llama-3.1-8B-Instruct-goose-abliterated-reflection and are newly initialized because the shapes did not match:
|
10 |
+
- model.embed_tokens.weight: found shape torch.Size([128256, 4096]) in the checkpoint and torch.Size([128262, 4096]) in the model instantiated
|
11 |
+
- lm_head.weight: found shape torch.Size([128256, 4096]) in the checkpoint and torch.Size([128262, 4096]) in the model instantiated
|
12 |
+
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|