Upload folder using huggingface_hub

#3
by pcuenq HF staff - opened
config.json CHANGED
@@ -22,6 +22,9 @@
22
  "qk_layernorm": true,
23
  "rope_scaling": null,
24
  "rope_theta": 25000.0,
 
 
 
25
  "tie_word_embeddings": false,
26
  "torch_dtype": "bfloat16",
27
  "transformers_version": "4.35.0.dev0",
 
22
  "qk_layernorm": true,
23
  "rope_scaling": null,
24
  "rope_theta": 25000.0,
25
+ "text_config": {
26
+ "model_type": "persimmon"
27
+ },
28
  "tie_word_embeddings": false,
29
  "torch_dtype": "bfloat16",
30
  "transformers_version": "4.35.0.dev0",
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86b2ecb494b4cbbb2eabe0e28510eed95db351e1001019356e5d45eabfaedaeb
3
- size 9956341424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0360c23415254b77cf052e5dc5b57ecd98f28d65e1e4cb7a222c9fa35e97c0e
3
+ size 9934219312
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6afb1f47511777d65d3e8cd1d6dcc5f826e550f4870b70afba2623a316512635
3
- size 8860203352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb67849dfa2f360d8bf653386a3cbc36002b8921de62f9916f70872ab929e3eb
3
+ size 8882334152
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff