LeroyDyer commited on
Commit
318d04e
1 Parent(s): b0aab04

(Trained with Unsloth)

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "LeroyDyer/Mixtral_AI_CyberBrain_3_0",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -20,11 +20,13 @@
20
  "num_hidden_layers": 32,
21
  "num_key_value_heads": 8,
22
  "rms_norm_eps": 1e-05,
 
23
  "rope_theta": 10000.0,
24
  "sliding_window": 4096,
25
  "tie_word_embeddings": false,
26
- "torch_dtype": "bfloat16",
27
- "transformers_version": "4.37.0.dev0",
 
28
  "use_cache": true,
29
  "use_complex_talk_head": true,
30
  "use_complex_think_head": false,
 
1
  {
2
+ "_name_or_path": "LeroyDyer/Mixtral_AI_CyberBrain_3.0",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
20
  "num_hidden_layers": 32,
21
  "num_key_value_heads": 8,
22
  "rms_norm_eps": 1e-05,
23
+ "rope_scaling": null,
24
  "rope_theta": 10000.0,
25
  "sliding_window": 4096,
26
  "tie_word_embeddings": false,
27
+ "torch_dtype": "float16",
28
+ "transformers_version": "4.40.0.dev0",
29
+ "unsloth_version": "2024.3",
30
  "use_cache": true,
31
  "use_complex_talk_head": true,
32
  "use_complex_think_head": false,
model-00001-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43a7b89698db41c12893cdacf27f6b75620c70a5f2ad5167ad8c8e2b1b9f8b1f
3
  size 1889603392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13f0403298b237fdef972835a728b3dd73fb62b6ff5045fbcc1a7f83b9780b3a
3
  size 1889603392
model-00002-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13c367ff186b247da401fb2ac65dd27cfe8a3a384888f97e3c7626c9aff885a1
3
  size 1946243896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e7673c6932fd497f568911dd8b11582c13bdbe8cf5ca785f4c7b9aaed4ec2de
3
  size 1946243896
model-00003-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f54586c6e7e055fe2e2b7ae604bba9c44c16c3b17497b3970a0a2cb02b1ca957
3
  size 1979781392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98ab95d023531246ab6c0788384e4565a4023d351d63b421848613ece7a05b8b
3
  size 1979781392
model-00004-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4309ab251636ecb7dff8f87cbc3c415e63f1a44b11cef0818b6dcdc325735d81
3
  size 1946243936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:128fc561e53f00066aa158fb1921abf73e8a2ad656c91315d7a9aa3817e81e1d
3
  size 1946243936
model-00005-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a32d7068f3c9ef207142f92cb8fe66a99e58b6070a6ebc20d507655697c0b21c
3
  size 1979781416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b258a99da847688b4efb47d3bf6c2add1b4a9e3e04bcc63c37749b04d2423a1
3
  size 1979781416
model-00006-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de67bc9d39baa7d2cc3983e494a7ea83e9be19659ab8915b5afdec8d9987ead9
3
  size 1946243936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:151cee02c522a17a02f1ef81c26dee71df3775cd107728e5285b799270b09d35
3
  size 1946243936
model-00007-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd5c86883b18b92a8394ebef60d1cd9a36945c18c4707485da9a2914963cb8c5
3
  size 1979781416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:880b0efcadbc14cdcc85c7931ebe3bf4b07becf261c87f6a03f794dedf71d99f
3
  size 1979781416
model-00008-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77bca604fa9c3a207f86e0cbb4a4a5fe3d63498d8507f0b90491567f96925313
3
  size 815851048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dc3589a195594dd02eea2dc123d2857067eeeae5bbd5f0a2dc0e4a563becae2
3
  size 815851048
special_tokens_map.json CHANGED
@@ -17,13 +17,7 @@
17
  "rstrip": false,
18
  "single_word": false
19
  },
20
- "pad_token": {
21
- "content": "</s>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false
26
- },
27
  "unk_token": {
28
  "content": "<unk>",
29
  "lstrip": false,
 
17
  "rstrip": false,
18
  "single_word": false
19
  },
20
+ "pad_token": "</s>",
 
 
 
 
 
 
21
  "unk_token": {
22
  "content": "<unk>",
23
  "lstrip": false,
tokenizer_config.json CHANGED
@@ -57,6 +57,7 @@
57
  "sp_model_kwargs": {},
58
  "spaces_between_special_tokens": false,
59
  "tokenizer_class": "LlamaTokenizer",
 
60
  "unk_token": "<unk>",
61
  "use_default_system_prompt": false
62
  }
 
57
  "sp_model_kwargs": {},
58
  "spaces_between_special_tokens": false,
59
  "tokenizer_class": "LlamaTokenizer",
60
+ "truncation": true,
61
  "unk_token": "<unk>",
62
  "use_default_system_prompt": false
63
  }