alexmarques commited on
Commit
c2b4a2e
1 Parent(s): 6644d0d

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "microsoft__Phi-3-medium-128k-instruct",
3
  "architectures": [
4
  "Phi3ForCausalLM"
5
  ],
@@ -38,13 +38,7 @@
38
  ],
39
  "kv_cache_scheme": null,
40
  "quant_method": "compressed-tensors",
41
- "quantization_status": "frozen",
42
- "sparsity_config": {
43
- "format": "dense",
44
- "global_sparsity": 1.3619957547314265,
45
- "registry_requires_subclass": false,
46
- "sparsity_structure": "unstructured"
47
- }
48
  },
49
  "embd_pdrop": 0.0,
50
  "eos_token_id": 32000,
@@ -200,7 +194,7 @@
200
  "sliding_window": 131072,
201
  "tie_word_embeddings": false,
202
  "torch_dtype": "bfloat16",
203
- "transformers_version": "4.42.3",
204
  "use_cache": true,
205
  "vocab_size": 32064
206
  }
 
1
  {
2
+ "_name_or_path": "/root/.cache/huggingface/hub/models--microsoft--Phi-3-medium-128k-instruct/snapshots/cae1d42b5577398fd1be9f0746052562ae552886",
3
  "architectures": [
4
  "Phi3ForCausalLM"
5
  ],
 
38
  ],
39
  "kv_cache_scheme": null,
40
  "quant_method": "compressed-tensors",
41
+ "quantization_status": "frozen"
 
 
 
 
 
 
42
  },
43
  "embd_pdrop": 0.0,
44
  "eos_token_id": 32000,
 
194
  "sliding_window": 131072,
195
  "tie_word_embeddings": false,
196
  "torch_dtype": "bfloat16",
197
+ "transformers_version": "4.44.0",
198
  "use_cache": true,
199
  "vocab_size": 32064
200
  }
generation_config.json CHANGED
@@ -7,5 +7,5 @@
7
  32007
8
  ],
9
  "pad_token_id": 32000,
10
- "transformers_version": "4.42.3"
11
  }
 
7
  32007
8
  ],
9
  "pad_token_id": 32000,
10
+ "transformers_version": "4.44.0"
11
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c78c728fbf14bc3c71982b7d19f55c7654adf1ed662a166bb98bf624fdf82df
3
  size 4825817208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b996802f676dfe1a9b7ec85969c81460b59b1d7a8f4c26dd9a40e0cc8e2dd370
3
  size 4825817208
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23f671b14a63c738d3865603c44c9d4792fc1cf10979a26ad94ba36d13c001fe
3
  size 4956409160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d0b30886ec45d29d9ec24d32bcb232657c55e06f90e5695a17fa212475f6037
3
  size 4956409160
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b763d02e15c5bd3c05398a3adab890e84c84208778d480f5b94fc00092f93da7
3
  size 4511129936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1843b208a870fb88b9d4d98abe7b683226b9671eb88b503c9db7fbf707d160fc
3
  size 4511129936
recipe.yaml CHANGED
@@ -4,7 +4,5 @@ quant_stage:
4
  sequential_update: false
5
  dampening_frac: 0.01
6
  ignore: [lm_head]
7
- config_groups:
8
- group_0:
9
- targets: [Linear]
10
- weights: {num_bits: 8, type: int, symmetric: true, strategy: channel}
 
4
  sequential_update: false
5
  dampening_frac: 0.01
6
  ignore: [lm_head]
7
+ scheme: W8A16
8
+ targets: Linear