kstecenko commited on
Commit
6be968a
1 Parent(s): 9ac12b8

Upload model

Browse files
Files changed (3) hide show
  1. README.md +3 -56
  2. adapter_config.json +3 -3
  3. adapter_model.bin +1 -1
README.md CHANGED
@@ -18,6 +18,7 @@ base_model: PY007/TinyLlama-1.1B-Chat-v0.1
18
 
19
 
20
  - **Developed by:** [More Information Needed]
 
21
  - **Shared by [optional]:** [More Information Needed]
22
  - **Model type:** [More Information Needed]
23
  - **Language(s) (NLP):** [More Information Needed]
@@ -76,7 +77,7 @@ Use the code below to get started with the model.
76
 
77
  ### Training Data
78
 
79
- <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
80
 
81
  [More Information Needed]
82
 
@@ -107,7 +108,7 @@ Use the code below to get started with the model.
107
 
108
  #### Testing Data
109
 
110
- <!-- This should link to a Data Card if possible. -->
111
 
112
  [More Information Needed]
113
 
@@ -233,58 +234,4 @@ The following `bitsandbytes` quantization config was used during training:
233
  ### Framework versions
234
 
235
 
236
- - PEFT 0.6.0.dev0
237
- ## Training procedure
238
-
239
-
240
- The following `bitsandbytes` quantization config was used during training:
241
- - load_in_8bit: False
242
- - load_in_4bit: True
243
- - llm_int8_threshold: 6.0
244
- - llm_int8_skip_modules: None
245
- - llm_int8_enable_fp32_cpu_offload: False
246
- - llm_int8_has_fp16_weight: False
247
- - bnb_4bit_quant_type: nf4
248
- - bnb_4bit_use_double_quant: True
249
- - bnb_4bit_compute_dtype: bfloat16
250
-
251
- ### Framework versions
252
-
253
-
254
- - PEFT 0.6.0.dev0
255
- ## Training procedure
256
-
257
-
258
- The following `bitsandbytes` quantization config was used during training:
259
- - load_in_8bit: False
260
- - load_in_4bit: True
261
- - llm_int8_threshold: 6.0
262
- - llm_int8_skip_modules: None
263
- - llm_int8_enable_fp32_cpu_offload: False
264
- - llm_int8_has_fp16_weight: False
265
- - bnb_4bit_quant_type: nf4
266
- - bnb_4bit_use_double_quant: True
267
- - bnb_4bit_compute_dtype: bfloat16
268
-
269
- ### Framework versions
270
-
271
-
272
- - PEFT 0.6.0.dev0
273
- ## Training procedure
274
-
275
-
276
- The following `bitsandbytes` quantization config was used during training:
277
- - load_in_8bit: False
278
- - load_in_4bit: True
279
- - llm_int8_threshold: 6.0
280
- - llm_int8_skip_modules: None
281
- - llm_int8_enable_fp32_cpu_offload: False
282
- - llm_int8_has_fp16_weight: False
283
- - bnb_4bit_quant_type: nf4
284
- - bnb_4bit_use_double_quant: True
285
- - bnb_4bit_compute_dtype: bfloat16
286
-
287
- ### Framework versions
288
-
289
-
290
  - PEFT 0.6.0.dev0
 
18
 
19
 
20
  - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
  - **Shared by [optional]:** [More Information Needed]
23
  - **Model type:** [More Information Needed]
24
  - **Language(s) (NLP):** [More Information Needed]
 
77
 
78
  ### Training Data
79
 
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
 
82
  [More Information Needed]
83
 
 
108
 
109
  #### Testing Data
110
 
111
+ <!-- This should link to a Dataset Card if possible. -->
112
 
113
  [More Information Needed]
114
 
 
234
  ### Framework versions
235
 
236
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
237
  - PEFT 0.6.0.dev0
adapter_config.json CHANGED
@@ -16,10 +16,10 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "self_attn.q_proj",
20
  "self_attn.o_proj",
21
- "self_attn.k_proj",
22
- "self_attn.v_proj"
 
23
  ],
24
  "task_type": "CAUSAL_LM"
25
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "self_attn.o_proj",
20
+ "self_attn.v_proj",
21
+ "self_attn.q_proj",
22
+ "self_attn.k_proj"
23
  ],
24
  "task_type": "CAUSAL_LM"
25
  }
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c222ffd04ca46128f0b7b352864385bc6880b8a54372ca55b84b47e531046d78
3
  size 18085517
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d51a883e4bc9d918ea6a23c4026f7325d613ddd33d62b65c2c9a2319cf05a77
3
  size 18085517