beamaia commited on
Commit
df8371f
1 Parent(s): 0690acb

Model save

Browse files
README.md CHANGED
@@ -18,15 +18,15 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.6907
22
- - Rewards/chosen: 0.0077
23
- - Rewards/rejected: 0.0025
24
- - Rewards/accuracies: 0.0472
25
- - Rewards/margins: 0.0052
26
- - Logps/rejected: -11.0078
27
- - Logps/chosen: -7.5850
28
- - Logits/rejected: -2.2114
29
- - Logits/chosen: -2.2106
30
 
31
  ## Model description
32
 
@@ -54,7 +54,7 @@ The following hyperparameters were used during training:
54
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
  - lr_scheduler_type: linear
56
  - lr_scheduler_warmup_ratio: 0.1
57
- - training_steps: 5
58
  - mixed_precision_training: Native AMP
59
 
60
  ### Training results
 
18
 
19
  This model is a fine-tuned version of [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.6931
22
+ - Rewards/chosen: 0.0
23
+ - Rewards/rejected: 0.0
24
+ - Rewards/accuracies: 0.0
25
+ - Rewards/margins: 0.0
26
+ - Logps/rejected: -8.1759
27
+ - Logps/chosen: -5.6437
28
+ - Logits/rejected: -2.2170
29
+ - Logits/chosen: -2.2168
30
 
31
  ## Model description
32
 
 
54
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
  - lr_scheduler_type: linear
56
  - lr_scheduler_warmup_ratio: 0.1
57
+ - training_steps: 1
58
  - mixed_precision_training: Native AMP
59
 
60
  ### Training results
adapter_config.json CHANGED
@@ -19,10 +19,10 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "o_proj",
23
  "q_proj",
24
- "v_proj",
25
- "k_proj"
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "q_proj",
23
+ "o_proj",
24
+ "k_proj",
25
+ "v_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01eff22f270c17d6a16cc359decf3316f37d553aa0632d6211fcd6c7a38470a6
3
  size 27297032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3452cef1e3289d786650c8637a888bfeabe74d15bf9a42d89e2cd7de7c0f993
3
  size 27297032
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a69cb15297afce38a7ecff1f96dc6e9d8a134c2fddf7a65c5382256f3bfe8c25
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e3ef142ff969fa6655db0c09c77ad43010cacec98d66a168e060189531be034
3
  size 5112