Mel-Iza0 commited on
Commit
c844c8c
1 Parent(s): 324ac13

Training in progress, step 100, checkpoint

Browse files
checkpoint-100/adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
  "v_proj",
25
  "o_proj",
26
- "k_proj"
 
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "v_proj",
24
  "o_proj",
25
+ "k_proj",
26
+ "q_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
checkpoint-100/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a09efd490b1230a03de342cc81ebd092ee71bdf105cf41f55fd564ccfb7fe67
3
  size 54560368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b522f4d26fb3f5c7f67fb6d1e57c69d93d012b49ae040b33d304619d5e700627
3
  size 54560368
checkpoint-100/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31ec931eae12d877a7421bb24a3cd866d5cb82e68cb89575eac37dd6eb7bf739
3
  size 109267450
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be7f37e8a7a420f9026fe26771af86c4119f4a70b6df31a422e59a46e283f0b9
3
  size 109267450
checkpoint-100/trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.47333332896232605,
3
  "best_model_checkpoint": "./zephyr/05-04-24-Weni-WeniGPT-Agents-Zephyr-1.0.1-KTO_testing kto dataset during training-3_max_steps-145_batch_16_2024-04-05_ppid_9/checkpoint-100",
4
  "epoch": 0.684931506849315,
5
  "eval_steps": 50,
@@ -10,95 +10,95 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.14,
13
- "grad_norm": 4.650723457336426,
14
- "kl": 4.865365505218506,
15
- "learning_rate": 0.00018285714285714286,
16
- "logps/chosen": -310.4614562988281,
17
- "logps/rejected": -364.2945556640625,
18
- "loss": 0.45,
19
- "rewards/chosen": -2.890953779220581,
20
- "rewards/margins": 3.259232521057129,
21
- "rewards/rejected": -5.586672306060791,
22
  "step": 20
23
  },
24
  {
25
  "epoch": 0.27,
26
- "grad_norm": 0.0,
27
- "kl": 1.256190538406372,
28
  "learning_rate": 0.0001542857142857143,
29
- "logps/chosen": -1522.9342041015625,
30
- "logps/rejected": -1577.1060791015625,
31
- "loss": 0.4463,
32
- "rewards/chosen": -123.79159545898438,
33
- "rewards/margins": 4.876855373382568,
34
- "rewards/rejected": -128.13319396972656,
35
  "step": 40
36
  },
37
  {
38
  "epoch": 0.34,
39
- "eval_kl": 0.0,
40
- "eval_logps/chosen": -1422.4874267578125,
41
- "eval_logps/rejected": -1508.553955078125,
42
- "eval_loss": 0.473336786031723,
43
- "eval_rewards/chosen": -114.58648681640625,
44
- "eval_rewards/margins": 10.318593978881836,
45
- "eval_rewards/rejected": -124.07293701171875,
46
- "eval_runtime": 140.0926,
47
- "eval_samples_per_second": 2.141,
48
- "eval_steps_per_second": 0.535,
49
  "step": 50
50
  },
51
  {
52
  "epoch": 0.41,
53
- "grad_norm": 3.258587184973294e-06,
54
- "kl": 0.0,
55
  "learning_rate": 0.00012571428571428572,
56
- "logps/chosen": -1605.4840087890625,
57
- "logps/rejected": -1731.5294189453125,
58
- "loss": 0.4812,
59
- "rewards/chosen": -133.641357421875,
60
- "rewards/margins": 11.06646728515625,
61
- "rewards/rejected": -144.48353576660156,
62
  "step": 60
63
  },
64
  {
65
  "epoch": 0.55,
66
- "grad_norm": 0.0,
67
- "kl": 0.0,
68
  "learning_rate": 9.714285714285715e-05,
69
- "logps/chosen": -2286.116455078125,
70
- "logps/rejected": -2312.602294921875,
71
- "loss": 0.4688,
72
- "rewards/chosen": -200.4071044921875,
73
- "rewards/margins": 2.4775140285491943,
74
- "rewards/rejected": -200.7866973876953,
75
  "step": 80
76
  },
77
  {
78
  "epoch": 0.68,
79
- "grad_norm": 0.0,
80
- "kl": 0.0,
81
  "learning_rate": 6.857142857142858e-05,
82
- "logps/chosen": -2744.5634765625,
83
- "logps/rejected": -2576.698974609375,
84
- "loss": 0.4719,
85
- "rewards/chosen": -246.3837890625,
86
- "rewards/margins": -13.795512199401855,
87
- "rewards/rejected": -228.7157440185547,
88
  "step": 100
89
  },
90
  {
91
  "epoch": 0.68,
92
- "eval_kl": 0.0,
93
- "eval_logps/chosen": -2469.165771484375,
94
- "eval_logps/rejected": -2225.683349609375,
95
- "eval_loss": 0.47333332896232605,
96
- "eval_rewards/chosen": -219.25433349609375,
97
- "eval_rewards/margins": -26.235870361328125,
98
- "eval_rewards/rejected": -195.78585815429688,
99
- "eval_runtime": 139.9824,
100
- "eval_samples_per_second": 2.143,
101
- "eval_steps_per_second": 0.536,
102
  "step": 100
103
  }
104
  ],
 
1
  {
2
+ "best_metric": 0.3754348158836365,
3
  "best_model_checkpoint": "./zephyr/05-04-24-Weni-WeniGPT-Agents-Zephyr-1.0.1-KTO_testing kto dataset during training-3_max_steps-145_batch_16_2024-04-05_ppid_9/checkpoint-100",
4
  "epoch": 0.684931506849315,
5
  "eval_steps": 50,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.14,
13
+ "grad_norm": 7.029934406280518,
14
+ "kl": 2.2542672157287598,
15
+ "learning_rate": 0.00018,
16
+ "logps/chosen": -356.6376037597656,
17
+ "logps/rejected": -393.9584655761719,
18
+ "loss": 0.4417,
19
+ "rewards/chosen": -3.5771310329437256,
20
+ "rewards/margins": 2.5301473140716553,
21
+ "rewards/rejected": -6.391368865966797,
22
  "step": 20
23
  },
24
  {
25
  "epoch": 0.27,
26
+ "grad_norm": 19.954599380493164,
27
+ "kl": 6.632220268249512,
28
  "learning_rate": 0.0001542857142857143,
29
+ "logps/chosen": -319.3048400878906,
30
+ "logps/rejected": -348.5643615722656,
31
+ "loss": 0.4315,
32
+ "rewards/chosen": -1.7924585342407227,
33
+ "rewards/margins": 1.6441251039505005,
34
+ "rewards/rejected": -3.656928539276123,
35
  "step": 40
36
  },
37
  {
38
  "epoch": 0.34,
39
+ "eval_kl": 32.934268951416016,
40
+ "eval_logps/chosen": -264.7897033691406,
41
+ "eval_logps/rejected": -267.6116027832031,
42
+ "eval_loss": 0.4092177450656891,
43
+ "eval_rewards/chosen": 5.2044806480407715,
44
+ "eval_rewards/margins": 2.720277786254883,
45
+ "eval_rewards/rejected": 2.5654892921447754,
46
+ "eval_runtime": 139.3761,
47
+ "eval_samples_per_second": 2.152,
48
+ "eval_steps_per_second": 0.538,
49
  "step": 50
50
  },
51
  {
52
  "epoch": 0.41,
53
+ "grad_norm": 4.9014763832092285,
54
+ "kl": 22.45343017578125,
55
  "learning_rate": 0.00012571428571428572,
56
+ "logps/chosen": -290.8168029785156,
57
+ "logps/rejected": -297.7256774902344,
58
+ "loss": 0.4154,
59
+ "rewards/chosen": 3.169417381286621,
60
+ "rewards/margins": 1.9798123836517334,
61
+ "rewards/rejected": 1.3037060499191284,
62
  "step": 60
63
  },
64
  {
65
  "epoch": 0.55,
66
+ "grad_norm": 2.188873291015625,
67
+ "kl": 15.003771781921387,
68
  "learning_rate": 9.714285714285715e-05,
69
+ "logps/chosen": -279.8699951171875,
70
+ "logps/rejected": -332.38848876953125,
71
+ "loss": 0.3492,
72
+ "rewards/chosen": 1.9723680019378662,
73
+ "rewards/margins": 4.276235103607178,
74
+ "rewards/rejected": -2.454475164413452,
75
  "step": 80
76
  },
77
  {
78
  "epoch": 0.68,
79
+ "grad_norm": 4.462972640991211,
80
+ "kl": 4.606294631958008,
81
  "learning_rate": 6.857142857142858e-05,
82
+ "logps/chosen": -278.8154602050781,
83
+ "logps/rejected": -356.94451904296875,
84
+ "loss": 0.3678,
85
+ "rewards/chosen": 1.00811767578125,
86
+ "rewards/margins": 5.261899471282959,
87
+ "rewards/rejected": -4.226025581359863,
88
  "step": 100
89
  },
90
  {
91
  "epoch": 0.68,
92
+ "eval_kl": 1.3235526084899902,
93
+ "eval_logps/chosen": -330.0657653808594,
94
+ "eval_logps/rejected": -357.2244873046875,
95
+ "eval_loss": 0.3754348158836365,
96
+ "eval_rewards/chosen": -1.3231240510940552,
97
+ "eval_rewards/margins": 5.1890459060668945,
98
+ "eval_rewards/rejected": -6.395801067352295,
99
+ "eval_runtime": 139.0506,
100
+ "eval_samples_per_second": 2.157,
101
+ "eval_steps_per_second": 0.539,
102
  "step": 100
103
  }
104
  ],
checkpoint-100/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91658cc872c39fb432cb322f2450e9b0c776e35ec752fc4911f1c95a922ea14a
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:918924d56c60c496a10afc622ed24a5105bed8686a07c0f35eeb9917bbadd344
3
  size 5624