adamjweintraut commited on
Commit
59ad45e
1 Parent(s): c8869fc

Training in progress, step 500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf8d3488e8a40d95324b5d59d9e28f847bb4ad4337cbab29278f2d71848c8154
3
  size 1625545896
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f04b87a850c8e1978cbcc919472d319a69545cea7249b6850d0c66548817b58c
3
  size 1625545896
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cba8df3e79cea54a0cf191a9d20a258a846fea397be360663639bf8078c188e1
3
  size 3250997519
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:111fc2c7739ba419a0649fff6c653fe1ac707bedea828f93414bbd7128ac5a24
3
  size 3250997519
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:292e3c34f31addb8e5d45c64c6c5a9fa41fab6787ec343f6cc668c14137bedb2
3
  size 14308
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b85836ff314397b6e9cda5d18b4e61802951efe26d6873089b5a8ec8a7fb5ae8
3
  size 14308
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d1b75caef8dcf16fc2c3f168095032c1c816c3b909d9e4856126e54ea281ebd
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbabefdd804dedc5c232b0c700653c9873b9270f3c9973815bb3914f4fa8dadb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,95 +1,35 @@
1
  {
2
- "best_metric": 0.45656052231788635,
3
- "best_model_checkpoint": "/content/drive/MyDrive/W210 Capstone - Lyric Generation with Melody/loaf/models/kwsylgen/bart/bart-finetuned-kwsylgen-64/checkpoint-2000",
4
- "epoch": 1.7385257301808066,
5
  "eval_steps": 500,
6
- "global_step": 2500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 0.35,
13
- "grad_norm": 0.5149683952331543,
14
- "learning_rate": 4.5679763560500696e-05,
15
- "loss": 0.4749,
16
  "step": 500
17
  },
18
  {
19
- "epoch": 0.35,
20
- "eval_loss": 0.4616529643535614,
21
- "eval_runtime": 93.1578,
22
- "eval_samples_per_second": 136.371,
23
- "eval_steps_per_second": 2.136,
24
  "step": 500
25
- },
26
- {
27
- "epoch": 0.7,
28
- "grad_norm": 0.5068113207817078,
29
- "learning_rate": 4.133344923504868e-05,
30
- "loss": 0.3922,
31
- "step": 1000
32
- },
33
- {
34
- "epoch": 0.7,
35
- "eval_loss": 0.4581904709339142,
36
- "eval_runtime": 93.9811,
37
- "eval_samples_per_second": 135.176,
38
- "eval_steps_per_second": 2.117,
39
- "step": 1000
40
- },
41
- {
42
- "epoch": 1.04,
43
- "grad_norm": 0.6911681890487671,
44
- "learning_rate": 3.698713490959667e-05,
45
- "loss": 0.3655,
46
- "step": 1500
47
- },
48
- {
49
- "epoch": 1.04,
50
- "eval_loss": 0.4568086564540863,
51
- "eval_runtime": 93.8201,
52
- "eval_samples_per_second": 135.408,
53
- "eval_steps_per_second": 2.121,
54
- "step": 1500
55
- },
56
- {
57
- "epoch": 1.39,
58
- "grad_norm": 0.5658309459686279,
59
- "learning_rate": 3.2640820584144646e-05,
60
- "loss": 0.3455,
61
- "step": 2000
62
- },
63
- {
64
- "epoch": 1.39,
65
- "eval_loss": 0.45656052231788635,
66
- "eval_runtime": 94.4047,
67
- "eval_samples_per_second": 134.57,
68
- "eval_steps_per_second": 2.108,
69
- "step": 2000
70
- },
71
- {
72
- "epoch": 1.74,
73
- "grad_norm": 0.4761568605899811,
74
- "learning_rate": 2.8294506258692632e-05,
75
- "loss": 0.3291,
76
- "step": 2500
77
- },
78
- {
79
- "epoch": 1.74,
80
- "eval_loss": 0.4570353031158447,
81
- "eval_runtime": 95.9797,
82
- "eval_samples_per_second": 132.361,
83
- "eval_steps_per_second": 2.073,
84
- "step": 2500
85
  }
86
  ],
87
  "logging_steps": 500,
88
- "max_steps": 5752,
89
  "num_input_tokens_seen": 0,
90
- "num_train_epochs": 4,
91
  "save_steps": 500,
92
- "total_flos": 2.166887891651789e+16,
93
  "train_batch_size": 64,
94
  "trial_name": null,
95
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.45187193155288696,
3
+ "best_model_checkpoint": "/content/drive/MyDrive/W210 Capstone - Lyric Generation with Melody/loaf/models/kwsylgen/bart/bart-finetuned-kwsylgen-64/checkpoint-500",
4
+ "epoch": 0.17940437746681018,
5
  "eval_steps": 500,
6
+ "global_step": 500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 0.18,
13
+ "grad_norm": 0.534104585647583,
14
+ "learning_rate": 4.910836024398995e-05,
15
+ "loss": 0.486,
16
  "step": 500
17
  },
18
  {
19
+ "epoch": 0.18,
20
+ "eval_loss": 0.45187193155288696,
21
+ "eval_runtime": 211.1424,
22
+ "eval_samples_per_second": 112.166,
23
+ "eval_steps_per_second": 1.757,
24
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  }
26
  ],
27
  "logging_steps": 500,
28
+ "max_steps": 27870,
29
  "num_input_tokens_seen": 0,
30
+ "num_train_epochs": 10,
31
  "save_steps": 500,
32
+ "total_flos": 4334209204224000.0,
33
  "train_batch_size": 64,
34
  "trial_name": null,
35
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17e6869522a097ec64314dcf3fed02e0d24916c5cedb8b01b2170c098b079f8a
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e26ab58717131c0abe643633f423bce131a8848cdbb71c5c35dc1974207a9c7
3
  size 5368