Upload prior/prior_config.json with huggingface_hub
Browse files- prior/prior_config.json +2 -1
prior/prior_config.json
CHANGED
@@ -34,7 +34,7 @@
|
|
34 |
},
|
35 |
"data": {
|
36 |
"batch_size": 200,
|
37 |
-
"num_data_points":
|
38 |
"eval_every_seconds": 1800,
|
39 |
"image_url": "s3://s-datasets/laion-aesthetic/embeddings/img_emb",
|
40 |
"meta_url": "s3://s-datasets/laion-aesthetic/embeddings/metadata",
|
@@ -51,6 +51,7 @@
|
|
51 |
"max_grad_norm": 0.5,
|
52 |
"use_ema": true,
|
53 |
"ema_beta": 0.9999,
|
|
|
54 |
"amp": false,
|
55 |
"save_every_seconds": 3600,
|
56 |
"eval_timesteps": [64, 250, 1000]
|
|
|
34 |
},
|
35 |
"data": {
|
36 |
"batch_size": 200,
|
37 |
+
"num_data_points": 50805086,
|
38 |
"eval_every_seconds": 1800,
|
39 |
"image_url": "s3://s-datasets/laion-aesthetic/embeddings/img_emb",
|
40 |
"meta_url": "s3://s-datasets/laion-aesthetic/embeddings/metadata",
|
|
|
51 |
"max_grad_norm": 0.5,
|
52 |
"use_ema": true,
|
53 |
"ema_beta": 0.9999,
|
54 |
+
"ema_update_after_step": 1,
|
55 |
"amp": false,
|
56 |
"save_every_seconds": 3600,
|
57 |
"eval_timesteps": [64, 250, 1000]
|