MaxReynolds
commited on
Commit
•
d2418d4
1
Parent(s):
25403b8
End of training
Browse files- README.md +4 -4
- feature_extractor/preprocessor_config.json +2 -1
- model_index.json +1 -1
- safety_checker/config.json +1 -1
- scheduler/scheduler_config.json +1 -1
- text_encoder/config.json +1 -1
- unet/config.json +1 -1
- unet/diffusion_pytorch_model.safetensors +1 -1
- vae/config.json +1 -1
- val_imgs_grid.png +0 -0
README.md
CHANGED
@@ -14,7 +14,7 @@ inference: true
|
|
14 |
|
15 |
# Text-to-image finetuning - MaxReynolds/MyPatternModel
|
16 |
|
17 |
-
This pipeline was finetuned from **CompVis/stable-diffusion-v1-4** on the **MaxReynolds/MyPatternDataset** dataset. Below are some example images generated with the finetuned pipeline using the following prompts: ['
|
18 |
|
19 |
![val_imgs_grid](./val_imgs_grid.png)
|
20 |
|
@@ -28,7 +28,7 @@ from diffusers import DiffusionPipeline
|
|
28 |
import torch
|
29 |
|
30 |
pipeline = DiffusionPipeline.from_pretrained("MaxReynolds/MyPatternModel", torch_dtype=torch.float16)
|
31 |
-
prompt = "
|
32 |
image = pipeline(prompt).images[0]
|
33 |
image.save("my_image.png")
|
34 |
```
|
@@ -37,7 +37,7 @@ image.save("my_image.png")
|
|
37 |
|
38 |
These are the key hyperparameters used during training:
|
39 |
|
40 |
-
* Epochs:
|
41 |
* Learning rate: 1e-05
|
42 |
* Batch size: 1
|
43 |
* Gradient accumulation steps: 4
|
@@ -45,4 +45,4 @@ These are the key hyperparameters used during training:
|
|
45 |
* Mixed-precision: fp16
|
46 |
|
47 |
|
48 |
-
More information on all the CLI arguments and the environment are available on your [`wandb` run page](https://wandb.ai/max-f-reynolds/text2image-fine-tune/runs/
|
|
|
14 |
|
15 |
# Text-to-image finetuning - MaxReynolds/MyPatternModel
|
16 |
|
17 |
+
This pipeline was finetuned from **CompVis/stable-diffusion-v1-4** on the **MaxReynolds/MyPatternDataset** dataset. Below are some example images generated with the finetuned pipeline using the following prompts: ['<r4nd0m-l4b3l>']:
|
18 |
|
19 |
![val_imgs_grid](./val_imgs_grid.png)
|
20 |
|
|
|
28 |
import torch
|
29 |
|
30 |
pipeline = DiffusionPipeline.from_pretrained("MaxReynolds/MyPatternModel", torch_dtype=torch.float16)
|
31 |
+
prompt = "<r4nd0m-l4b3l>"
|
32 |
image = pipeline(prompt).images[0]
|
33 |
image.save("my_image.png")
|
34 |
```
|
|
|
37 |
|
38 |
These are the key hyperparameters used during training:
|
39 |
|
40 |
+
* Epochs: 22
|
41 |
* Learning rate: 1e-05
|
42 |
* Batch size: 1
|
43 |
* Gradient accumulation steps: 4
|
|
|
45 |
* Mixed-precision: fp16
|
46 |
|
47 |
|
48 |
+
More information on all the CLI arguments and the environment are available on your [`wandb` run page](https://wandb.ai/max-f-reynolds/text2image-fine-tune/runs/kybv4sem).
|
feature_extractor/preprocessor_config.json
CHANGED
@@ -24,5 +24,6 @@
|
|
24 |
"rescale_factor": 0.00392156862745098,
|
25 |
"size": {
|
26 |
"shortest_edge": 224
|
27 |
-
}
|
|
|
28 |
}
|
|
|
24 |
"rescale_factor": 0.00392156862745098,
|
25 |
"size": {
|
26 |
"shortest_edge": 224
|
27 |
+
},
|
28 |
+
"use_square_size": false
|
29 |
}
|
model_index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "StableDiffusionPipeline",
|
3 |
-
"_diffusers_version": "0.
|
4 |
"_name_or_path": "CompVis/stable-diffusion-v1-4",
|
5 |
"feature_extractor": [
|
6 |
"transformers",
|
|
|
1 |
{
|
2 |
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.23.0.dev0",
|
4 |
"_name_or_path": "CompVis/stable-diffusion-v1-4",
|
5 |
"feature_extractor": [
|
6 |
"transformers",
|
safety_checker/config.json
CHANGED
@@ -15,7 +15,7 @@
|
|
15 |
"num_attention_heads": 12
|
16 |
},
|
17 |
"torch_dtype": "float32",
|
18 |
-
"transformers_version": "4.
|
19 |
"vision_config": {
|
20 |
"dropout": 0.0,
|
21 |
"hidden_size": 1024,
|
|
|
15 |
"num_attention_heads": 12
|
16 |
},
|
17 |
"torch_dtype": "float32",
|
18 |
+
"transformers_version": "4.35.0",
|
19 |
"vision_config": {
|
20 |
"dropout": 0.0,
|
21 |
"hidden_size": 1024,
|
scheduler/scheduler_config.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "PNDMScheduler",
|
3 |
-
"_diffusers_version": "0.
|
4 |
"beta_end": 0.012,
|
5 |
"beta_schedule": "scaled_linear",
|
6 |
"beta_start": 0.00085,
|
|
|
1 |
{
|
2 |
"_class_name": "PNDMScheduler",
|
3 |
+
"_diffusers_version": "0.23.0.dev0",
|
4 |
"beta_end": 0.012,
|
5 |
"beta_schedule": "scaled_linear",
|
6 |
"beta_start": 0.00085,
|
text_encoder/config.json
CHANGED
@@ -20,6 +20,6 @@
|
|
20 |
"pad_token_id": 1,
|
21 |
"projection_dim": 512,
|
22 |
"torch_dtype": "float16",
|
23 |
-
"transformers_version": "4.
|
24 |
"vocab_size": 49408
|
25 |
}
|
|
|
20 |
"pad_token_id": 1,
|
21 |
"projection_dim": 512,
|
22 |
"torch_dtype": "float16",
|
23 |
+
"transformers_version": "4.35.0",
|
24 |
"vocab_size": 49408
|
25 |
}
|
unet/config.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
-
"_diffusers_version": "0.
|
4 |
"_name_or_path": "CompVis/stable-diffusion-v1-4",
|
5 |
"act_fn": "silu",
|
6 |
"addition_embed_type": null,
|
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
+
"_diffusers_version": "0.23.0.dev0",
|
4 |
"_name_or_path": "CompVis/stable-diffusion-v1-4",
|
5 |
"act_fn": "silu",
|
6 |
"addition_embed_type": null,
|
unet/diffusion_pytorch_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3438167536
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc3d75a7ba6037cd8999e81d7be6b02cc0955b0221f2e7189a29cb553b1d7439
|
3 |
size 3438167536
|
vae/config.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
-
"_diffusers_version": "0.
|
4 |
"_name_or_path": "CompVis/stable-diffusion-v1-4",
|
5 |
"act_fn": "silu",
|
6 |
"block_out_channels": [
|
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
+
"_diffusers_version": "0.23.0.dev0",
|
4 |
"_name_or_path": "CompVis/stable-diffusion-v1-4",
|
5 |
"act_fn": "silu",
|
6 |
"block_out_channels": [
|
val_imgs_grid.png
CHANGED