adding 3D style
#3
by
Guizmus
- opened
- README.md +64 -22
- unet/diffusion_pytorch_model.bin → ckpt/3DStyle-v1.ckpt +2 -2
- safety_checker/pytorch_model.bin → ckpt/3DStyle-v1_with_optimizers.ckpt +2 -2
- AnimeChanStyle_v1.ckpt → ckpt/AnimeChanStyle-v1.ckpt +0 -0
- AnimeChanStyle-v2.ckpt → ckpt/AnimeChanStyle-v2.ckpt +0 -0
- text_encoder/pytorch_model.bin → datasets/3DChanStyle-v1.zip +2 -2
- AnimeChan Style.zip → datasets/AnimeChanStyle-v1.zip +0 -0
- vae/diffusion_pytorch_model.bin → datasets/AnimeChanStyle-v2.zip +2 -2
- feature_extractor/preprocessor_config.json +0 -20
- images/showcase_3DChanStyle-v1.jpg +0 -0
- showcase.jpg → images/showcase_AnimeChan-v1.jpg +0 -0
- showcase_AnimeChan.jpg → images/showcase_AnimeChan-v2.jpg +0 -0
- images/showcase_main.jpg +0 -0
- model_index.json +0 -32
- safety_checker/config.json +0 -179
- scheduler/scheduler_config.json +0 -12
- text_encoder/config.json +0 -25
- tokenizer/merges.txt +0 -0
- tokenizer/special_tokens_map.json +0 -24
- tokenizer/tokenizer_config.json +0 -34
- tokenizer/vocab.json +0 -0
- unet/config.json +0 -36
- vae/config.json +0 -29
README.md
CHANGED
@@ -2,46 +2,88 @@
|
|
2 |
language:
|
3 |
- en
|
4 |
license: creativeml-openrail-m
|
5 |
-
thumbnail: "https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/
|
6 |
tags:
|
7 |
- stable-diffusion
|
8 |
- text-to-image
|
9 |
- image-to-image
|
10 |
library_name: "EveryDream"
|
11 |
-
datasets:
|
12 |
-
- Guizmus/AnimeChanStyle
|
13 |
inference: false
|
14 |
---
|
15 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
# AnimeChan Style
|
17 |
-
<p>
|
18 |
-
<img alt="Showcase" src="https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/showcase_AnimeChan.jpg"/><br/>
|
19 |
-
This model was based on <a href="https://huggingface.co/naclbit/trinart_stable_diffusion_v2">Trinart</a> model.<br/>
|
20 |
-
The dataset was a collaborative effort of the Stable Diffusion #anime channel, made of pictures from the users themselves using their different techniques.<br/>
|
21 |
-
100 total pictures in the dataset, 300 repeats total each, over 6 Epoch on LR1e-6.<br/>
|
22 |
-
This was trained using EveryDream with a full caption of all training pictures. The dataset can be found <a href="https://huggingface.co/datasets/Guizmus/AnimeChanStyle">here</a>.<br/>
|
23 |
-
<br/>
|
24 |
-
The style will be called by the use of the token <b>AnimeChan Style</b>.<br/>
|
25 |
-
<br/>
|
26 |
-
To access this model, you can download the CKPT file below, or use the <a href="https://huggingface.co/Guizmus/AnimeChanStyle/tree/main">diffusers</a>
|
27 |
-
</p>
|
28 |
|
29 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
|
31 |
-
[
|
32 |
|
33 |
-
|
34 |
|
35 |
-
|
36 |
|
37 |
-
[first version download link](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/AnimeChanStyle_v1.ckpt)
|
38 |
|
39 |
-
[dataset for the first version](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/AnimeChan%20Style.zip)
|
40 |
|
41 |
|
42 |
-
|
43 |
|
44 |
-
|
45 |
The CreativeML OpenRAIL License specifies:
|
46 |
|
47 |
1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
|
|
|
2 |
language:
|
3 |
- en
|
4 |
license: creativeml-openrail-m
|
5 |
+
thumbnail: "https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/images/showcase_main.jpg"
|
6 |
tags:
|
7 |
- stable-diffusion
|
8 |
- text-to-image
|
9 |
- image-to-image
|
10 |
library_name: "EveryDream"
|
|
|
|
|
11 |
inference: false
|
12 |
---
|
13 |
|
14 |
+
# Introduction
|
15 |
+
|
16 |
+
This is a collection of models made from and for the users of the Stable Diffusion Discord server. Different categories of channel exist, the "Dreamers Communities" presenting a panel of subjects, like Anime, 3D, or Architectural. Each of these channels has users posting images made through the use of Stable diffusion. After asking the users, and, depending on the activity of each channel, collecting a dataset from new submissions or from the history of the channel, I intend to build multiple models representing the style of each, so that users can produce things in the style they like and mix it with other things more easily.
|
17 |
+
|
18 |
+
Those are mainly done through the use of EveryDream, and should result in a Mega Model towards the end for the datasets that are compatible. Some model like the Anime one require to stay on a different starting point, and may not get merged.
|
19 |
+
|
20 |
+
|
21 |
+
|
22 |
+
# 3DChanStyle Style
|
23 |
+
|
24 |
+
|
25 |
+
## Dataset & training
|
26 |
+
|
27 |
+
This model was based on [RunwayML SD 1.5](https://huggingface.co/runwayml/stable-diffusion-v1-5) model with updated VAE.
|
28 |
+
|
29 |
+
The dataset was a collaborative effort of the Stable Diffusion #3D channel, made of pictures from the users themselves using their different techniques.
|
30 |
+
|
31 |
+
120 total pictures in the dataset, 500 repeats total each, over 10 Epoch on LR1e-6.
|
32 |
+
|
33 |
+
This was trained using EveryDream with a full caption of all training pictures.
|
34 |
+
|
35 |
+
The style will be called by the use of the token **3D Style**.
|
36 |
+
|
37 |
+
Other significant tokens : rick roll, fullbody shot, bad cosplay man
|
38 |
+
|
39 |
+
## Showcase & Downloads
|
40 |
+
|
41 |
+
![Showcase](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/images/showcase_3DChanStyle-v1.jpg)
|
42 |
+
|
43 |
+
[CKPT (2GB)](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/ckpt/3DStyle-v1.ckpt)
|
44 |
+
|
45 |
+
[CKPT with training optimizers (11GB)](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/ckpt/3DStyle-v1_with_optimizers.ckpt)
|
46 |
+
|
47 |
+
[Dataset](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/datasets/3DChanStyle-v1.zip)
|
48 |
+
|
49 |
+
|
50 |
+
|
51 |
# AnimeChan Style
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
52 |
|
53 |
+
## Dataset & training
|
54 |
+
|
55 |
+
This model was based on [Trinart](https://huggingface.co/naclbit/trinart_stable_diffusion_v2) model.
|
56 |
+
|
57 |
+
The dataset was a collaborative effort of the Stable Diffusion #anime channel, made of pictures from the users themselves using their different techniques.
|
58 |
+
|
59 |
+
100 total pictures in the dataset, 300 repeats total each, over 6 Epoch on LR1e-6.
|
60 |
+
|
61 |
+
This was trained using EveryDream with a full caption of all training pictures.
|
62 |
+
|
63 |
+
The style will be called by the use of the token **AnimeChan Style**.
|
64 |
+
|
65 |
+
## Downloads v2
|
66 |
+
|
67 |
+
![Showcase](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/images/showcase_AnimeChan-v2.jpg)
|
68 |
+
|
69 |
+
[CKPT (2GB)](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/ckpt/AnimeChanStyle-v2.ckpt)
|
70 |
+
|
71 |
+
[Dataset](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/datasets/AnimeChanStyle-v2.zip)
|
72 |
+
|
73 |
+
## Downloads v1
|
74 |
|
75 |
+
![Showcase](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/images/showcase_AnimeChan-v1.jpg)
|
76 |
|
77 |
+
[CKPT (2GB)](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/ckpt/AnimeChanStyle-v1.ckpt)
|
78 |
|
79 |
+
[Dataset](https://huggingface.co/Guizmus/SD_DreamerCommunities_Collection/resolve/main/datasets/AnimeChanStyle-v1.zip)
|
80 |
|
|
|
81 |
|
|
|
82 |
|
83 |
|
84 |
+
# License
|
85 |
|
86 |
+
These models are open access and available to all, with a CreativeML OpenRAIL-M license further specifying rights and usage.
|
87 |
The CreativeML OpenRAIL License specifies:
|
88 |
|
89 |
1. You can't use the model to deliberately produce nor share illegal or harmful outputs or content
|
unet/diffusion_pytorch_model.bin → ckpt/3DStyle-v1.ckpt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9698c46f614733b95d75cae29e777be15b2f13758f92d8eb113d566d79b665a6
|
3 |
+
size 2132888989
|
safety_checker/pytorch_model.bin → ckpt/3DStyle-v1_with_optimizers.ckpt
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8bca5d3ef403ad1ecfa7d3c9c5717a3bd5fa7899f93cd5526e4d55bbbd380147
|
3 |
+
size 12126930715
|
AnimeChanStyle_v1.ckpt → ckpt/AnimeChanStyle-v1.ckpt
RENAMED
File without changes
|
AnimeChanStyle-v2.ckpt → ckpt/AnimeChanStyle-v2.ckpt
RENAMED
File without changes
|
text_encoder/pytorch_model.bin → datasets/3DChanStyle-v1.zip
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0b659800b6b5140f8e953be05d7acd6d02296dbcbbd43dae944af84b2c5d131a
|
3 |
+
size 29246736
|
AnimeChan Style.zip → datasets/AnimeChanStyle-v1.zip
RENAMED
File without changes
|
vae/diffusion_pytorch_model.bin → datasets/AnimeChanStyle-v2.zip
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:09f5678e615988ce35d01ce9980a546568e27ec52065d2578c265275b070ea27
|
3 |
+
size 98956573
|
feature_extractor/preprocessor_config.json
DELETED
@@ -1,20 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"crop_size": 224,
|
3 |
-
"do_center_crop": true,
|
4 |
-
"do_convert_rgb": true,
|
5 |
-
"do_normalize": true,
|
6 |
-
"do_resize": true,
|
7 |
-
"feature_extractor_type": "CLIPFeatureExtractor",
|
8 |
-
"image_mean": [
|
9 |
-
0.48145466,
|
10 |
-
0.4578275,
|
11 |
-
0.40821073
|
12 |
-
],
|
13 |
-
"image_std": [
|
14 |
-
0.26862954,
|
15 |
-
0.26130258,
|
16 |
-
0.27577711
|
17 |
-
],
|
18 |
-
"resample": 3,
|
19 |
-
"size": 224
|
20 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
images/showcase_3DChanStyle-v1.jpg
ADDED
showcase.jpg → images/showcase_AnimeChan-v1.jpg
RENAMED
File without changes
|
showcase_AnimeChan.jpg → images/showcase_AnimeChan-v2.jpg
RENAMED
File without changes
|
images/showcase_main.jpg
ADDED
model_index.json
DELETED
@@ -1,32 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_class_name": "StableDiffusionPipeline",
|
3 |
-
"_diffusers_version": "0.8.0.dev0",
|
4 |
-
"feature_extractor": [
|
5 |
-
"transformers",
|
6 |
-
"CLIPFeatureExtractor"
|
7 |
-
],
|
8 |
-
"safety_checker": [
|
9 |
-
"stable_diffusion",
|
10 |
-
"StableDiffusionSafetyChecker"
|
11 |
-
],
|
12 |
-
"scheduler": [
|
13 |
-
"diffusers",
|
14 |
-
"PNDMScheduler"
|
15 |
-
],
|
16 |
-
"text_encoder": [
|
17 |
-
"transformers",
|
18 |
-
"CLIPTextModel"
|
19 |
-
],
|
20 |
-
"tokenizer": [
|
21 |
-
"transformers",
|
22 |
-
"CLIPTokenizer"
|
23 |
-
],
|
24 |
-
"unet": [
|
25 |
-
"diffusers",
|
26 |
-
"UNet2DConditionModel"
|
27 |
-
],
|
28 |
-
"vae": [
|
29 |
-
"diffusers",
|
30 |
-
"AutoencoderKL"
|
31 |
-
]
|
32 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
safety_checker/config.json
DELETED
@@ -1,179 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_commit_hash": "4bb648a606ef040e7685bde262611766a5fdd67b",
|
3 |
-
"_name_or_path": "CompVis/stable-diffusion-safety-checker",
|
4 |
-
"architectures": [
|
5 |
-
"StableDiffusionSafetyChecker"
|
6 |
-
],
|
7 |
-
"initializer_factor": 1.0,
|
8 |
-
"logit_scale_init_value": 2.6592,
|
9 |
-
"model_type": "clip",
|
10 |
-
"projection_dim": 768,
|
11 |
-
"text_config": {
|
12 |
-
"_name_or_path": "",
|
13 |
-
"add_cross_attention": false,
|
14 |
-
"architectures": null,
|
15 |
-
"attention_dropout": 0.0,
|
16 |
-
"bad_words_ids": null,
|
17 |
-
"begin_suppress_tokens": null,
|
18 |
-
"bos_token_id": 0,
|
19 |
-
"chunk_size_feed_forward": 0,
|
20 |
-
"cross_attention_hidden_size": null,
|
21 |
-
"decoder_start_token_id": null,
|
22 |
-
"diversity_penalty": 0.0,
|
23 |
-
"do_sample": false,
|
24 |
-
"dropout": 0.0,
|
25 |
-
"early_stopping": false,
|
26 |
-
"encoder_no_repeat_ngram_size": 0,
|
27 |
-
"eos_token_id": 2,
|
28 |
-
"exponential_decay_length_penalty": null,
|
29 |
-
"finetuning_task": null,
|
30 |
-
"forced_bos_token_id": null,
|
31 |
-
"forced_eos_token_id": null,
|
32 |
-
"hidden_act": "quick_gelu",
|
33 |
-
"hidden_size": 768,
|
34 |
-
"id2label": {
|
35 |
-
"0": "LABEL_0",
|
36 |
-
"1": "LABEL_1"
|
37 |
-
},
|
38 |
-
"initializer_factor": 1.0,
|
39 |
-
"initializer_range": 0.02,
|
40 |
-
"intermediate_size": 3072,
|
41 |
-
"is_decoder": false,
|
42 |
-
"is_encoder_decoder": false,
|
43 |
-
"label2id": {
|
44 |
-
"LABEL_0": 0,
|
45 |
-
"LABEL_1": 1
|
46 |
-
},
|
47 |
-
"layer_norm_eps": 1e-05,
|
48 |
-
"length_penalty": 1.0,
|
49 |
-
"max_length": 20,
|
50 |
-
"max_position_embeddings": 77,
|
51 |
-
"min_length": 0,
|
52 |
-
"model_type": "clip_text_model",
|
53 |
-
"no_repeat_ngram_size": 0,
|
54 |
-
"num_attention_heads": 12,
|
55 |
-
"num_beam_groups": 1,
|
56 |
-
"num_beams": 1,
|
57 |
-
"num_hidden_layers": 12,
|
58 |
-
"num_return_sequences": 1,
|
59 |
-
"output_attentions": false,
|
60 |
-
"output_hidden_states": false,
|
61 |
-
"output_scores": false,
|
62 |
-
"pad_token_id": 1,
|
63 |
-
"prefix": null,
|
64 |
-
"problem_type": null,
|
65 |
-
"pruned_heads": {},
|
66 |
-
"remove_invalid_values": false,
|
67 |
-
"repetition_penalty": 1.0,
|
68 |
-
"return_dict": true,
|
69 |
-
"return_dict_in_generate": false,
|
70 |
-
"sep_token_id": null,
|
71 |
-
"suppress_tokens": null,
|
72 |
-
"task_specific_params": null,
|
73 |
-
"temperature": 1.0,
|
74 |
-
"tf_legacy_loss": false,
|
75 |
-
"tie_encoder_decoder": false,
|
76 |
-
"tie_word_embeddings": true,
|
77 |
-
"tokenizer_class": null,
|
78 |
-
"top_k": 50,
|
79 |
-
"top_p": 1.0,
|
80 |
-
"torch_dtype": null,
|
81 |
-
"torchscript": false,
|
82 |
-
"transformers_version": "4.24.0",
|
83 |
-
"typical_p": 1.0,
|
84 |
-
"use_bfloat16": false,
|
85 |
-
"vocab_size": 49408
|
86 |
-
},
|
87 |
-
"text_config_dict": {
|
88 |
-
"hidden_size": 768,
|
89 |
-
"intermediate_size": 3072,
|
90 |
-
"num_attention_heads": 12,
|
91 |
-
"num_hidden_layers": 12
|
92 |
-
},
|
93 |
-
"torch_dtype": "float32",
|
94 |
-
"transformers_version": null,
|
95 |
-
"vision_config": {
|
96 |
-
"_name_or_path": "",
|
97 |
-
"add_cross_attention": false,
|
98 |
-
"architectures": null,
|
99 |
-
"attention_dropout": 0.0,
|
100 |
-
"bad_words_ids": null,
|
101 |
-
"begin_suppress_tokens": null,
|
102 |
-
"bos_token_id": null,
|
103 |
-
"chunk_size_feed_forward": 0,
|
104 |
-
"cross_attention_hidden_size": null,
|
105 |
-
"decoder_start_token_id": null,
|
106 |
-
"diversity_penalty": 0.0,
|
107 |
-
"do_sample": false,
|
108 |
-
"dropout": 0.0,
|
109 |
-
"early_stopping": false,
|
110 |
-
"encoder_no_repeat_ngram_size": 0,
|
111 |
-
"eos_token_id": null,
|
112 |
-
"exponential_decay_length_penalty": null,
|
113 |
-
"finetuning_task": null,
|
114 |
-
"forced_bos_token_id": null,
|
115 |
-
"forced_eos_token_id": null,
|
116 |
-
"hidden_act": "quick_gelu",
|
117 |
-
"hidden_size": 1024,
|
118 |
-
"id2label": {
|
119 |
-
"0": "LABEL_0",
|
120 |
-
"1": "LABEL_1"
|
121 |
-
},
|
122 |
-
"image_size": 224,
|
123 |
-
"initializer_factor": 1.0,
|
124 |
-
"initializer_range": 0.02,
|
125 |
-
"intermediate_size": 4096,
|
126 |
-
"is_decoder": false,
|
127 |
-
"is_encoder_decoder": false,
|
128 |
-
"label2id": {
|
129 |
-
"LABEL_0": 0,
|
130 |
-
"LABEL_1": 1
|
131 |
-
},
|
132 |
-
"layer_norm_eps": 1e-05,
|
133 |
-
"length_penalty": 1.0,
|
134 |
-
"max_length": 20,
|
135 |
-
"min_length": 0,
|
136 |
-
"model_type": "clip_vision_model",
|
137 |
-
"no_repeat_ngram_size": 0,
|
138 |
-
"num_attention_heads": 16,
|
139 |
-
"num_beam_groups": 1,
|
140 |
-
"num_beams": 1,
|
141 |
-
"num_channels": 3,
|
142 |
-
"num_hidden_layers": 24,
|
143 |
-
"num_return_sequences": 1,
|
144 |
-
"output_attentions": false,
|
145 |
-
"output_hidden_states": false,
|
146 |
-
"output_scores": false,
|
147 |
-
"pad_token_id": null,
|
148 |
-
"patch_size": 14,
|
149 |
-
"prefix": null,
|
150 |
-
"problem_type": null,
|
151 |
-
"pruned_heads": {},
|
152 |
-
"remove_invalid_values": false,
|
153 |
-
"repetition_penalty": 1.0,
|
154 |
-
"return_dict": true,
|
155 |
-
"return_dict_in_generate": false,
|
156 |
-
"sep_token_id": null,
|
157 |
-
"suppress_tokens": null,
|
158 |
-
"task_specific_params": null,
|
159 |
-
"temperature": 1.0,
|
160 |
-
"tf_legacy_loss": false,
|
161 |
-
"tie_encoder_decoder": false,
|
162 |
-
"tie_word_embeddings": true,
|
163 |
-
"tokenizer_class": null,
|
164 |
-
"top_k": 50,
|
165 |
-
"top_p": 1.0,
|
166 |
-
"torch_dtype": null,
|
167 |
-
"torchscript": false,
|
168 |
-
"transformers_version": "4.24.0",
|
169 |
-
"typical_p": 1.0,
|
170 |
-
"use_bfloat16": false
|
171 |
-
},
|
172 |
-
"vision_config_dict": {
|
173 |
-
"hidden_size": 1024,
|
174 |
-
"intermediate_size": 4096,
|
175 |
-
"num_attention_heads": 16,
|
176 |
-
"num_hidden_layers": 24,
|
177 |
-
"patch_size": 14
|
178 |
-
}
|
179 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
scheduler/scheduler_config.json
DELETED
@@ -1,12 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_class_name": "PNDMScheduler",
|
3 |
-
"_diffusers_version": "0.8.0.dev0",
|
4 |
-
"beta_end": 0.012,
|
5 |
-
"beta_schedule": "scaled_linear",
|
6 |
-
"beta_start": 0.00085,
|
7 |
-
"num_train_timesteps": 1000,
|
8 |
-
"set_alpha_to_one": false,
|
9 |
-
"skip_prk_steps": true,
|
10 |
-
"steps_offset": 1,
|
11 |
-
"trained_betas": null
|
12 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
text_encoder/config.json
DELETED
@@ -1,25 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_name_or_path": "openai/clip-vit-large-patch14",
|
3 |
-
"architectures": [
|
4 |
-
"CLIPTextModel"
|
5 |
-
],
|
6 |
-
"attention_dropout": 0.0,
|
7 |
-
"bos_token_id": 0,
|
8 |
-
"dropout": 0.0,
|
9 |
-
"eos_token_id": 2,
|
10 |
-
"hidden_act": "quick_gelu",
|
11 |
-
"hidden_size": 768,
|
12 |
-
"initializer_factor": 1.0,
|
13 |
-
"initializer_range": 0.02,
|
14 |
-
"intermediate_size": 3072,
|
15 |
-
"layer_norm_eps": 1e-05,
|
16 |
-
"max_position_embeddings": 77,
|
17 |
-
"model_type": "clip_text_model",
|
18 |
-
"num_attention_heads": 12,
|
19 |
-
"num_hidden_layers": 12,
|
20 |
-
"pad_token_id": 1,
|
21 |
-
"projection_dim": 768,
|
22 |
-
"torch_dtype": "float32",
|
23 |
-
"transformers_version": "4.24.0",
|
24 |
-
"vocab_size": 49408
|
25 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
tokenizer/merges.txt
DELETED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer/special_tokens_map.json
DELETED
@@ -1,24 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"bos_token": {
|
3 |
-
"content": "<|startoftext|>",
|
4 |
-
"lstrip": false,
|
5 |
-
"normalized": true,
|
6 |
-
"rstrip": false,
|
7 |
-
"single_word": false
|
8 |
-
},
|
9 |
-
"eos_token": {
|
10 |
-
"content": "<|endoftext|>",
|
11 |
-
"lstrip": false,
|
12 |
-
"normalized": true,
|
13 |
-
"rstrip": false,
|
14 |
-
"single_word": false
|
15 |
-
},
|
16 |
-
"pad_token": "<|endoftext|>",
|
17 |
-
"unk_token": {
|
18 |
-
"content": "<|endoftext|>",
|
19 |
-
"lstrip": false,
|
20 |
-
"normalized": true,
|
21 |
-
"rstrip": false,
|
22 |
-
"single_word": false
|
23 |
-
}
|
24 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
tokenizer/tokenizer_config.json
DELETED
@@ -1,34 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"add_prefix_space": false,
|
3 |
-
"bos_token": {
|
4 |
-
"__type": "AddedToken",
|
5 |
-
"content": "<|startoftext|>",
|
6 |
-
"lstrip": false,
|
7 |
-
"normalized": true,
|
8 |
-
"rstrip": false,
|
9 |
-
"single_word": false
|
10 |
-
},
|
11 |
-
"do_lower_case": true,
|
12 |
-
"eos_token": {
|
13 |
-
"__type": "AddedToken",
|
14 |
-
"content": "<|endoftext|>",
|
15 |
-
"lstrip": false,
|
16 |
-
"normalized": true,
|
17 |
-
"rstrip": false,
|
18 |
-
"single_word": false
|
19 |
-
},
|
20 |
-
"errors": "replace",
|
21 |
-
"model_max_length": 77,
|
22 |
-
"name_or_path": "openai/clip-vit-large-patch14",
|
23 |
-
"pad_token": "<|endoftext|>",
|
24 |
-
"special_tokens_map_file": "./special_tokens_map.json",
|
25 |
-
"tokenizer_class": "CLIPTokenizer",
|
26 |
-
"unk_token": {
|
27 |
-
"__type": "AddedToken",
|
28 |
-
"content": "<|endoftext|>",
|
29 |
-
"lstrip": false,
|
30 |
-
"normalized": true,
|
31 |
-
"rstrip": false,
|
32 |
-
"single_word": false
|
33 |
-
}
|
34 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
tokenizer/vocab.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
unet/config.json
DELETED
@@ -1,36 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_class_name": "UNet2DConditionModel",
|
3 |
-
"_diffusers_version": "0.8.0.dev0",
|
4 |
-
"act_fn": "silu",
|
5 |
-
"attention_head_dim": 8,
|
6 |
-
"block_out_channels": [
|
7 |
-
320,
|
8 |
-
640,
|
9 |
-
1280,
|
10 |
-
1280
|
11 |
-
],
|
12 |
-
"center_input_sample": false,
|
13 |
-
"cross_attention_dim": 768,
|
14 |
-
"down_block_types": [
|
15 |
-
"CrossAttnDownBlock2D",
|
16 |
-
"CrossAttnDownBlock2D",
|
17 |
-
"CrossAttnDownBlock2D",
|
18 |
-
"DownBlock2D"
|
19 |
-
],
|
20 |
-
"downsample_padding": 1,
|
21 |
-
"flip_sin_to_cos": true,
|
22 |
-
"freq_shift": 0,
|
23 |
-
"in_channels": 4,
|
24 |
-
"layers_per_block": 2,
|
25 |
-
"mid_block_scale_factor": 1,
|
26 |
-
"norm_eps": 1e-05,
|
27 |
-
"norm_num_groups": 32,
|
28 |
-
"out_channels": 4,
|
29 |
-
"sample_size": 32,
|
30 |
-
"up_block_types": [
|
31 |
-
"UpBlock2D",
|
32 |
-
"CrossAttnUpBlock2D",
|
33 |
-
"CrossAttnUpBlock2D",
|
34 |
-
"CrossAttnUpBlock2D"
|
35 |
-
]
|
36 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
vae/config.json
DELETED
@@ -1,29 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"_class_name": "AutoencoderKL",
|
3 |
-
"_diffusers_version": "0.8.0.dev0",
|
4 |
-
"act_fn": "silu",
|
5 |
-
"block_out_channels": [
|
6 |
-
128,
|
7 |
-
256,
|
8 |
-
512,
|
9 |
-
512
|
10 |
-
],
|
11 |
-
"down_block_types": [
|
12 |
-
"DownEncoderBlock2D",
|
13 |
-
"DownEncoderBlock2D",
|
14 |
-
"DownEncoderBlock2D",
|
15 |
-
"DownEncoderBlock2D"
|
16 |
-
],
|
17 |
-
"in_channels": 3,
|
18 |
-
"latent_channels": 4,
|
19 |
-
"layers_per_block": 2,
|
20 |
-
"norm_num_groups": 32,
|
21 |
-
"out_channels": 3,
|
22 |
-
"sample_size": 256,
|
23 |
-
"up_block_types": [
|
24 |
-
"UpDecoderBlock2D",
|
25 |
-
"UpDecoderBlock2D",
|
26 |
-
"UpDecoderBlock2D",
|
27 |
-
"UpDecoderBlock2D"
|
28 |
-
]
|
29 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|