Upload 15 files
Browse files- tokenizer/tokenizer_config.json +0 -1
- unet/config.json +1 -1
- unet/diffusion_pytorch_model.bin +1 -1
tokenizer/tokenizer_config.json
CHANGED
@@ -21,7 +21,6 @@
|
|
21 |
"errors": "replace",
|
22 |
"model_max_length": 77,
|
23 |
"pad_token": "<|endoftext|>",
|
24 |
-
"special_tokens_map_file": "./special_tokens_map.json",
|
25 |
"tokenizer_class": "CLIPTokenizer",
|
26 |
"unk_token": {
|
27 |
"__type": "AddedToken",
|
|
|
21 |
"errors": "replace",
|
22 |
"model_max_length": 77,
|
23 |
"pad_token": "<|endoftext|>",
|
|
|
24 |
"tokenizer_class": "CLIPTokenizer",
|
25 |
"unk_token": {
|
26 |
"__type": "AddedToken",
|
unet/config.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
"_diffusers_version": "0.15.0.dev0",
|
4 |
-
"_name_or_path": "
|
5 |
"act_fn": "silu",
|
6 |
"attention_head_dim": 8,
|
7 |
"block_out_channels": [
|
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
"_diffusers_version": "0.15.0.dev0",
|
4 |
+
"_name_or_path": "/content/drive/MyDrive/Trained Model Data/checkpoint-2500",
|
5 |
"act_fn": "silu",
|
6 |
"attention_head_dim": 8,
|
7 |
"block_out_channels": [
|
unet/diffusion_pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3438375973
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae4153f5607af9eb1a58d816e1c22ad715a7718fc2945916fda5929d66074208
|
3 |
size 3438375973
|