import torch from torch import autocast from diffusers import StableDiffusionPipeline pipe = StableDiffusionPipeline.from_pretrained( 'hakurei/waifu-diffusion', torch_dtype=torch.float32 ).to('cuda') prompt = "1girl, aqua eyes, without clothes, blonde hair, open mouth, earrings, beach background, hat, hoop earrings, jewelry, looking at viewer, without shirt, short hair, simple background, solo, upper body, without shirt" with autocast("cuda"): image = pipe(prompt, guidance_scale=6)["sample"][0] image.save("test.png")
#32
by
Eireyn35
- opened
- .gitattributes +0 -8
- README.md +6 -8
- feature_extractor/preprocessor_config.json +2 -10
- model_index.json +3 -4
- safety_checker/config.json +3 -9
- safety_checker/model.fp16.safetensors +0 -3
- safety_checker/model.safetensors +0 -3
- safety_checker/pytorch_model.bin +2 -2
- safety_checker/pytorch_model.fp16.bin +0 -3
- scheduler/scheduler_config.json +2 -7
- text_encoder/config.json +8 -8
- text_encoder/model.fp16.safetensors +0 -3
- text_encoder/model.safetensors +0 -3
- text_encoder/pytorch_model.bin +2 -2
- text_encoder/pytorch_model.fp16.bin +0 -3
- tokenizer/special_tokens_map.json +1 -1
- tokenizer/tokenizer_config.json +1 -1
- unet/config.json +5 -16
- unet/diffusion_pytorch_model.bin +2 -2
- unet/diffusion_pytorch_model.fp16.bin +0 -3
- unet/diffusion_pytorch_model.fp16.safetensors +0 -3
- unet/diffusion_pytorch_model.safetensors +0 -3
- vae/config.json +2 -3
- vae/diffusion_pytorch_model.bin +2 -2
- vae/diffusion_pytorch_model.fp16.bin +0 -3
- vae/diffusion_pytorch_model.fp16.safetensors +0 -3
- vae/diffusion_pytorch_model.safetensors +0 -3
.gitattributes
CHANGED
@@ -33,11 +33,3 @@ safety_checker/pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
|
33 |
text_encoder/pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
34 |
unet/diffusion_pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
35 |
vae/diffusion_pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
36 |
-
text_encoder/model.safetensors filter=lfs diff=lfs merge=lfs -text
|
37 |
-
unet/diffusion_pytorch_model.safetensors filter=lfs diff=lfs merge=lfs -text
|
38 |
-
vae/diffusion_pytorch_model.safetensors filter=lfs diff=lfs merge=lfs -text
|
39 |
-
safety_checker/model.safetensors filter=lfs diff=lfs merge=lfs -text
|
40 |
-
unet/diffusion_pytorch_model.fp16.safetensors filter=lfs diff=lfs merge=lfs -text
|
41 |
-
text_encoder/model.fp16.safetensors filter=lfs diff=lfs merge=lfs -text
|
42 |
-
vae/diffusion_pytorch_model.fp16.safetensors filter=lfs diff=lfs merge=lfs -text
|
43 |
-
safety_checker/model.fp16.safetensors filter=lfs diff=lfs merge=lfs -text
|
|
|
33 |
text_encoder/pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
34 |
unet/diffusion_pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
35 |
vae/diffusion_pytorch_model.bin filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
README.md
CHANGED
@@ -5,19 +5,17 @@ tags:
|
|
5 |
- stable-diffusion
|
6 |
- text-to-image
|
7 |
license: creativeml-openrail-m
|
8 |
-
inference:
|
9 |
|
10 |
---
|
11 |
|
12 |
-
# waifu-diffusion v1.
|
13 |
|
14 |
waifu-diffusion is a latent text-to-image diffusion model that has been conditioned on high-quality anime images through fine-tuning.
|
15 |
|
16 |
-
|
17 |
|
18 |
-
|
19 |
-
|
20 |
-
[Original Weights](https://huggingface.co/hakurei/waifu-diffusion-v1-4)
|
21 |
|
22 |
# Gradio & Colab
|
23 |
|
@@ -64,9 +62,9 @@ image.save("test.png")
|
|
64 |
|
65 |
## Team Members and Acknowledgements
|
66 |
|
67 |
-
This project would not have been possible without the incredible work by
|
68 |
|
69 |
-
- [
|
70 |
- [Salt](https://github.com/sALTaccount/)
|
71 |
- [Sta @ Bit192](https://twitter.com/naclbbr)
|
72 |
|
|
|
5 |
- stable-diffusion
|
6 |
- text-to-image
|
7 |
license: creativeml-openrail-m
|
8 |
+
inference: false
|
9 |
|
10 |
---
|
11 |
|
12 |
+
# waifu-diffusion v1.3 - Diffusion for Weebs
|
13 |
|
14 |
waifu-diffusion is a latent text-to-image diffusion model that has been conditioned on high-quality anime images through fine-tuning.
|
15 |
|
16 |
+
<img src=https://i.imgur.com/Y5Tmw1S.png width=75% height=75%>
|
17 |
|
18 |
+
[Original Weights](https://huggingface.co/hakurei/waifu-diffusion-v1-3)
|
|
|
|
|
19 |
|
20 |
# Gradio & Colab
|
21 |
|
|
|
62 |
|
63 |
## Team Members and Acknowledgements
|
64 |
|
65 |
+
This project would not have been possible without the incredible work by the [CompVis Researchers](https://ommer-lab.com/).
|
66 |
|
67 |
+
- [Anthony Mercurio](https://github.com/harubaru)
|
68 |
- [Salt](https://github.com/sALTaccount/)
|
69 |
- [Sta @ Bit192](https://twitter.com/naclbbr)
|
70 |
|
feature_extractor/preprocessor_config.json
CHANGED
@@ -1,12 +1,8 @@
|
|
1 |
{
|
2 |
-
"crop_size":
|
3 |
-
"height": 224,
|
4 |
-
"width": 224
|
5 |
-
},
|
6 |
"do_center_crop": true,
|
7 |
"do_convert_rgb": true,
|
8 |
"do_normalize": true,
|
9 |
-
"do_rescale": true,
|
10 |
"do_resize": true,
|
11 |
"feature_extractor_type": "CLIPFeatureExtractor",
|
12 |
"image_mean": [
|
@@ -14,15 +10,11 @@
|
|
14 |
0.4578275,
|
15 |
0.40821073
|
16 |
],
|
17 |
-
"image_processor_type": "CLIPImageProcessor",
|
18 |
"image_std": [
|
19 |
0.26862954,
|
20 |
0.26130258,
|
21 |
0.27577711
|
22 |
],
|
23 |
"resample": 3,
|
24 |
-
"
|
25 |
-
"size": {
|
26 |
-
"shortest_edge": 224
|
27 |
-
}
|
28 |
}
|
|
|
1 |
{
|
2 |
+
"crop_size": 224,
|
|
|
|
|
|
|
3 |
"do_center_crop": true,
|
4 |
"do_convert_rgb": true,
|
5 |
"do_normalize": true,
|
|
|
6 |
"do_resize": true,
|
7 |
"feature_extractor_type": "CLIPFeatureExtractor",
|
8 |
"image_mean": [
|
|
|
10 |
0.4578275,
|
11 |
0.40821073
|
12 |
],
|
|
|
13 |
"image_std": [
|
14 |
0.26862954,
|
15 |
0.26130258,
|
16 |
0.27577711
|
17 |
],
|
18 |
"resample": 3,
|
19 |
+
"size": 224
|
|
|
|
|
|
|
20 |
}
|
model_index.json
CHANGED
@@ -1,18 +1,17 @@
|
|
1 |
{
|
2 |
"_class_name": "StableDiffusionPipeline",
|
3 |
-
"_diffusers_version": "0.
|
4 |
"feature_extractor": [
|
5 |
"transformers",
|
6 |
-
"
|
7 |
],
|
8 |
-
"requires_safety_checker": true,
|
9 |
"safety_checker": [
|
10 |
"stable_diffusion",
|
11 |
"StableDiffusionSafetyChecker"
|
12 |
],
|
13 |
"scheduler": [
|
14 |
"diffusers",
|
15 |
-
"
|
16 |
],
|
17 |
"text_encoder": [
|
18 |
"transformers",
|
|
|
1 |
{
|
2 |
"_class_name": "StableDiffusionPipeline",
|
3 |
+
"_diffusers_version": "0.4.1",
|
4 |
"feature_extractor": [
|
5 |
"transformers",
|
6 |
+
"CLIPFeatureExtractor"
|
7 |
],
|
|
|
8 |
"safety_checker": [
|
9 |
"stable_diffusion",
|
10 |
"StableDiffusionSafetyChecker"
|
11 |
],
|
12 |
"scheduler": [
|
13 |
"diffusers",
|
14 |
+
"LMSDiscreteScheduler"
|
15 |
],
|
16 |
"text_encoder": [
|
17 |
"transformers",
|
safety_checker/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_commit_hash":
|
3 |
"_name_or_path": "CompVis/stable-diffusion-safety-checker",
|
4 |
"architectures": [
|
5 |
"StableDiffusionSafetyChecker"
|
@@ -14,7 +14,6 @@
|
|
14 |
"architectures": null,
|
15 |
"attention_dropout": 0.0,
|
16 |
"bad_words_ids": null,
|
17 |
-
"begin_suppress_tokens": null,
|
18 |
"bos_token_id": 0,
|
19 |
"chunk_size_feed_forward": 0,
|
20 |
"cross_attention_hidden_size": null,
|
@@ -62,14 +61,12 @@
|
|
62 |
"pad_token_id": 1,
|
63 |
"prefix": null,
|
64 |
"problem_type": null,
|
65 |
-
"projection_dim": 512,
|
66 |
"pruned_heads": {},
|
67 |
"remove_invalid_values": false,
|
68 |
"repetition_penalty": 1.0,
|
69 |
"return_dict": true,
|
70 |
"return_dict_in_generate": false,
|
71 |
"sep_token_id": null,
|
72 |
-
"suppress_tokens": null,
|
73 |
"task_specific_params": null,
|
74 |
"temperature": 1.0,
|
75 |
"tf_legacy_loss": false,
|
@@ -80,7 +77,7 @@
|
|
80 |
"top_p": 1.0,
|
81 |
"torch_dtype": null,
|
82 |
"torchscript": false,
|
83 |
-
"transformers_version": "4.
|
84 |
"typical_p": 1.0,
|
85 |
"use_bfloat16": false,
|
86 |
"vocab_size": 49408
|
@@ -99,7 +96,6 @@
|
|
99 |
"architectures": null,
|
100 |
"attention_dropout": 0.0,
|
101 |
"bad_words_ids": null,
|
102 |
-
"begin_suppress_tokens": null,
|
103 |
"bos_token_id": null,
|
104 |
"chunk_size_feed_forward": 0,
|
105 |
"cross_attention_hidden_size": null,
|
@@ -149,14 +145,12 @@
|
|
149 |
"patch_size": 14,
|
150 |
"prefix": null,
|
151 |
"problem_type": null,
|
152 |
-
"projection_dim": 512,
|
153 |
"pruned_heads": {},
|
154 |
"remove_invalid_values": false,
|
155 |
"repetition_penalty": 1.0,
|
156 |
"return_dict": true,
|
157 |
"return_dict_in_generate": false,
|
158 |
"sep_token_id": null,
|
159 |
-
"suppress_tokens": null,
|
160 |
"task_specific_params": null,
|
161 |
"temperature": 1.0,
|
162 |
"tf_legacy_loss": false,
|
@@ -167,7 +161,7 @@
|
|
167 |
"top_p": 1.0,
|
168 |
"torch_dtype": null,
|
169 |
"torchscript": false,
|
170 |
-
"transformers_version": "4.
|
171 |
"typical_p": 1.0,
|
172 |
"use_bfloat16": false
|
173 |
},
|
|
|
1 |
{
|
2 |
+
"_commit_hash": null,
|
3 |
"_name_or_path": "CompVis/stable-diffusion-safety-checker",
|
4 |
"architectures": [
|
5 |
"StableDiffusionSafetyChecker"
|
|
|
14 |
"architectures": null,
|
15 |
"attention_dropout": 0.0,
|
16 |
"bad_words_ids": null,
|
|
|
17 |
"bos_token_id": 0,
|
18 |
"chunk_size_feed_forward": 0,
|
19 |
"cross_attention_hidden_size": null,
|
|
|
61 |
"pad_token_id": 1,
|
62 |
"prefix": null,
|
63 |
"problem_type": null,
|
|
|
64 |
"pruned_heads": {},
|
65 |
"remove_invalid_values": false,
|
66 |
"repetition_penalty": 1.0,
|
67 |
"return_dict": true,
|
68 |
"return_dict_in_generate": false,
|
69 |
"sep_token_id": null,
|
|
|
70 |
"task_specific_params": null,
|
71 |
"temperature": 1.0,
|
72 |
"tf_legacy_loss": false,
|
|
|
77 |
"top_p": 1.0,
|
78 |
"torch_dtype": null,
|
79 |
"torchscript": false,
|
80 |
+
"transformers_version": "4.22.2",
|
81 |
"typical_p": 1.0,
|
82 |
"use_bfloat16": false,
|
83 |
"vocab_size": 49408
|
|
|
96 |
"architectures": null,
|
97 |
"attention_dropout": 0.0,
|
98 |
"bad_words_ids": null,
|
|
|
99 |
"bos_token_id": null,
|
100 |
"chunk_size_feed_forward": 0,
|
101 |
"cross_attention_hidden_size": null,
|
|
|
145 |
"patch_size": 14,
|
146 |
"prefix": null,
|
147 |
"problem_type": null,
|
|
|
148 |
"pruned_heads": {},
|
149 |
"remove_invalid_values": false,
|
150 |
"repetition_penalty": 1.0,
|
151 |
"return_dict": true,
|
152 |
"return_dict_in_generate": false,
|
153 |
"sep_token_id": null,
|
|
|
154 |
"task_specific_params": null,
|
155 |
"temperature": 1.0,
|
156 |
"tf_legacy_loss": false,
|
|
|
161 |
"top_p": 1.0,
|
162 |
"torch_dtype": null,
|
163 |
"torchscript": false,
|
164 |
+
"transformers_version": "4.22.2",
|
165 |
"typical_p": 1.0,
|
166 |
"use_bfloat16": false
|
167 |
},
|
safety_checker/model.fp16.safetensors
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:08902f19b1cfebd7c989f152fc0507bef6898c706a91d666509383122324b511
|
3 |
-
size 608018440
|
|
|
|
|
|
|
|
safety_checker/model.safetensors
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:9d6a233ff6fd5ccb9f76fd99618d73369c52dd3d8222376384d0e601911089e8
|
3 |
-
size 1215981830
|
|
|
|
|
|
|
|
safety_checker/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:193490b58ef62739077262e833bf091c66c29488058681ac25cf7df3d8190974
|
3 |
+
size 1216061799
|
safety_checker/pytorch_model.fp16.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:22ba87205445ad5def13e54919b038dcfb7321ec1c3f4b12487d4fba6036125f
|
3 |
-
size 608103564
|
|
|
|
|
|
|
|
scheduler/scheduler_config.json
CHANGED
@@ -1,14 +1,9 @@
|
|
1 |
{
|
2 |
-
"_class_name": "
|
3 |
-
"_diffusers_version": "0.
|
4 |
"beta_end": 0.012,
|
5 |
"beta_schedule": "scaled_linear",
|
6 |
"beta_start": 0.00085,
|
7 |
-
"clip_sample": false,
|
8 |
"num_train_timesteps": 1000,
|
9 |
-
"prediction_type": "epsilon",
|
10 |
-
"set_alpha_to_one": false,
|
11 |
-
"skip_prk_steps": true,
|
12 |
-
"steps_offset": 1,
|
13 |
"trained_betas": null
|
14 |
}
|
|
|
1 |
{
|
2 |
+
"_class_name": "LMSDiscreteScheduler",
|
3 |
+
"_diffusers_version": "0.4.1",
|
4 |
"beta_end": 0.012,
|
5 |
"beta_schedule": "scaled_linear",
|
6 |
"beta_start": 0.00085,
|
|
|
7 |
"num_train_timesteps": 1000,
|
|
|
|
|
|
|
|
|
8 |
"trained_betas": null
|
9 |
}
|
text_encoder/config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "/
|
3 |
"architectures": [
|
4 |
"CLIPTextModel"
|
5 |
],
|
@@ -7,19 +7,19 @@
|
|
7 |
"bos_token_id": 0,
|
8 |
"dropout": 0.0,
|
9 |
"eos_token_id": 2,
|
10 |
-
"hidden_act": "
|
11 |
-
"hidden_size":
|
12 |
"initializer_factor": 1.0,
|
13 |
"initializer_range": 0.02,
|
14 |
-
"intermediate_size":
|
15 |
"layer_norm_eps": 1e-05,
|
16 |
"max_position_embeddings": 77,
|
17 |
"model_type": "clip_text_model",
|
18 |
-
"num_attention_heads":
|
19 |
-
"num_hidden_layers":
|
20 |
"pad_token_id": 1,
|
21 |
-
"projection_dim":
|
22 |
"torch_dtype": "float32",
|
23 |
-
"transformers_version": "4.
|
24 |
"vocab_size": 49408
|
25 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "openai/clip-vit-large-patch14",
|
3 |
"architectures": [
|
4 |
"CLIPTextModel"
|
5 |
],
|
|
|
7 |
"bos_token_id": 0,
|
8 |
"dropout": 0.0,
|
9 |
"eos_token_id": 2,
|
10 |
+
"hidden_act": "quick_gelu",
|
11 |
+
"hidden_size": 768,
|
12 |
"initializer_factor": 1.0,
|
13 |
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 3072,
|
15 |
"layer_norm_eps": 1e-05,
|
16 |
"max_position_embeddings": 77,
|
17 |
"model_type": "clip_text_model",
|
18 |
+
"num_attention_heads": 12,
|
19 |
+
"num_hidden_layers": 12,
|
20 |
"pad_token_id": 1,
|
21 |
+
"projection_dim": 768,
|
22 |
"torch_dtype": "float32",
|
23 |
+
"transformers_version": "4.22.2",
|
24 |
"vocab_size": 49408
|
25 |
}
|
text_encoder/model.fp16.safetensors
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:22bc8e104d064b678ef7d2d2b217d4a8c9bfb79fb35792417cdf228e70adc7fb
|
3 |
-
size 680821096
|
|
|
|
|
|
|
|
text_encoder/model.safetensors
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:661a5d7f8e19fce696aa9d932ab97b546b4d4a2a2d87238a17761bef2704269f
|
3 |
-
size 1361597016
|
|
|
|
|
|
|
|
text_encoder/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:770a47a9ffdcfda0b05506a7888ed714d06131d60267e6cf52765d61cf59fd67
|
3 |
+
size 492305335
|
text_encoder/pytorch_model.fp16.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:ba47a6c751cce5b6c4d5c79c8cd63bab63bfce3539e7805a70b8dca9d1f2f151
|
3 |
-
size 680900852
|
|
|
|
|
|
|
|
tokenizer/special_tokens_map.json
CHANGED
@@ -13,7 +13,7 @@
|
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
15 |
},
|
16 |
-
"pad_token": "
|
17 |
"unk_token": {
|
18 |
"content": "<|endoftext|>",
|
19 |
"lstrip": false,
|
|
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
15 |
},
|
16 |
+
"pad_token": "<|endoftext|>",
|
17 |
"unk_token": {
|
18 |
"content": "<|endoftext|>",
|
19 |
"lstrip": false,
|
tokenizer/tokenizer_config.json
CHANGED
@@ -19,7 +19,7 @@
|
|
19 |
},
|
20 |
"errors": "replace",
|
21 |
"model_max_length": 77,
|
22 |
-
"name_or_path": "/
|
23 |
"pad_token": "<|endoftext|>",
|
24 |
"special_tokens_map_file": "./special_tokens_map.json",
|
25 |
"tokenizer_class": "CLIPTokenizer",
|
|
|
19 |
},
|
20 |
"errors": "replace",
|
21 |
"model_max_length": 77,
|
22 |
+
"name_or_path": "openai/clip-vit-large-patch14",
|
23 |
"pad_token": "<|endoftext|>",
|
24 |
"special_tokens_map_file": "./special_tokens_map.json",
|
25 |
"tokenizer_class": "CLIPTokenizer",
|
unet/config.json
CHANGED
@@ -1,14 +1,8 @@
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
-
"_diffusers_version": "0.
|
4 |
-
"_name_or_path": "/mnt/sd-finetune-data/finetunes/step_57000",
|
5 |
"act_fn": "silu",
|
6 |
-
"attention_head_dim":
|
7 |
-
5,
|
8 |
-
10,
|
9 |
-
20,
|
10 |
-
20
|
11 |
-
],
|
12 |
"block_out_channels": [
|
13 |
320,
|
14 |
640,
|
@@ -16,7 +10,7 @@
|
|
16 |
1280
|
17 |
],
|
18 |
"center_input_sample": false,
|
19 |
-
"cross_attention_dim":
|
20 |
"down_block_types": [
|
21 |
"CrossAttnDownBlock2D",
|
22 |
"CrossAttnDownBlock2D",
|
@@ -24,7 +18,6 @@
|
|
24 |
"DownBlock2D"
|
25 |
],
|
26 |
"downsample_padding": 1,
|
27 |
-
"dual_cross_attention": false,
|
28 |
"flip_sin_to_cos": true,
|
29 |
"freq_shift": 0,
|
30 |
"in_channels": 4,
|
@@ -32,16 +25,12 @@
|
|
32 |
"mid_block_scale_factor": 1,
|
33 |
"norm_eps": 1e-05,
|
34 |
"norm_num_groups": 32,
|
35 |
-
"num_class_embeds": null,
|
36 |
-
"only_cross_attention": false,
|
37 |
"out_channels": 4,
|
38 |
-
"sample_size":
|
39 |
"up_block_types": [
|
40 |
"UpBlock2D",
|
41 |
"CrossAttnUpBlock2D",
|
42 |
"CrossAttnUpBlock2D",
|
43 |
"CrossAttnUpBlock2D"
|
44 |
-
]
|
45 |
-
"upcast_attention": false,
|
46 |
-
"use_linear_projection": true
|
47 |
}
|
|
|
1 |
{
|
2 |
"_class_name": "UNet2DConditionModel",
|
3 |
+
"_diffusers_version": "0.4.1",
|
|
|
4 |
"act_fn": "silu",
|
5 |
+
"attention_head_dim": 8,
|
|
|
|
|
|
|
|
|
|
|
6 |
"block_out_channels": [
|
7 |
320,
|
8 |
640,
|
|
|
10 |
1280
|
11 |
],
|
12 |
"center_input_sample": false,
|
13 |
+
"cross_attention_dim": 768,
|
14 |
"down_block_types": [
|
15 |
"CrossAttnDownBlock2D",
|
16 |
"CrossAttnDownBlock2D",
|
|
|
18 |
"DownBlock2D"
|
19 |
],
|
20 |
"downsample_padding": 1,
|
|
|
21 |
"flip_sin_to_cos": true,
|
22 |
"freq_shift": 0,
|
23 |
"in_channels": 4,
|
|
|
25 |
"mid_block_scale_factor": 1,
|
26 |
"norm_eps": 1e-05,
|
27 |
"norm_num_groups": 32,
|
|
|
|
|
28 |
"out_channels": 4,
|
29 |
+
"sample_size": 32,
|
30 |
"up_block_types": [
|
31 |
"UpBlock2D",
|
32 |
"CrossAttnUpBlock2D",
|
33 |
"CrossAttnUpBlock2D",
|
34 |
"CrossAttnUpBlock2D"
|
35 |
+
]
|
|
|
|
|
36 |
}
|
unet/diffusion_pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f47e5665f0e85155a5f6f58683b04940c6b132023d584396226bf54419a78831
|
3 |
+
size 3438354725
|
unet/diffusion_pytorch_model.fp16.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:96edda2701914e1e248197bb205305e6aa9cfc776c3372cff9eaf62d4706a3cf
|
3 |
-
size 1732107093
|
|
|
|
|
|
|
|
unet/diffusion_pytorch_model.fp16.safetensors
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:7b5cdd1c15f025166ded673f898e09e621c9ff828d6508a81e83378a6d0ba8dd
|
3 |
-
size 1731904736
|
|
|
|
|
|
|
|
unet/diffusion_pytorch_model.safetensors
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:dda5a15fe85e6ea7fe0e21d06264611246ab60bdbf7001daa1e48028a49cd2e3
|
3 |
-
size 3463726500
|
|
|
|
|
|
|
|
vae/config.json
CHANGED
@@ -1,7 +1,6 @@
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
-
"_diffusers_version": "0.
|
4 |
-
"_name_or_path": "/mnt/sd-finetune-data/base/vae",
|
5 |
"act_fn": "silu",
|
6 |
"block_out_channels": [
|
7 |
128,
|
@@ -20,7 +19,7 @@
|
|
20 |
"layers_per_block": 2,
|
21 |
"norm_num_groups": 32,
|
22 |
"out_channels": 3,
|
23 |
-
"sample_size":
|
24 |
"up_block_types": [
|
25 |
"UpDecoderBlock2D",
|
26 |
"UpDecoderBlock2D",
|
|
|
1 |
{
|
2 |
"_class_name": "AutoencoderKL",
|
3 |
+
"_diffusers_version": "0.4.1",
|
|
|
4 |
"act_fn": "silu",
|
5 |
"block_out_channels": [
|
6 |
128,
|
|
|
19 |
"layers_per_block": 2,
|
20 |
"norm_num_groups": 32,
|
21 |
"out_channels": 3,
|
22 |
+
"sample_size": 256,
|
23 |
"up_block_types": [
|
24 |
"UpDecoderBlock2D",
|
25 |
"UpDecoderBlock2D",
|
vae/diffusion_pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b134cded8eb78b184aefb8805b6b572f36fa77b255c483665dda931fa0130c5
|
3 |
+
size 334707217
|
vae/diffusion_pytorch_model.fp16.bin
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:d207e4928394a2002bcb7fff829e93bbd2a44bc323e597fdb690d3fc2d064de2
|
3 |
-
size 167405651
|
|
|
|
|
|
|
|
vae/diffusion_pytorch_model.fp16.safetensors
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:9d8fb415ab4f9782232e7bb82e618e2c0cef0be3593c77a35f5733d8fdd3530f
|
3 |
-
size 167335342
|
|
|
|
|
|
|
|
vae/diffusion_pytorch_model.safetensors
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:d55443a2d9d4d9decdbe669c51cc6d91eb6a2297477624e2e16a3054f30c2f5a
|
3 |
-
size 334643276
|
|
|
|
|
|
|
|