hollowstrawberry
commited on
Commit
•
ec2c447
1
Parent(s):
82cc7ad
Update README.md
Browse files
README.md
CHANGED
@@ -8,26 +8,26 @@ tags:
|
|
8 |
- stable-diffusion-webui
|
9 |
---
|
10 |
|
11 |
-
# Index <a name="index"></a>
|
12 |
-
|
13 |
-
* [Introduction](#intro)
|
14 |
-
* [Installation](#install)
|
15 |
-
* [Getting Started](#start)
|
16 |
-
1. [Edit your starting parameters](#params)
|
17 |
-
1. [Getting a model](#model)
|
18 |
-
1. [Getting a VAE](#vae)
|
19 |
-
1. [Launching and settings](#launch)
|
20 |
-
1. [Prompts](#prompt)
|
21 |
-
1. [Generation parameters](#gen)
|
22 |
-
* [Extensions](#extensions)
|
23 |
-
* [Loras](#lora)
|
24 |
-
* [Upscalers](#upscale)
|
25 |
* ControlNet
|
26 |
* Tips for training character Loras
|
27 |
|
28 |
|
29 |
|
30 |
-
# Introduction <a name="intro"></a>[▲](#index)
|
31 |
|
32 |
Stable Diffusion is a very powerful AI image generation software you can run on your own home computer. It uses "models", which function like the brain of the AI, and can make almost anything given that someone has trained it to do it. The biggest uses are anime art, photorealism, and NSFW content.
|
33 |
|
@@ -35,7 +35,7 @@ The images you create may be used for any purpose, depending on the used model's
|
|
35 |
|
36 |
|
37 |
|
38 |
-
# Installation <a name="install"></a>[▲](#index)
|
39 |
|
40 |
* __**Requirements:**__ To run Stable Diffusion on your own computer you'll need at least 16 GB of RAM and 4 GB of VRAM. I will only cover the case where you are running Windows 10/11 and using an NVIDIA graphics card series 16XX, 20XX or 30XX (though 10XX also work). AMD users are out of luck, as it's very inconsistent to get it working. Same with Linux and Mac, though it's possible to do it.
|
41 |
|
@@ -45,11 +45,11 @@ The images you create may be used for any purpose, depending on the used model's
|
|
45 |
|
46 |
|
47 |
|
48 |
-
# Getting Started <a name="start"></a>[▲](#index)
|
49 |
|
50 |
Before generating some images, here are some useful steps you can follow to improve your experience.
|
51 |
|
52 |
-
1. **Edit your starting parameters** <a name="params"></a>[
|
53 |
|
54 |
If you're using the collab, skip this step.
|
55 |
|
@@ -60,7 +60,7 @@ Before generating some images, here are some useful steps you can follow to impr
|
|
60 |
* If you're using the original stable-diffusion-webui, you can add these parameters by editing your webui-user.bat, right next to `set COMMANDLINE_ARGS=`
|
61 |
* Full list of possible parameters [here](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Command-Line-Arguments-and-Settings)
|
62 |
|
63 |
-
1. **Getting a model** <a name="model"></a>[▲](#index)
|
64 |
|
65 |
The model is the brain of your AI, designed for the purpose of producing certain types of images. There are many options, most of which are on [civitai](https://civitai.com). But which to choose? These are my recommendations:
|
66 |
* For anime, [7th Heaven Mix](https://civitai.com/models/4669/corneos-7th-heaven-mix) has a nice aesthetic similar to anime movies, while [Abyss Orange Mix 3](https://civitai.com/models/9942/abyssorangemix3-aom3) *(__Note:__ scroll down and choose the AOM3 option)* offers more realism in the form of advanced lighting and softer shading, as well as more lewdness. I remixed the two options above into [Heaven Orange Mix](https://civitai.com/models/14305/heavenorangemix). While AOM3 is extremely capable for NSFW, the popular [Grapefruit](https://civitai.com/models/2583/grapefruit-hentai-model) hentai model may also fit your needs.
|
@@ -74,7 +74,7 @@ Before generating some images, here are some useful steps you can follow to impr
|
|
74 |
|
75 |
Please note that models in the format `.safetensors` are safe to use while `.ckpt` **may** contain viruses. Be careful.
|
76 |
|
77 |
-
1. **Getting a VAE** <a name="vae"></a>[▲](#index)
|
78 |
|
79 |
Most models don't come with a VAE built in. The VAE is a small separate model, which "converts your image from AI format into human format". Without it, you'll get faded colors and ugly eyes, among other things.
|
80 |
|
@@ -87,7 +87,7 @@ Before generating some images, here are some useful steps you can follow to impr
|
|
87 |
|
88 |
*Collab:* You will have to place it in your Google Drive, in `MyDrive/sd/stable-diffusion-webui/models/VAE`.
|
89 |
|
90 |
-
1. **Launching and settings** <a name="launch"></a>[▲](#index)
|
91 |
|
92 |
It is finally time to launch the WebUI.
|
93 |
*Launcher:* Press the button on the launcher and wait patiently for it to start. Then, it will open the interface in your browser. It's like a website, but on your computer.
|
@@ -98,7 +98,7 @@ Before generating some images, here are some useful steps you can follow to impr
|
|
98 |
* In the *User Interface* section, scroll down to **Quicksettings list** and change it to `sd_model_checkpoint, sd_vae`.
|
99 |
* Scroll back up, click the big orange **Apply settings** button, then **Reload UI** next to it. You can now change your model as well as your VAE from the top of the page at any time.
|
100 |
|
101 |
-
1. **Prompts** <a name="prompt"></a>[▲](#index)
|
102 |
|
103 |
On the first tab, **txt2img**, you'll be making most of your images. This is where you'll find your *prompt* and *negative prompt*.
|
104 |
Stable Diffusion is not like Midjourney or other popular image generation software, you can't just ask it what you want and get a good image. You have to be specific. *Very* specific.
|
@@ -120,7 +120,7 @@ Before generating some images, here are some useful steps you can follow to impr
|
|
120 |
|
121 |
Note that when you surround something in `(parentheses)`, it will have emphasis or more **weight** in your resulting image, equal to `1.1`. The normal weight is 1, and each parentheses will multiply by an additional 1.1. You can also specify the weight yourself, like this: `(full body:1.4)`. You can also go below 1 to de-emphasize a word: `[brackets]` will multiply by 0.9, but you still use normal parentheses to go lower, like `(this:0.5)`.
|
122 |
|
123 |
-
1. **Generation parameters** <a name="gen"></a>[▲](#index)
|
124 |
|
125 |
* *Sampling method:* These dictate how your image is formulated, and each produce different results. The default of `Euler a` is almost always the best. There are also very good results for `DPM++ 2M Karras` and `DPM++ SDE Karras`.
|
126 |
* *Sampling steps:* These are "calculated" beforehand, and so more steps doesn't always mean more detail. I always go with 30, you may go from 20-50 and find good results.
|
@@ -130,7 +130,7 @@ Before generating some images, here are some useful steps you can follow to impr
|
|
130 |
* *Seed:* A number that guides the creation of your image. The same seed with the same prompt and parameters produces almost exacly the same image every time.
|
131 |
|
132 |
*Hires. fix:* Lets you create larger images without distortion. Usually used at 2x scale. When selected, more options appear:
|
133 |
-
* *Upscaler:* The algorithm to upscale with. `Latent` and its variations produce creative results, and you may also like `R-ESRGAN 4x+` and its anime version. Also see [Upscalers](#upscale).
|
134 |
* *Hires steps:* I recommend at least half as many as your sampling steps. Higher values aren't always better.
|
135 |
* *Denoising strength:* The most important parameter. Near 0, no detail will be added to the image. Near 1, the image will be changed completely. I recommend something between 0.2 and 0.6 depending on the image.
|
136 |
|
@@ -141,7 +141,7 @@ Before generating some images, here are some useful steps you can follow to impr
|
|
141 |
|
142 |
|
143 |
|
144 |
-
# Extensions <a name="extensions"></a>[▲](#index)
|
145 |
|
146 |
*Stable Diffusion WebUI* supports extensions to add additional functionality and quality of life. These can be added by going into the **Extensions** tab, then **Install from URL**, and pasting the links found here or elsewhere. Then, click *Install* and wait for it to finish. Then, go to **Installed** and click *Apply and restart UI*.
|
147 |
|
@@ -156,19 +156,19 @@ Here are some useful extensions, I hugely recommend the first 2:
|
|
156 |
|
157 |
|
158 |
|
159 |
-
# Loras <a name="lora"></a>[▲](#index)
|
160 |
|
161 |
-
LoRA or *Low-Rank Adaptation* is a form of **Extra Network** and the latest technology that lets you append a smaller model to any of your full models. They are similar to embeddings, one of which you might've seen [earlier](#prompt), but Loras are larger and often more capable. Technical details omitted.
|
162 |
|
163 |
Loras can represent a character, an artstyle, poses, clothes, or even a human face (though I do not endorse this). Models are usually capable enough for general work, but when it comes to specific details with little existing examples, they fall short. That's where Loras come in. They can be downloaded from [civitai](https://civitai.com) and are 144 MB by default, but they can go as low as 1 MB and sometimes several hundreds of MB. Bigger Loras are not necessarily better. They come in `.safetensor` format, same as models.
|
164 |
|
165 |
-
Place your lora files in the `stable-diffusion-webui/models/Lora` folder. Then, look for the 🎴 *Show extra networks* button below the big orange Generate button. It will open a new section. Click on the Lora tab and press the **Refresh** button, and your loras should appear. When you click a Lora in that menu it will get added to your prompt. It will look like this: `<lora:filename:1` The start is always the same. The filename will be the exact filename in your system without the `.safetensors` extension. Finally, the number is the weight, like we saw in [Prompts](#prompt). Most Loras work between 0.5 and 1 weight, and too high values might "fry" your image.
|
166 |
|
167 |
An example of a Lora is [Thicker Lines Anime Style](https://civitai.com/models/13910/thicker-lines-anime-style-lora-mix), which is perfect if you want your images to look more like traditional anime.
|
168 |
|
169 |
|
170 |
|
171 |
-
# Upscalers <a name="upscale"></a>[▲](#index)
|
172 |
|
173 |
You can download additional upscalers and put them in your `stable-diffusion-webui/models/ESRGAN` folder.
|
174 |
|
|
|
8 |
- stable-diffusion-webui
|
9 |
---
|
10 |
|
11 |
+
# Index <a name="sdguide-index"></a>
|
12 |
+
|
13 |
+
* [Introduction](#sdguide-intro)
|
14 |
+
* [Installation](#sdguide-install)
|
15 |
+
* [Getting Started](#sdguide-start)
|
16 |
+
1. [Edit your starting parameters](#sdguide-params)
|
17 |
+
1. [Getting a model](#sdguide-model)
|
18 |
+
1. [Getting a VAE](#sdguide-vae)
|
19 |
+
1. [Launching and settings](#sdguide-launch)
|
20 |
+
1. [Prompts](#sdguide-prompt)
|
21 |
+
1. [Generation parameters](#sdguide-gen)
|
22 |
+
* [Extensions](#sdguide-extensions)
|
23 |
+
* [Loras](#sdguide-lora)
|
24 |
+
* [Upscalers](#sdguide-upscale)
|
25 |
* ControlNet
|
26 |
* Tips for training character Loras
|
27 |
|
28 |
|
29 |
|
30 |
+
# Introduction <a name="sdguide-intro"></a>[▲](#sdguide-index)
|
31 |
|
32 |
Stable Diffusion is a very powerful AI image generation software you can run on your own home computer. It uses "models", which function like the brain of the AI, and can make almost anything given that someone has trained it to do it. The biggest uses are anime art, photorealism, and NSFW content.
|
33 |
|
|
|
35 |
|
36 |
|
37 |
|
38 |
+
# Installation <a name="sdguide-install"></a>[▲](#sdguide-index)
|
39 |
|
40 |
* __**Requirements:**__ To run Stable Diffusion on your own computer you'll need at least 16 GB of RAM and 4 GB of VRAM. I will only cover the case where you are running Windows 10/11 and using an NVIDIA graphics card series 16XX, 20XX or 30XX (though 10XX also work). AMD users are out of luck, as it's very inconsistent to get it working. Same with Linux and Mac, though it's possible to do it.
|
41 |
|
|
|
45 |
|
46 |
|
47 |
|
48 |
+
# Getting Started <a name="sdguide-start"></a>[▲](#sdguide-index)
|
49 |
|
50 |
Before generating some images, here are some useful steps you can follow to improve your experience.
|
51 |
|
52 |
+
1. **Edit your starting parameters** <a name="sdguide-params"></a>[▲](#sdguide-index)
|
53 |
|
54 |
If you're using the collab, skip this step.
|
55 |
|
|
|
60 |
* If you're using the original stable-diffusion-webui, you can add these parameters by editing your webui-user.bat, right next to `set COMMANDLINE_ARGS=`
|
61 |
* Full list of possible parameters [here](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/Command-Line-Arguments-and-Settings)
|
62 |
|
63 |
+
1. **Getting a model** <a name="sdguide-model"></a>[▲](#sdguide-index)
|
64 |
|
65 |
The model is the brain of your AI, designed for the purpose of producing certain types of images. There are many options, most of which are on [civitai](https://civitai.com). But which to choose? These are my recommendations:
|
66 |
* For anime, [7th Heaven Mix](https://civitai.com/models/4669/corneos-7th-heaven-mix) has a nice aesthetic similar to anime movies, while [Abyss Orange Mix 3](https://civitai.com/models/9942/abyssorangemix3-aom3) *(__Note:__ scroll down and choose the AOM3 option)* offers more realism in the form of advanced lighting and softer shading, as well as more lewdness. I remixed the two options above into [Heaven Orange Mix](https://civitai.com/models/14305/heavenorangemix). While AOM3 is extremely capable for NSFW, the popular [Grapefruit](https://civitai.com/models/2583/grapefruit-hentai-model) hentai model may also fit your needs.
|
|
|
74 |
|
75 |
Please note that models in the format `.safetensors` are safe to use while `.ckpt` **may** contain viruses. Be careful.
|
76 |
|
77 |
+
1. **Getting a VAE** <a name="sdguide-vae"></a>[▲](#sdguide-index)
|
78 |
|
79 |
Most models don't come with a VAE built in. The VAE is a small separate model, which "converts your image from AI format into human format". Without it, you'll get faded colors and ugly eyes, among other things.
|
80 |
|
|
|
87 |
|
88 |
*Collab:* You will have to place it in your Google Drive, in `MyDrive/sd/stable-diffusion-webui/models/VAE`.
|
89 |
|
90 |
+
1. **Launching and settings** <a name="sdguide-launch"></a>[▲](#sdguide-index)
|
91 |
|
92 |
It is finally time to launch the WebUI.
|
93 |
*Launcher:* Press the button on the launcher and wait patiently for it to start. Then, it will open the interface in your browser. It's like a website, but on your computer.
|
|
|
98 |
* In the *User Interface* section, scroll down to **Quicksettings list** and change it to `sd_model_checkpoint, sd_vae`.
|
99 |
* Scroll back up, click the big orange **Apply settings** button, then **Reload UI** next to it. You can now change your model as well as your VAE from the top of the page at any time.
|
100 |
|
101 |
+
1. **Prompts** <a name="sdguide-prompt"></a>[▲](#sdguide-index)
|
102 |
|
103 |
On the first tab, **txt2img**, you'll be making most of your images. This is where you'll find your *prompt* and *negative prompt*.
|
104 |
Stable Diffusion is not like Midjourney or other popular image generation software, you can't just ask it what you want and get a good image. You have to be specific. *Very* specific.
|
|
|
120 |
|
121 |
Note that when you surround something in `(parentheses)`, it will have emphasis or more **weight** in your resulting image, equal to `1.1`. The normal weight is 1, and each parentheses will multiply by an additional 1.1. You can also specify the weight yourself, like this: `(full body:1.4)`. You can also go below 1 to de-emphasize a word: `[brackets]` will multiply by 0.9, but you still use normal parentheses to go lower, like `(this:0.5)`.
|
122 |
|
123 |
+
1. **Generation parameters** <a name="sdguide-gen"></a>[▲](#sdguide-index)
|
124 |
|
125 |
* *Sampling method:* These dictate how your image is formulated, and each produce different results. The default of `Euler a` is almost always the best. There are also very good results for `DPM++ 2M Karras` and `DPM++ SDE Karras`.
|
126 |
* *Sampling steps:* These are "calculated" beforehand, and so more steps doesn't always mean more detail. I always go with 30, you may go from 20-50 and find good results.
|
|
|
130 |
* *Seed:* A number that guides the creation of your image. The same seed with the same prompt and parameters produces almost exacly the same image every time.
|
131 |
|
132 |
*Hires. fix:* Lets you create larger images without distortion. Usually used at 2x scale. When selected, more options appear:
|
133 |
+
* *Upscaler:* The algorithm to upscale with. `Latent` and its variations produce creative results, and you may also like `R-ESRGAN 4x+` and its anime version. Also see [Upscalers](#sdguide-upscale).
|
134 |
* *Hires steps:* I recommend at least half as many as your sampling steps. Higher values aren't always better.
|
135 |
* *Denoising strength:* The most important parameter. Near 0, no detail will be added to the image. Near 1, the image will be changed completely. I recommend something between 0.2 and 0.6 depending on the image.
|
136 |
|
|
|
141 |
|
142 |
|
143 |
|
144 |
+
# Extensions <a name="sdguide-extensions"></a>[▲](#sdguide-index)
|
145 |
|
146 |
*Stable Diffusion WebUI* supports extensions to add additional functionality and quality of life. These can be added by going into the **Extensions** tab, then **Install from URL**, and pasting the links found here or elsewhere. Then, click *Install* and wait for it to finish. Then, go to **Installed** and click *Apply and restart UI*.
|
147 |
|
|
|
156 |
|
157 |
|
158 |
|
159 |
+
# Loras <a name="sdguide-lora"></a>[▲](#sdguide-index)
|
160 |
|
161 |
+
LoRA or *Low-Rank Adaptation* is a form of **Extra Network** and the latest technology that lets you append a smaller model to any of your full models. They are similar to embeddings, one of which you might've seen [earlier](#sdguide-prompt), but Loras are larger and often more capable. Technical details omitted.
|
162 |
|
163 |
Loras can represent a character, an artstyle, poses, clothes, or even a human face (though I do not endorse this). Models are usually capable enough for general work, but when it comes to specific details with little existing examples, they fall short. That's where Loras come in. They can be downloaded from [civitai](https://civitai.com) and are 144 MB by default, but they can go as low as 1 MB and sometimes several hundreds of MB. Bigger Loras are not necessarily better. They come in `.safetensor` format, same as models.
|
164 |
|
165 |
+
Place your lora files in the `stable-diffusion-webui/models/Lora` folder. Then, look for the 🎴 *Show extra networks* button below the big orange Generate button. It will open a new section. Click on the Lora tab and press the **Refresh** button, and your loras should appear. When you click a Lora in that menu it will get added to your prompt. It will look like this: `<lora:filename:1` The start is always the same. The filename will be the exact filename in your system without the `.safetensors` extension. Finally, the number is the weight, like we saw in [Prompts](#sdguide-prompt). Most Loras work between 0.5 and 1 weight, and too high values might "fry" your image.
|
166 |
|
167 |
An example of a Lora is [Thicker Lines Anime Style](https://civitai.com/models/13910/thicker-lines-anime-style-lora-mix), which is perfect if you want your images to look more like traditional anime.
|
168 |
|
169 |
|
170 |
|
171 |
+
# Upscalers <a name="sdguide-upscale"></a>[▲](#sdguide-index)
|
172 |
|
173 |
You can download additional upscalers and put them in your `stable-diffusion-webui/models/ESRGAN` folder.
|
174 |
|