hollowstrawberry
commited on
Commit
•
463cf24
1
Parent(s):
7d549ab
Update README.md
Browse files
README.md
CHANGED
@@ -178,15 +178,15 @@ Here you can select your checkpoint and VAE. We will go over what these are and
|
|
178 |
|
179 |
![Parameters](images/parameters.png)
|
180 |
|
181 |
-
* **Sampling method:** This is the algorithm that formulates your image, and each produce different results. The default of `Euler a` is
|
182 |
-
* **Sampling steps:** These are "calculated" beforehand, and so more steps doesn't always mean more detail. I always go with 30, you may go from 20-50 and find consistently good results.
|
183 |
* **Width and Height:** 512x512 is the default, and you should almost never go above 768 in either direction as it may distort and deform your image. To produce bigger images see `Hires fix`.
|
184 |
* **Batch Count and Batch Size:** Batch *size* is how many images your graphics card will generate at the same time, which is limited by its VRAM. Batch *count* is how many times to repeat those. Batches have consecutive seeds, more on seeds below.
|
185 |
* **CFG Scale:** "Lower values produce more creative results". You should almost always stick to 7, but 4 to 10 is an acceptable range.
|
186 |
* **Seed:** A number that guides the creation of your image. The same seed with the same prompt and parameters produces the same image every time, except for small details and under some circumstances.
|
187 |
|
188 |
**Hires fix:** Lets you create larger images without distortion. Often used at 2x scale. When selected, more options appear:
|
189 |
-
* **Upscaler:** The algorithm to upscale with. `Latent` and its variations
|
190 |
* **Hires steps:** I recommend at least half as many as your sampling steps. Higher values aren't always better, and they take a long time, so be conservative here.
|
191 |
* **Denoising strength:** The most important parameter. Near 0.0, no detail will be added to the image. Near 1.0, the image will be changed completely. I recommend something between 0.2 and 0.6 depending on the image, to add enough detail as the image gets larger, without *destroying* any original details you like.
|
192 |
|
@@ -195,6 +195,22 @@ Here you can select your checkpoint and VAE. We will go over what these are and
|
|
195 |
* **Tiling:** Used to produce repeating textures to put on a grid. Not very useful.
|
196 |
* **Script:** Lets you access useful features and extensions, such as [X/Y/Z Plot ▼](#plot) which lets you compare images with varying parameters on a grid. Very powerful.
|
197 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
198 |
|
199 |
|
200 |
# Extensions <a name="extensions"></a>[▲](#index)
|
@@ -243,7 +259,7 @@ The collab in this guide comes with several of them, including **Remacri**, whic
|
|
243 |
Here are some comparisons. All of them were done at 0.4 denoising strength. Note that some of the differences may be completely up to random chance.
|
244 |
|
245 |
<details>
|
246 |
-
|
247 |
|
248 |
**Some details to consider:** The fireballs to the left and right, the texture of the fire around her, the grass and its flowers, the ghost's face, the flowers in her hat, the hands, the eyes (which should be flower-shaped), the things on her waist.
|
249 |
|
@@ -255,7 +271,7 @@ Here are some comparisons. All of them were done at 0.4 denoising strength. Note
|
|
255 |
</details>
|
256 |
|
257 |
<details>
|
258 |
-
|
259 |
|
260 |
**Some details to consider:** The background, the flower and symbol on her hat, the flowers on the branches to the sides, the eyes (which should be flower-shaped), the emblem below her neck, The pattern on the lower half of her dress, as well as the nearby frills and folds.
|
261 |
|
@@ -268,7 +284,7 @@ Here are some comparisons. All of them were done at 0.4 denoising strength. Note
|
|
268 |
</details>
|
269 |
|
270 |
<details>
|
271 |
-
|
272 |
|
273 |
**Some details to consider:** The eye on the left, the finger creases, the bracelet, the edge trim on the vest, the flower on the vest, the brooches on the vest, the rocks and vegetation on the bottom left, the trees on the top left, the waterfalls of course.
|
274 |
|
@@ -297,7 +313,7 @@ Scripts can be found at the bottom of your generation parameters in txt2img or i
|
|
297 |
Here I made a comparison between different **models** (columns) and faces of different ethnicities via **S/R Prompt** (rows):
|
298 |
|
299 |
<details>
|
300 |
-
<summary>X/Y/Z Plot example
|
301 |
|
302 |
![X Y Z plot of models and ethnicities](images/XYZplot.png)
|
303 |
</details>
|
@@ -313,12 +329,14 @@ Scripts can be found at the bottom of your generation parameters in txt2img or i
|
|
313 |
<a name="matrixneg"></a>Here is a comparison using the negative prompts I showed you in [Prompts ▲](#prompt). We can see how EasyNegative affects the image, as well as how the rest of the prompt affects the image, then both together:
|
314 |
|
315 |
<details>
|
316 |
-
<summary>Prompt matrix examples
|
317 |
|
318 |
![Prompt matrix of anime negative prompt sections](images/promptmatrix1.png)
|
319 |
![Prompt matrix of photorealistic negative prompt sections](images/promptmatrix2.png)
|
320 |
</details>
|
321 |
|
|
|
|
|
322 |
* **Ultimate Upscaler** <a name="ultimate"></a>[▲](#index)
|
323 |
|
324 |
An improved version of a builtin script, it can be added as an [extension ▲](#extensions) and used from within **img2img**. Its purpose is to resize an image and add more detail way past the normal limits of your VRAM by splitting it into chunks, although slower. Here are the steps:
|
@@ -358,7 +376,7 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
|
|
358 |
The Canny method extracts the hard edges of the sample image. It is useful for many different types of images, specially where you want to preserve small details and the general look of an image. Observe:
|
359 |
|
360 |
<details>
|
361 |
-
<summary>Canny example
|
362 |
|
363 |
![Canny preprocessed image](images/canny1.png)
|
364 |
![Canny output image](images/canny2.png)
|
@@ -369,7 +387,7 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
|
|
369 |
The Depth method extracts the 3D elements of the sample image. It is best suited for complex environments and general composition. Observe:
|
370 |
|
371 |
<details>
|
372 |
-
<summary>Depth example
|
373 |
|
374 |
![Depth preprocessed image](images/depth1.png)
|
375 |
![Depth output image](images/depth2.png)
|
@@ -380,7 +398,7 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
|
|
380 |
The Openpose method extracts the human poses of the sample image. It helps tremendously to get the desired shot and composition of your generated characters. Observe:
|
381 |
|
382 |
<details>
|
383 |
-
<summary>Openpose example
|
384 |
|
385 |
![Open Pose preprocessed image](images/openpose1.png)
|
386 |
![Open Pose output image](images/openpose2.png)
|
@@ -391,7 +409,7 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
|
|
391 |
Lets you make a simple sketch and convert it into a finished piece with the help of your prompt. This is the only example not using the sample image above.
|
392 |
|
393 |
<details>
|
394 |
-
<summary>Scribble example
|
395 |
|
396 |
![Scribble sample image](images/scribble1.jpg)
|
397 |
![Scribble output image](images/scribble2.png)
|
@@ -402,8 +420,8 @@ You will notice that there are 2 results for each method except Scribble. The fi
|
|
402 |
In the Settings tab there is a ControlNet section where you can enable *multiple controlnets at once*. One particularly good use is when one of them is Openpose, to get a specific character pose in a specific environment, or with specific hand gestures or details. Observe:
|
403 |
|
404 |
<details>
|
405 |
-
<summary>Openpose+Canny example
|
406 |
-
|
407 |
![Open Pose + Canny](images/openpose_canny.png)
|
408 |
</details>
|
409 |
|
|
|
178 |
|
179 |
![Parameters](images/parameters.png)
|
180 |
|
181 |
+
* **Sampling method:** This is the algorithm that formulates your image, and each produce different results. The default of `Euler a` is often the best. There are also very good results for `DPM++ 2M Karras` and `DPM++ SDE Karras`. See below for a comparison.
|
182 |
+
* **Sampling steps:** These are "calculated" beforehand, and so more steps doesn't always mean more detail. I always go with 30, you may go from 20-50 and find consistently good results. See below for a comparison.
|
183 |
* **Width and Height:** 512x512 is the default, and you should almost never go above 768 in either direction as it may distort and deform your image. To produce bigger images see `Hires fix`.
|
184 |
* **Batch Count and Batch Size:** Batch *size* is how many images your graphics card will generate at the same time, which is limited by its VRAM. Batch *count* is how many times to repeat those. Batches have consecutive seeds, more on seeds below.
|
185 |
* **CFG Scale:** "Lower values produce more creative results". You should almost always stick to 7, but 4 to 10 is an acceptable range.
|
186 |
* **Seed:** A number that guides the creation of your image. The same seed with the same prompt and parameters produces the same image every time, except for small details and under some circumstances.
|
187 |
|
188 |
**Hires fix:** Lets you create larger images without distortion. Often used at 2x scale. When selected, more options appear:
|
189 |
+
* **Upscaler:** The algorithm to upscale with. `Latent` and its variations produce creative and detailed results, but you may also like `R-ESRGAN 4x+` and its anime version. [More explanation and some comparisons further down ▼](#upscale).
|
190 |
* **Hires steps:** I recommend at least half as many as your sampling steps. Higher values aren't always better, and they take a long time, so be conservative here.
|
191 |
* **Denoising strength:** The most important parameter. Near 0.0, no detail will be added to the image. Near 1.0, the image will be changed completely. I recommend something between 0.2 and 0.6 depending on the image, to add enough detail as the image gets larger, without *destroying* any original details you like.
|
192 |
|
|
|
195 |
* **Tiling:** Used to produce repeating textures to put on a grid. Not very useful.
|
196 |
* **Script:** Lets you access useful features and extensions, such as [X/Y/Z Plot ▼](#plot) which lets you compare images with varying parameters on a grid. Very powerful.
|
197 |
|
198 |
+
Here is a comparison of a few popular samplers and various sampling steps:
|
199 |
+
|
200 |
+
<details>
|
201 |
+
<summary>(Click) Sampler comparison - Photography</summary>
|
202 |
+
|
203 |
+
![samplers with photos](images/samplers1.png)
|
204 |
+
<details>
|
205 |
+
|
206 |
+
<details>
|
207 |
+
<summary>(Click) Sampler comparison - Anime</summary>
|
208 |
+
|
209 |
+
![samplers with anime](images/samplers2.png)
|
210 |
+
<details>
|
211 |
+
|
212 |
+
An explanation of the samplers used above: `Euler` is a basic sampler. `DDIM` is a faster version, while `DPM++ 2M Karras` is an improved version. Meanwhile we have `Euler a` or "ancestral" which produces more creative results, and `DPM++ 2S a Karras` which is also ancestral and thus similar. Finally `DPM++ SDE Karras` is the slowest and quite unique. There are many other samplers not shown here but most of them are related.
|
213 |
+
|
214 |
|
215 |
|
216 |
# Extensions <a name="extensions"></a>[▲](#index)
|
|
|
259 |
Here are some comparisons. All of them were done at 0.4 denoising strength. Note that some of the differences may be completely up to random chance.
|
260 |
|
261 |
<details>
|
262 |
+
<summary>(Click) Comparison 1: Anime, stylized, fantasy</summary>
|
263 |
|
264 |
**Some details to consider:** The fireballs to the left and right, the texture of the fire around her, the grass and its flowers, the ghost's face, the flowers in her hat, the hands, the eyes (which should be flower-shaped), the things on her waist.
|
265 |
|
|
|
271 |
</details>
|
272 |
|
273 |
<details>
|
274 |
+
<summary>(Click) Comparison 2: Anime, detailed, soft lighting</summary>
|
275 |
|
276 |
**Some details to consider:** The background, the flower and symbol on her hat, the flowers on the branches to the sides, the eyes (which should be flower-shaped), the emblem below her neck, The pattern on the lower half of her dress, as well as the nearby frills and folds.
|
277 |
|
|
|
284 |
</details>
|
285 |
|
286 |
<details>
|
287 |
+
<summary>(Click) Comparison 3: Photography, human, nature</summary>
|
288 |
|
289 |
**Some details to consider:** The eye on the left, the finger creases, the bracelet, the edge trim on the vest, the flower on the vest, the brooches on the vest, the rocks and vegetation on the bottom left, the trees on the top left, the waterfalls of course.
|
290 |
|
|
|
313 |
Here I made a comparison between different **models** (columns) and faces of different ethnicities via **S/R Prompt** (rows):
|
314 |
|
315 |
<details>
|
316 |
+
<summary>(Click) X/Y/Z Plot example</summary>
|
317 |
|
318 |
![X Y Z plot of models and ethnicities](images/XYZplot.png)
|
319 |
</details>
|
|
|
329 |
<a name="matrixneg"></a>Here is a comparison using the negative prompts I showed you in [Prompts ▲](#prompt). We can see how EasyNegative affects the image, as well as how the rest of the prompt affects the image, then both together:
|
330 |
|
331 |
<details>
|
332 |
+
<summary>(Click) Prompt matrix examples</summary>
|
333 |
|
334 |
![Prompt matrix of anime negative prompt sections](images/promptmatrix1.png)
|
335 |
![Prompt matrix of photorealistic negative prompt sections](images/promptmatrix2.png)
|
336 |
</details>
|
337 |
|
338 |
+
**Tip:** When using prompt matrix, the Batch Size will let you generate multiple images or the whole grid all at once.
|
339 |
+
|
340 |
* **Ultimate Upscaler** <a name="ultimate"></a>[▲](#index)
|
341 |
|
342 |
An improved version of a builtin script, it can be added as an [extension ▲](#extensions) and used from within **img2img**. Its purpose is to resize an image and add more detail way past the normal limits of your VRAM by splitting it into chunks, although slower. Here are the steps:
|
|
|
376 |
The Canny method extracts the hard edges of the sample image. It is useful for many different types of images, specially where you want to preserve small details and the general look of an image. Observe:
|
377 |
|
378 |
<details>
|
379 |
+
<summary>(Click) Canny example</summary>
|
380 |
|
381 |
![Canny preprocessed image](images/canny1.png)
|
382 |
![Canny output image](images/canny2.png)
|
|
|
387 |
The Depth method extracts the 3D elements of the sample image. It is best suited for complex environments and general composition. Observe:
|
388 |
|
389 |
<details>
|
390 |
+
<summary>(Click) Depth example</summary>
|
391 |
|
392 |
![Depth preprocessed image](images/depth1.png)
|
393 |
![Depth output image](images/depth2.png)
|
|
|
398 |
The Openpose method extracts the human poses of the sample image. It helps tremendously to get the desired shot and composition of your generated characters. Observe:
|
399 |
|
400 |
<details>
|
401 |
+
<summary>(Click) Openpose example</summary>
|
402 |
|
403 |
![Open Pose preprocessed image](images/openpose1.png)
|
404 |
![Open Pose output image](images/openpose2.png)
|
|
|
409 |
Lets you make a simple sketch and convert it into a finished piece with the help of your prompt. This is the only example not using the sample image above.
|
410 |
|
411 |
<details>
|
412 |
+
<summary>(Click) Scribble example</summary>
|
413 |
|
414 |
![Scribble sample image](images/scribble1.jpg)
|
415 |
![Scribble output image](images/scribble2.png)
|
|
|
420 |
In the Settings tab there is a ControlNet section where you can enable *multiple controlnets at once*. One particularly good use is when one of them is Openpose, to get a specific character pose in a specific environment, or with specific hand gestures or details. Observe:
|
421 |
|
422 |
<details>
|
423 |
+
<summary>(Click) Openpose+Canny example</summary>
|
424 |
+
|
425 |
![Open Pose + Canny](images/openpose_canny.png)
|
426 |
</details>
|
427 |
|