hollowstrawberry commited on
Commit
164bf0e
1 Parent(s): bf8fdf4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -7
README.md CHANGED
@@ -263,13 +263,33 @@ Scripts can be found at the bottom of your generation parameters in txt2img or i
263
  ![Prompt matrix of anime negative prompt sections](images/promptmatrix1.png)
264
  ![Prompt matrix of photorealistic negative prompt sections](images/promptmatrix2.png)
265
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
266
   
267
 
268
  # ControlNet <a name="controlnet"></a>[▲](#index)
269
 
270
  ControlNet is an extremely powerful recent technology for Stable Diffusion. It lets you analyze information about any previously existing image and use it to guide the generation of your AI images. We'll see what this means in a moment.
271
 
272
- If you're on collab, you should enable the `all_control_models` option. On Windows, you should first install the ControlNet [extension ▲](#extensions), then go [here](https://civitai.com/models/9251/controlnet-pre-trained-models) to download some models which you'll need to place in `stable-diffusion-webui/extensions/sd-webui-controlnet/models`. I recommend at least Canny, Depth, Openpose and Scribble, which I will show here.
273
 
274
  I will demonstrate how ControlNet may be used. For this I chose a popular image online as our "sample image". It's not necessary for you to follow along, but you can download the images and put them in the **PNG Info** tab to view their generation data.
275
 
@@ -282,7 +302,7 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
282
  The Canny method extracts the hard edges of the sample image. It is useful for many different types of images, specially where you want to preserve small details and the general look of an image. Observe:
283
 
284
  <details>
285
- <summary>Canny example, click to open</summary>
286
 
287
  ![Canny preprocessed image](images/canny1.png)
288
  ![Canny output image](images/canny2.png)
@@ -293,7 +313,7 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
293
  The Depth method extracts the 3D elements of the sample image. It is best suited for complex environments and general composition. Observe:
294
 
295
  <details>
296
- <summary>Depth example, click to open</summary>
297
 
298
  ![Depth preprocessed image](images/depth1.png)
299
  ![Depth output image](images/depth2.png)
@@ -304,7 +324,7 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
304
  The Openpose method extracts the human poses of the sample image. It helps tremendously to get the desired shot and composition of your generated characters. Observe:
305
 
306
  <details>
307
- <summary>Openpose example, click to open</summary>
308
 
309
  ![Open Pose preprocessed image](images/openpose1.png)
310
  ![Open Pose output image](images/openpose2.png)
@@ -312,10 +332,10 @@ First, you must scroll down in the txt2img page and click on ControlNet to open
312
 
313
  * **Scribble**
314
 
315
- Lets you make a simple sketch and convert it into a finished piece with the help of your prompt. For this one I asked a friend:
316
 
317
  <details>
318
- <summary>Scribble example, click to open</summary>
319
 
320
  ![Scribble sample image](images/scribble1.png)
321
  ![Scribble output image](images/scribble2.png)
@@ -335,6 +355,6 @@ There are also alternative **diff** versions of each ControlNet model, which pro
335
 
336
  # Lora Training <a name="train"></a>[▲](#index)
337
 
338
- * **Character Loras** <a name="trainchars"></abe[▲](#index)
339
 
340
  Coming soon.
 
263
  ![Prompt matrix of anime negative prompt sections](images/promptmatrix1.png)
264
  ![Prompt matrix of photorealistic negative prompt sections](images/promptmatrix2.png)
265
 
266
+ * **Ultimate Upscale** <a name="ultimate"></a>[▲](#index)
267
+
268
+ An improved version of a builtin script, it can be added as an [extension ▲] and used from within **img2img**. Its purpose is to resize an image and add more detail way past the normal limits of your VRAM by splitting it into chunks, although slower. Here are the steps:
269
+
270
+ 1. Generate your image normally up to 768 width and height, you can then apply hires fix if you are able to.
271
+
272
+ 1. From txt2img or the Image Browser extension send it directly into img2img, along with its prompt and parameters.
273
+
274
+ 1. Set the **Denoising** somewhere between 0.1 and 0.4. If you go higher you most likely will experience mutations.
275
+
276
+ 1. Go down to **Scripts** and choose **Ultimate SD Upscale**. Then, set your parameters like this, with your desired size and upscaler, and the **"Chess" Type**:
277
+
278
+ ![Ultimate upscale parameters](images/ultimate.png)
279
+
280
+ * If you have enough VRAM, you may increase the **Tile width** as well as the **Padding**. For example, doubling both of them. **Tile height** can remain at 0 and it'll match the width.
281
+
282
+ * It is not necessary to set the **Seams fix** unless you encounter visible seams between regions in the final image.
283
+
284
+ 1. Generate your image and wait. You can watch the squares get sharper if you have image previews enabled.
285
+
286
  &nbsp;
287
 
288
  # ControlNet <a name="controlnet"></a>[▲](#index)
289
 
290
  ControlNet is an extremely powerful recent technology for Stable Diffusion. It lets you analyze information about any previously existing image and use it to guide the generation of your AI images. We'll see what this means in a moment.
291
 
292
+ If you're using the collab in this guide, you should enable the `all_control_models` option. Otherwise, you should first install the ControlNet [extension ▲](#extensions), then go [here](https://civitai.com/models/9251/controlnet-pre-trained-models) to download some models which you'll need to place in `stable-diffusion-webui/extensions/sd-webui-controlnet/models`. I recommend at least Canny, Depth, Openpose and Scribble, which I will show here.
293
 
294
  I will demonstrate how ControlNet may be used. For this I chose a popular image online as our "sample image". It's not necessary for you to follow along, but you can download the images and put them in the **PNG Info** tab to view their generation data.
295
 
 
302
  The Canny method extracts the hard edges of the sample image. It is useful for many different types of images, specially where you want to preserve small details and the general look of an image. Observe:
303
 
304
  <details>
305
+ <summary>Canny example, click to expand</summary>
306
 
307
  ![Canny preprocessed image](images/canny1.png)
308
  ![Canny output image](images/canny2.png)
 
313
  The Depth method extracts the 3D elements of the sample image. It is best suited for complex environments and general composition. Observe:
314
 
315
  <details>
316
+ <summary>Depth example, click to expand</summary>
317
 
318
  ![Depth preprocessed image](images/depth1.png)
319
  ![Depth output image](images/depth2.png)
 
324
  The Openpose method extracts the human poses of the sample image. It helps tremendously to get the desired shot and composition of your generated characters. Observe:
325
 
326
  <details>
327
+ <summary>Openpose example, click to expand</summary>
328
 
329
  ![Open Pose preprocessed image](images/openpose1.png)
330
  ![Open Pose output image](images/openpose2.png)
 
332
 
333
  * **Scribble**
334
 
335
+ Lets you make a simple sketch and convert it into a finished piece with the help of your prompt. This is the only example not using the sample image above.
336
 
337
  <details>
338
+ <summary>Scribble example, click to expand</summary>
339
 
340
  ![Scribble sample image](images/scribble1.png)
341
  ![Scribble output image](images/scribble2.png)
 
355
 
356
  # Lora Training <a name="train"></a>[▲](#index)
357
 
358
+ * **Character Loras** <a name="trainchars"></a>[▲](#index)
359
 
360
  Coming soon.