hollowstrawberry commited on
Commit
7c62070
1 Parent(s): 2b21060

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -1
README.md CHANGED
@@ -30,6 +30,7 @@ language:
30
  1. [Generation parameters](#gen)
31
  * [Extensions](#extensions)
32
  * [Loras](#lora)
 
33
  * [Upscaling](#upscale)
34
  * [Scripts](#imgscripts)
35
  * [X/Y/Z Plot](#plot)
@@ -222,6 +223,7 @@ Here you can select your checkpoint and VAE. We will go over what these are and
222
  Here are some useful extensions. If you're using the colab in this guide you already have most of these, otherwise I hugely recommend you manually add the first 2:
223
  * [Image Browser (fixed fork)](https://github.com/aka7774/sd_images_browser) - This will let you browse your past generated images very efficiently, as well as directly sending their prompts and parameters back to txt2img, img2img, etc.
224
  * [TagComplete](https://github.com/DominikDoom/a1111-sd-webui-tagcomplete) - Absolutely essential for anime art. It will show you the matching booru tags as you type. Anime models work via booru tags, and prompts without them usually don't work, so knowing them is godmode. Not all tags will work well in all models though, specially if they're rare.
 
225
  * [ControlNet](https://github.com/Mikubill/sd-webui-controlnet) - A huge extension deserving of [its own guide ▼](#controlnet). It lets you analyze any image and use it as an referene for your own image. Practically speaking, it can create any pose or environment you want.
226
  * [Ultimate Upscale](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111) - A script usable from the img2img section to make really large images, where normally you can only go as high as your VRAM allows. See [Ultimate Upscaler ▼](#ultimate).
227
  * [Two-shot](https://github.com/opparco/stable-diffusion-webui-two-shot) - Normally you can't create more than one distinct character in the same image without them blending together. This extension lets you divide the image into parts; full, left side, right side; allowing you to make nice 2-character images.
@@ -242,6 +244,16 @@ Place your Lora files in the `stable-diffusion-webui/models/Lora` folder, or if
242
 
243
  An example of a Lora is [Thicker Lines Anime Style](https://civitai.com/models/13910/thicker-lines-anime-style-lora-mix), which is perfect if you want your images to look more like traditional anime.
244
 
 
 
 
 
 
 
 
 
 
 
245
   
246
 
247
  # Upscaling <a name="upscale"></a>[▲](#index)
@@ -443,7 +455,7 @@ With those way smarter resources out of the way, I'll try to produce a simple gu
443
 
444
  2. **Trainer Colab** <a name="traincolab"></a>[▲](#index) ![Trainer colab](images/trainercollab.png)
445
 
446
- We will be using [MY NEW TRAINER COLAB](https://colab.research.google.com/drive/1fs7oHytA4Va0IzDK-F8q_q9RIJRIh5Dv?usp=sharing). Here's what settings you should use under each section:
447
 
448
  * **▶️Setup**
449
 
@@ -463,6 +475,10 @@ With those way smarter resources out of the way, I'll try to produce a simple gu
463
 
464
  The `unet_lr` or "learning rate" is the most important parameter. 5e-4 unet is the default in this guide and good for characters, but you can try 1e-3 if you have very few images, or 1e-4/1e-5 if you have lots of images and want slower cooking. You can ignore the rest of the settings this time.
465
 
 
 
 
 
466
  * **▶️Ready**
467
 
468
  You can now press the play button that's been floating on the left side. It will set up the training environment then start the process. Check the progress via the outputs at the bottom, it should take 20 to 60 minutes. Good luck! If you encounter an error, read it carefully in case it's easy to fix, otherwise seek help online or contact me.
 
30
  1. [Generation parameters](#gen)
31
  * [Extensions](#extensions)
32
  * [Loras](#lora)
33
+ * [Lycoris](#lycoris)
34
  * [Upscaling](#upscale)
35
  * [Scripts](#imgscripts)
36
  * [X/Y/Z Plot](#plot)
 
223
  Here are some useful extensions. If you're using the colab in this guide you already have most of these, otherwise I hugely recommend you manually add the first 2:
224
  * [Image Browser (fixed fork)](https://github.com/aka7774/sd_images_browser) - This will let you browse your past generated images very efficiently, as well as directly sending their prompts and parameters back to txt2img, img2img, etc.
225
  * [TagComplete](https://github.com/DominikDoom/a1111-sd-webui-tagcomplete) - Absolutely essential for anime art. It will show you the matching booru tags as you type. Anime models work via booru tags, and prompts without them usually don't work, so knowing them is godmode. Not all tags will work well in all models though, specially if they're rare.
226
+ * [Locon](https://github.com/KohakuBlueleaf/a1111-sd-webui-locon) - Lets you use LoCons and LoHas. More info [below ▼](#lycoris).
227
  * [ControlNet](https://github.com/Mikubill/sd-webui-controlnet) - A huge extension deserving of [its own guide ▼](#controlnet). It lets you analyze any image and use it as an referene for your own image. Practically speaking, it can create any pose or environment you want.
228
  * [Ultimate Upscale](https://github.com/Coyote-A/ultimate-upscale-for-automatic1111) - A script usable from the img2img section to make really large images, where normally you can only go as high as your VRAM allows. See [Ultimate Upscaler ▼](#ultimate).
229
  * [Two-shot](https://github.com/opparco/stable-diffusion-webui-two-shot) - Normally you can't create more than one distinct character in the same image without them blending together. This extension lets you divide the image into parts; full, left side, right side; allowing you to make nice 2-character images.
 
244
 
245
  An example of a Lora is [Thicker Lines Anime Style](https://civitai.com/models/13910/thicker-lines-anime-style-lora-mix), which is perfect if you want your images to look more like traditional anime.
246
 
247
+ * Lycoris <a name="lycoris"></a>[▲](#index)
248
+
249
+ LyCORIS is a new development that lets LoRAs learn more layers. [Learn about it here](https://github.com/KohakuBlueleaf/Lycoris). You'll need [this extension](https://github.com/KohakuBlueleaf/a1111-sd-webui-locon) to use them.
250
+
251
+ As of now there are 2 new LoRA types:
252
+ * **LoCon** - Said to be good with styles
253
+ * **LoHa** - Said to be good with styles that also contain characters
254
+
255
+ You can make your own in the [trainer further down ▼](#traincolab).
256
+
257
  &nbsp;
258
 
259
  # Upscaling <a name="upscale"></a>[▲](#index)
 
455
 
456
  2. **Trainer Colab** <a name="traincolab"></a>[▲](#index) ![Trainer colab](images/trainercollab.png)
457
 
458
+ We will be using [MY NEW TRAINER COLAB](https://colab.research.google.com/drive/1fs7oHytA4Va0IzDK-F8q_q9RIJRIh5Dv?usp=sharing). It has been revamped to be easier to use and have more options, as well as [LoCon and LoHa ▲](#lycoris) support. Here's what settings you should use under each section:
459
 
460
  * **▶️Setup**
461
 
 
475
 
476
  The `unet_lr` or "learning rate" is the most important parameter. 5e-4 unet is the default in this guide and good for characters, but you can try 1e-3 if you have very few images, or 1e-4/1e-5 if you have lots of images and want slower cooking. You can ignore the rest of the settings this time.
477
 
478
+ * **▶️Lora Type**
479
+
480
+ Don't touch this if you don't know what it is.
481
+
482
  * **▶️Ready**
483
 
484
  You can now press the play button that's been floating on the left side. It will set up the training environment then start the process. Check the progress via the outputs at the bottom, it should take 20 to 60 minutes. Good luck! If you encounter an error, read it carefully in case it's easy to fix, otherwise seek help online or contact me.