breakcore2
commited on
Commit
•
3caf86c
1
Parent(s):
c15682b
Update locon/notes.md
Browse files- locon/notes.md +21 -0
locon/notes.md
CHANGED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# LoCon Testing
|
2 |
+
https://github.com/KohakuBlueleaf/LoCon
|
3 |
+
|
4 |
+
## Dim/Alpha
|
5 |
+
Testing with
|
6 |
+
conv: 1dim 1alpha (due to coding error the examples are using 4alpha, has since been fixed in the code base)
|
7 |
+
linear(network): 4dim 4alpha
|
8 |
+
We will refer to dim and alpha in the format of dim/alpha, ie 1/1, 4/4
|
9 |
+
|
10 |
+
## Training Time
|
11 |
+
Character appears to gain significant accuracy sooner than using Lora (600 steps vs 800). The time per step is a little bit slower but a better analysis of this is required.
|
12 |
+
We suspect that we can train characters faster in terms of real time.
|
13 |
+
|
14 |
+
## Mixing
|
15 |
+
Applying style on character appears to work well. The example wlop style applied on Amber is 1/0.25 and 4/1. The character loses some stability but the style LoCon can easily be lowered and still retain great style retention while significantly reducing the impact to character correctness.
|
16 |
+
Overall it appears that mixing is more stable with LoCon than lora but this could be due to the dim/alpha and training settings between these two locon being closer compared to LoRA land where popular dim sizes range from 8 to 128.
|
17 |
+
|
18 |
+
## Finetune model extractions
|
19 |
+
To be tested. We suspect the performance should be much better than LoRA extraction since all the model will be extracted from.
|
20 |
+
https://github.com/KohakuBlueleaf/LoCon/blob/main/extract_locon.py
|
21 |
+
|