LeroyDyer commited on
Commit
e272686
1 Parent(s): e0123b6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +73 -0
README.md CHANGED
@@ -17,6 +17,33 @@ tags:
17
  - **Developed by:** LeroyDyer
18
  - **License:** apache-2.0
19
  - **Finetuned from model :** LeroyDyer/SpydazWeb_AI_HumanAI_004
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  Reinforcement learning for Roleplay and NsFW !
22
  These are also a part of the humanization process :
@@ -41,3 +68,49 @@ SO for these subsequet specialist we are actually really only speciallzing some
41
  as well as this asociated coders and sumarizers !
42
 
43
  SO Agent training !
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  - **Developed by:** LeroyDyer
18
  - **License:** apache-2.0
19
  - **Finetuned from model :** LeroyDyer/SpydazWeb_AI_HumanAI_004
20
+ -
21
+ # HUMAN JUDGEMENT: or REASONING !
22
+
23
+ How do we choose ?
24
+
25
+ what should we choose from what we should not choose ?
26
+
27
+ What is the correct moral pathway?
28
+
29
+ this is the current idea! ...
30
+
31
+ A model need to choose good or bad ?
32
+
33
+ right or wrong ? What is ethically correct and what is imorrally wrong !
34
+
35
+ This does not effect roleplaying abilitys or the emotional content of the model !
36
+
37
+ it effect how the model chooses ... SO the model has been trained on many dpo sets swaying the morality of he model either way !
38
+
39
+ IE : some angry response and some rude or chatty responses with avoidance ...
40
+ Ways to invoke a conversation or reason about a topic from various perspectives ie the good or the bad ..killer or victim !
41
+ this ability to postion a self in another persons shoes ! it would seem like role playing but its more humanistic !
42
+
43
+
44
+
45
+ ## Training
46
+
47
 
48
  Reinforcement learning for Roleplay and NsFW !
49
  These are also a part of the humanization process :
 
68
  as well as this asociated coders and sumarizers !
69
 
70
  SO Agent training !
71
+
72
+
73
+ ## Text Visionn !
74
+ Currently designing a few datasets which have tasks !... The covenesion of the images to bas64 .. I forgot about sound for the moment ! ( as i would like to refine the method for making spectograms into a more simplr procexs but retian all the paramets discovered during this current process : i think that the anyalsing of a specrogram should be much more intricate .. before converting to base64 s well as the detailled caption associated with it !
75
+ it is also important to have a wide range of sounds to generate as well as learn . so that the task training can beginn !
76
+ With the imahes i was lucky to find some good datasets which are highly generalised but also retain some important fucitonality such as charts and digrams and chemical structures etc : i do have lots of dna files ( i used to work with dna data in trie trees ! ) Finding patterns in data so i will convert some fo these dna chains and do some patern detection , as well as some familty recognition !
77
+ as this data is already as text ! , Just the embeddings need to be trained to create new Chunks which apply to these long dna words which will enhance the embedding space with recognizan=ble patterns ! ) as all dna patterns contain simular strings ! ( very short ) we ignorw these for longer paterns which are less common . but these freuqnet chuck can become new tokens to the byte pair encoding strategy to manage ! As well as attention will work very well for this !
78
+
79
+ ## Data searching
80
+ I am very interested to seen how it goes as i have traied the model on lots of complex strings ! as well as trainned the embeddinngs to accept 512k sequences ! right now i dont have the GPU powers for the full 512k
81
+
82
+ which will be needed to trian for more medically challenging problems oand tasks :
83
+
84
+ I am also searching for more complexed calculus tasks ! so the model can learn the many steps it takes as well as the repeatble formles used to solve these equasions ! the meta math datyasets ar finne for some basic maths but in multui stepped process it fails !
85
+ hence wirthout a GRaph or chain or set of sub tools the modle cannot solve this !
86
+
87
+ I have also Run away from tools ! ad back to traiing the modle for tasks ! It does not need tools ! It ca make them on the fly and dispose of them .. hence the dats neneds to frame the task with the tool code and the input and putput given .
88
+ fubction calling datsets are genrally random and do not follow a methodology of teching gradiuallly !!
89
+
90
+
91
+ )
92
+
93
+ # TOP TRIANING TIP !
94
+
95
+ First over fit the model on 100-200-500 samples before training a dataset !,
96
+ merhging the lora on this first over fit stage ! My parameters are always :
97
+
98
+
99
+ ```yaml
100
+ model = FastLanguageModel.get_peft_model(
101
+ model,
102
+ r = 32, # Choose any number > 0 ! Suggested 8, 16, 32, 64, 128
103
+ target_modules = ["q_proj", "k_proj", "v_proj","o_proj",],
104
+ lora_alpha = 64
105
+
106
+ ....
107
+
108
+ 27,262,976 parameters ( this is when you train embeddings and learning rates!!
109
+ ```
110
+ Notice Sometimes ( ie in my case so many tasks have been trained that i must choose only the attention mechanizim also !
111
+
112
+ but the important factor here is THE ora Alpha must be higher than the Rank R
113
+
114
+ these numbers can be reduced in subsequent trains ! ( ie the model knows the task ! )
115
+ Now you can do the long train .. or high batch size training steps ie ( 100 sample steps large ones and walk through the dataset 5000-10000) after this the model will not need the dataset!!
116
+ But we can prompt teain this task now and begin geralsistion of this task ! ( or simply in some model abliate the model !)