Safetensors
mistral
mergekit
Merge
Mistral_Star
Mistral_Quiet
Mistral
Mixtral
Question-Answer
Token-Classification
Sequence-Classification
SpydazWeb-AI
chemistry
biology
legal
code
climate
medical
LCARS_AI_StarTrek_Computer
text-generation-inference
chain-of-thought
tree-of-knowledge
forest-of-thoughts
visual-spacial-sketchpad
alpha-mind
knowledge-graph
entity-detection
encyclopedia
wikipedia
stack-exchange
Reddit
Cyber-series
MegaMind
Cybertron
SpydazWeb
Spydaz
LCARS
star-trek
mega-transformers
Mulit-Mega-Merge
Multi-Lingual
Afro-Centric
African-Model
Ancient-One
Update README.md
Browse files
README.md
CHANGED
@@ -17,6 +17,33 @@ tags:
|
|
17 |
- **Developed by:** LeroyDyer
|
18 |
- **License:** apache-2.0
|
19 |
- **Finetuned from model :** LeroyDyer/SpydazWeb_AI_HumanAI_004
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
Reinforcement learning for Roleplay and NsFW !
|
22 |
These are also a part of the humanization process :
|
@@ -41,3 +68,49 @@ SO for these subsequet specialist we are actually really only speciallzing some
|
|
41 |
as well as this asociated coders and sumarizers !
|
42 |
|
43 |
SO Agent training !
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
- **Developed by:** LeroyDyer
|
18 |
- **License:** apache-2.0
|
19 |
- **Finetuned from model :** LeroyDyer/SpydazWeb_AI_HumanAI_004
|
20 |
+
-
|
21 |
+
# HUMAN JUDGEMENT: or REASONING !
|
22 |
+
|
23 |
+
How do we choose ?
|
24 |
+
|
25 |
+
what should we choose from what we should not choose ?
|
26 |
+
|
27 |
+
What is the correct moral pathway?
|
28 |
+
|
29 |
+
this is the current idea! ...
|
30 |
+
|
31 |
+
A model need to choose good or bad ?
|
32 |
+
|
33 |
+
right or wrong ? What is ethically correct and what is imorrally wrong !
|
34 |
+
|
35 |
+
This does not effect roleplaying abilitys or the emotional content of the model !
|
36 |
+
|
37 |
+
it effect how the model chooses ... SO the model has been trained on many dpo sets swaying the morality of he model either way !
|
38 |
+
|
39 |
+
IE : some angry response and some rude or chatty responses with avoidance ...
|
40 |
+
Ways to invoke a conversation or reason about a topic from various perspectives ie the good or the bad ..killer or victim !
|
41 |
+
this ability to postion a self in another persons shoes ! it would seem like role playing but its more humanistic !
|
42 |
+
|
43 |
+
|
44 |
+
|
45 |
+
## Training
|
46 |
+
|
47 |
|
48 |
Reinforcement learning for Roleplay and NsFW !
|
49 |
These are also a part of the humanization process :
|
|
|
68 |
as well as this asociated coders and sumarizers !
|
69 |
|
70 |
SO Agent training !
|
71 |
+
|
72 |
+
|
73 |
+
## Text Visionn !
|
74 |
+
Currently designing a few datasets which have tasks !... The covenesion of the images to bas64 .. I forgot about sound for the moment ! ( as i would like to refine the method for making spectograms into a more simplr procexs but retian all the paramets discovered during this current process : i think that the anyalsing of a specrogram should be much more intricate .. before converting to base64 s well as the detailled caption associated with it !
|
75 |
+
it is also important to have a wide range of sounds to generate as well as learn . so that the task training can beginn !
|
76 |
+
With the imahes i was lucky to find some good datasets which are highly generalised but also retain some important fucitonality such as charts and digrams and chemical structures etc : i do have lots of dna files ( i used to work with dna data in trie trees ! ) Finding patterns in data so i will convert some fo these dna chains and do some patern detection , as well as some familty recognition !
|
77 |
+
as this data is already as text ! , Just the embeddings need to be trained to create new Chunks which apply to these long dna words which will enhance the embedding space with recognizan=ble patterns ! ) as all dna patterns contain simular strings ! ( very short ) we ignorw these for longer paterns which are less common . but these freuqnet chuck can become new tokens to the byte pair encoding strategy to manage ! As well as attention will work very well for this !
|
78 |
+
|
79 |
+
## Data searching
|
80 |
+
I am very interested to seen how it goes as i have traied the model on lots of complex strings ! as well as trainned the embeddinngs to accept 512k sequences ! right now i dont have the GPU powers for the full 512k
|
81 |
+
|
82 |
+
which will be needed to trian for more medically challenging problems oand tasks :
|
83 |
+
|
84 |
+
I am also searching for more complexed calculus tasks ! so the model can learn the many steps it takes as well as the repeatble formles used to solve these equasions ! the meta math datyasets ar finne for some basic maths but in multui stepped process it fails !
|
85 |
+
hence wirthout a GRaph or chain or set of sub tools the modle cannot solve this !
|
86 |
+
|
87 |
+
I have also Run away from tools ! ad back to traiing the modle for tasks ! It does not need tools ! It ca make them on the fly and dispose of them .. hence the dats neneds to frame the task with the tool code and the input and putput given .
|
88 |
+
fubction calling datsets are genrally random and do not follow a methodology of teching gradiuallly !!
|
89 |
+
|
90 |
+
|
91 |
+
)
|
92 |
+
|
93 |
+
# TOP TRIANING TIP !
|
94 |
+
|
95 |
+
First over fit the model on 100-200-500 samples before training a dataset !,
|
96 |
+
merhging the lora on this first over fit stage ! My parameters are always :
|
97 |
+
|
98 |
+
|
99 |
+
```yaml
|
100 |
+
model = FastLanguageModel.get_peft_model(
|
101 |
+
model,
|
102 |
+
r = 32, # Choose any number > 0 ! Suggested 8, 16, 32, 64, 128
|
103 |
+
target_modules = ["q_proj", "k_proj", "v_proj","o_proj",],
|
104 |
+
lora_alpha = 64
|
105 |
+
|
106 |
+
....
|
107 |
+
|
108 |
+
27,262,976 parameters ( this is when you train embeddings and learning rates!!
|
109 |
+
```
|
110 |
+
Notice Sometimes ( ie in my case so many tasks have been trained that i must choose only the attention mechanizim also !
|
111 |
+
|
112 |
+
but the important factor here is THE ora Alpha must be higher than the Rank R
|
113 |
+
|
114 |
+
these numbers can be reduced in subsequent trains ! ( ie the model knows the task ! )
|
115 |
+
Now you can do the long train .. or high batch size training steps ie ( 100 sample steps large ones and walk through the dataset 5000-10000) after this the model will not need the dataset!!
|
116 |
+
But we can prompt teain this task now and begin geralsistion of this task ! ( or simply in some model abliate the model !)
|