OpenChatRWKV 430m r2
This is a finetune of RWKV-v4neo 430m on openchatgpt safe r2
dataset. r2
shares no data with the r1
, even the greetings, making this finetune, in some ways, inferiour to the original on r1
.
Key differences with openchatrwkv-430m
One of the key differences is using an actual token for the instant message separation, apart from the new dataset.
Differences with openchatgpt-neox-125m
Increased parameter size and different dataset.
Training data
New dataset was obviously made at a later point in time. Many speculate that ChatGPT has degraded over the months, and I am a strong believer of that aswell - style in which model speaks started to sound different compared to 2-3 months ago.
This model was trained on a mix of natural language and code.
This model does not know how to greet you.