File size: 576 Bytes
28c2541 0c93012 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 |
Base Model: GPT4-x-Alpaca full fine tune by Chavinlo -> https://huggingface.co/chavinlo/gpt4-x-alpaca
LORA fine tune using the Roleplay Instruct from GPT4 generated dataset -> https://github.com/teknium1/GPTeacher/tree/main/Roleplay
LORA Adapter Only: https://huggingface.co/ZeusLabs/gpt4-x-alpaca-rp-lora/tree/main/gpt-rp-instruct-1
Merged LORA to the model.
Instruct it same way as alpaca / gpt4xalpaca:
```
### Instruction:
<prompt>
### Response:
```
or
```
### Instruction:
<prompt>
### Input:
<specific data to manipulate for the instruction
### Response:
``` |