File size: 1,697 Bytes
afa4667 4a1f591 afa4667 181703e 4a1f591 afa4667 4a1f591 afa4667 4a1f591 afa4667 6095b57 4a1f591 afa4667 4a1f591 b96f19b 4a1f591 b224f45 4a1f591 b224f45 4a1f591 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 |
---
language:
- en
license: cc-by-nc-4.0
tags:
- text-generation-inference
- transformers
- unsloth
- mistral
base_model: alpindale/Mistral-7B-v0.2-hf
datasets:
- maldv/cyberpunk
- microsoft/orca-math-word-problems-200k
- Weyaxi/sci-datasets
- grimulkan/theory-of-mind
- ResplendentAI/Synthetic_Soul_1k
- GraphWiz/GraphInstruct-RFT-72K
---
# Electric Mist 7B
- **Developed by:** maldv
- **License:** cc-by-nc-4.0
- **Finetuned from model:** alpindale/Mistral-7B-v0.2-hf
- **Methodology:** Simple newline delimited, rolling window book and stripped conversation data.
## Have you learned anything?
Yes, I learned that if you try to train models that aren't the base model, that the results are trash. I have heard rumors that merging the LoRA's works,
which is why the companion LoRA for this is published as well.
## Will It Write
It's good. It goes page after page. It needs an authors note to stay on track though.
## Data
90% sci-fi fiction text data (with a lot of the pulpiest removed), then 10% of the other datasets mixed together; around 6000 8192 context samples, lora r 64, lr .00005, 2 epochs.
## Chat Template
It was trained to follow no prompt at all, just to start going; which means the best results are from when you start the story. There is explicitly no chat in the training data. Simply double newline delimited (even with the orca, math, etc).
## Issues
Punctuation isn't perfect, and has spacing issues, but I have yet to see it collapse even after dumping 40000 tokens in a rolling 8192 context.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="50"/>](https://github.com/unslothai/unsloth) |