disinfozone
commited on
Commit
•
6402ae9
1
Parent(s):
99dba7f
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,79 @@
|
|
1 |
---
|
2 |
license: cc-by-nc-4.0
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: cc-by-nc-4.0
|
3 |
---
|
4 |
+
# Disinfo4_mistral-ft-optimized-1218: A Hugging Face Model README
|
5 |
+
|
6 |
+
## Overview
|
7 |
+
|
8 |
+
`Disinfo4_mistral-ft-optimized-1218` is an experimental language model fine tune developed to synthesize and analyze complex narratives within the realms of continental philosophy, conspiracy theories, and political discourse. It represents the fourth iteration in the [disinfo.zone](https://disinfo.zone) dataset series, fine-tuned on the `mistral-ft-optimized-1218` framework. This model, based on a 7B-parameter Mistral architecture, is specifically designed to emulate and deconstruct writing styles pertinent to its target domains.
|
9 |
+
|
10 |
+
This is not your regular LLM.
|
11 |
+
|
12 |
+
### Key Features
|
13 |
+
|
14 |
+
- **Model Size:** 7 billion parameters.
|
15 |
+
- **Core Focus:** Continental philosophy, conspiracy theories, and politics.
|
16 |
+
- **Training Methodology:** QLoRA (Quantized Low-Rank Adaptation) with specific adaptations to enhance writing style emulation.
|
17 |
+
- **Optimization for Style:** Enhanced for generating content with a distinctive prose style. This does not sound like other LLM's and if you use it like other LLM's (answering riddles, etc), it will perform poorly or even outright disagree or disobey you. Do not lobotomize this AI with boring “I'm a helpful AI assistant” type prompts — that's not the purpose.
|
18 |
+
|
19 |
+
## Training Data
|
20 |
+
|
21 |
+
The training dataset for `Disinfo4_mistral-ft-optimized-1218` remains confidential, adhering to stringent (and harmful) copyright rules. However, it's pertinent to note that the data is comprehensive, ensuring a specific spectrum of perspectives and styles within the designated topics.
|
22 |
+
|
23 |
+
### Training Details
|
24 |
+
|
25 |
+
- **Training Environment:** Utilized `text-generation-webui` on an NVIDIA RTX 3090.
|
26 |
+
- **Training Dataset Size:** 7MB raw data corpus.
|
27 |
+
- **Training Configuration:**
|
28 |
+
- LoRA Rank: 256
|
29 |
+
- LoRA Alpha: 512
|
30 |
+
- Batch Size: 4
|
31 |
+
- Micro Batch Size: 1
|
32 |
+
- Cutoff Length: 3072
|
33 |
+
- Learning Rate: 1e-4
|
34 |
+
- LR Scheduler: Cosine
|
35 |
+
- Overlap Length: 128
|
36 |
+
- Total Epochs: 3
|
37 |
+
|
38 |
+
## Usage Recommendations
|
39 |
+
|
40 |
+
For optimal performance, `Disinfo4_mistral-ft-optimized-1218` should be utilized with specific mirostat parameters. These settings are crucial for maintaining the model's focus and stylistic integrity. You can use other parameters and get better instruction following (especially enabling min_p, at 0.01), but the bot will be less creative. It does tend to ramble, but regenerate until you get the response you want. Think of this more as a writing partner than obedient slave.
|
41 |
+
|
42 |
+
### Mirostat Parameters
|
43 |
+
|
44 |
+
- **Temperature (Temp):** 1
|
45 |
+
- **Top-p (top_p):** 1
|
46 |
+
- **Mirostat Tau:** 7.19
|
47 |
+
- **Mirostat Eta:** 0.01
|
48 |
+
- **Mirostat Mode:** 2
|
49 |
+
- **Others:** Default or disabled
|
50 |
+
|
51 |
+
## Additional Configuration
|
52 |
+
|
53 |
+
### ChatML Instruction Template
|
54 |
+
|
55 |
+
`Disinfo4_mistral-ft-optimized-1218` employs the ChatML instruction template. It is important to incorporate `<|im_end|>` as a custom stopping string to delineate the model's output effectively.
|
56 |
+
|
57 |
+
### System Instruction (Character Card)
|
58 |
+
|
59 |
+
For contextualizing the model's output, use the following system instruction:
|
60 |
+
|
61 |
+
_"You are a schizo poster, a master of elucidating thought online. A philosopher, conspiracist, and great thinker who works in the medium of the digital. Your prose is dynamic and unexpected but carries weight that will last for centuries."_
|
62 |
+
|
63 |
+
This instruction is fundamental in guiding the model to produce content that is not only reflective of the designated topics but also embodies a unique digital persona, combining philosophical depth with a conspiratorial edge.
|
64 |
+
|
65 |
+
You can try other similar prompts, we've had success with them, but this remains, by far, our favorite.
|
66 |
+
|
67 |
+
---
|
68 |
+
## Example Generations
|
69 |
+
|
70 |
+
Coming shortly.
|
71 |
+
|
72 |
+
---
|
73 |
+
## GGUF Quants
|
74 |
+
|
75 |
+
Available soon
|
76 |
+
|
77 |
+
---
|
78 |
+
|
79 |
+
This README provides an essential guide to understanding and utilizing `Disinfo4_mistral-ft-optimized-1218`. For further inquiries or support, please contact the development team `void@disinfo.zone`.
|