nbeerbower commited on
Commit
e52f7b7
1 Parent(s): 24d1eeb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -1
README.md CHANGED
@@ -9,4 +9,22 @@ datasets:
9
  - nbeerbower/gutenberg-moderne-dpo
10
  base_model:
11
  - nbeerbower/Mahou-1.5-mistral-nemo-12B-lorablated
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  - nbeerbower/gutenberg-moderne-dpo
10
  base_model:
11
  - nbeerbower/Mahou-1.5-mistral-nemo-12B-lorablated
12
+ ---
13
+
14
+ ![image/png](https://huggingface.co/nbeerbower/MN-Moderne-EXPERIMENT/resolve/main/moderne-fft-cover.png?download=true)
15
+
16
+ > 🧪 **Just Another Model Experiment**
17
+ >
18
+ > This is one of many experimental iterations I'm sharing publicly while I mess around with training parameters and ideas. It's not a "real" release - just me being transparent about my learning process. Feel free to look under the hood, but don't expect anything production-ready!
19
+
20
+ # Mistral-Nemo-Moderne-12B-FFT-experimental
21
+
22
+ [Mahou-1.5-mistral-nemo-12B-lorablated](https://huggingface.co/nbeerbower/Mahou-1.5-mistral-nemo-12B-lorablated) finetuned on [gutenberg2-dpo](https://huggingface.co/datasets/nbeerbower/gutenberg2-dpo) and [gutenberg-moderne-dpo](https://huggingface.co/datasets/nbeerbower/gutenberg-moderne-dpo).
23
+
24
+ **This model has erratic behavior and poor performance**
25
+
26
+ ### Method
27
+
28
+ [ORPO tuned](https://mlabonne.github.io/blog/posts/2024-04-19_Fine_tune_Llama_3_with_ORPO.html) with 8x A100 for 1.5 epochs.
29
+
30
+ This was a full finetune. I think the issues with the model can be chalked up to conflicts with Mistral Instruct and ChatML.