BramVanroy commited on
Commit
788cbef
1 Parent(s): 227a3af

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -8
README.md CHANGED
@@ -2,34 +2,41 @@
2
  license: cc-by-nc-4.0
3
  datasets:
4
  - BramVanroy/alpaca-cleaned-dutch
 
 
 
5
  model-index:
6
  - name: falcon-7b-ft-alpaca-cleaned-dutch
7
  results: []
8
  ---
9
 
10
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
- should probably proofread and complete it, then remove this comment. -->
12
 
13
  # falcon-7b-ft-alpaca-cleaned-dutch
14
 
15
- This model is a fine-tuned version of [ybelkada/falcon-7b-sharded-bf16](https://huggingface.co/ybelkada/falcon-7b-sharded-bf16) on the BramVanroy/alpaca-cleaned-dutch dataset.
16
- It achieves the following results on the evaluation set:
17
- - Loss: 1.5448
18
 
19
  ## Model description
20
 
21
- More information needed
 
 
22
 
23
  ## Intended uses & limitations
24
 
25
- More information needed
 
 
26
 
27
  ## Training and evaluation data
28
 
29
- More information needed
 
30
 
31
  ## Training procedure
32
 
 
 
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
 
2
  license: cc-by-nc-4.0
3
  datasets:
4
  - BramVanroy/alpaca-cleaned-dutch
5
+ language:
6
+ - nl
7
+ inference: false
8
  model-index:
9
  - name: falcon-7b-ft-alpaca-cleaned-dutch
10
  results: []
11
  ---
12
 
13
+
 
14
 
15
  # falcon-7b-ft-alpaca-cleaned-dutch
16
 
17
+
 
 
18
 
19
  ## Model description
20
 
21
+ This model is a fine-tuned version of [ybelkada/falcon-7b-sharded-bf16](https://huggingface.co/ybelkada/falcon-7b-sharded-bf16) on the [BramVanroy/alpaca-cleaned-dutch](https://huggingface.co/datasets/BramVanroy/alpaca-cleaned-dutch) dataset.
22
+ See the original [Falcon 7B model](https://huggingface.co/tiiuae/falcon-7b/) for more information, intended use, and biases.
23
+
24
 
25
  ## Intended uses & limitations
26
 
27
+ This model is intended as a baseline for Dutch generative LLMs. It by no means aims to provide SOTA performance and is specifically intended for research purposes.
28
+
29
+ Interestingly, the original Falcon 7B model was only trained on English and French. Therefore, Dutch generations should be taken with a massive grain of salt.
30
 
31
  ## Training and evaluation data
32
 
33
+ Trained on the synthetic [BramVanroy/alpaca-cleaned-dutch](https://huggingface.co/datasets/BramVanroy/alpaca-cleaned-dutch) instruction dataset.
34
+ Therefore, commercial use of this model is forbidden. The model is intended for research purposes only.
35
 
36
  ## Training procedure
37
 
38
+ Trained with LoRA and merged before upload.
39
+
40
  ### Training hyperparameters
41
 
42
  The following hyperparameters were used during training: