|
--- |
|
language: |
|
- nl |
|
license: cc-by-nc-sa-4.0 |
|
tags: |
|
- generated_from_trainer |
|
- simplification |
|
datasets: |
|
- BramVanroy/chatgpt-dutch-simplification |
|
metrics: |
|
- rouge |
|
- sari |
|
task_categories: |
|
- text2text-generation |
|
task_ids: |
|
- text-simplification |
|
widget: |
|
- example_title: Cooking |
|
text: Op bepaalde tijdstippen verlang ik naar de smaakvolle culinaire creaties welke |
|
door de ambachtelijke expertise van mijn grootmoeder zijn vervaardigd. |
|
base_model: yhavinga/ul2-base-dutch |
|
model-index: |
|
- name: BramVanroy/ul2-base-dutch-simplification-mai-2023 |
|
results: |
|
- task: |
|
type: text-simplification |
|
name: Text Simplification |
|
dataset: |
|
name: ChatGPT Dutch Simplification |
|
type: BramVanroy/chatgpt-dutch-simplification |
|
metrics: |
|
- type: rouge |
|
value: 41.5749 |
|
name: Eval Rouge-1 |
|
- type: rouge |
|
value: 19.9 |
|
name: Eval Rouge-2 |
|
- type: rouge |
|
value: 36.3204 |
|
name: Eval RougeL |
|
- type: rouge |
|
value: 36.2596 |
|
name: Eval RougeLsum |
|
- type: sari |
|
value: 53.0091 |
|
name: Eval SARI |
|
- type: rouge |
|
value: 44.2877 |
|
name: Test Rouge-1 |
|
- type: rouge |
|
value: 20.8132 |
|
name: Test Rouge-2 |
|
- type: rouge |
|
value: 39.0951 |
|
name: Test RougeL |
|
- type: rouge |
|
value: 39.2709 |
|
name: Test RougeLsum |
|
- type: sari |
|
value: 52.9621 |
|
name: Test SARI |
|
--- |
|
|
|
|
|
# ul2-base-dutch-simplification-mai-2023 |
|
|
|
This model is intended to simplify Dutch sentences. |
|
|
|
This model is a fine-tuned version of [yhavinga/ul2-base-dutch](https://huggingface.co/yhavinga/ul2-base-dutch) on |
|
the [BramVanroy/chatgpt-dutch-simplification](https://huggingface.co/datasets/BramVanroy/chatgpt-dutch-simplification) |
|
dataset. |
|
|
|
The model was created in light of the master thesis of Charlotte Van de Velde in the Master of Science in Artificial |
|
Intelligence (MAI) at KU Leuven in 2023. Charlotte is supervised by Vincent Vandeghinste and Bram Vanroy. |
|
Dataset creation by Charlotte, model training by Bram. |
|
|
|
## Quick links |
|
|
|
- [Repository](https://github.com/BramVanroy/mai-simplification-nl-2023#22-hyperparameter-sweep): includes training code and model creation log |
|
- [Dataset](https://huggingface.co/datasets/BramVanroy/chatgpt-dutch-simplification): `BramVanroy/chatgpt-dutch-simplification` |
|
- [Parent model](https://huggingface.co/yhavinga/ul2-base-dutch): this model was finetuned on `yhavinga/ul2-base-dutch` |
|
- [Demo](https://huggingface.co/spaces/BramVanroy/mai-simplification-nl-2023-demo): shows the this model in action (don't rely on the "Hosted inference API" widget on this page, it does not work very well) |
|
|
|
## Intended uses & limitations, and dataset |
|
|
|
The model is intended for sentence-level simplification of Dutch. It might extend to document-level simplification |
|
but most of the dataset is limited to sentences so document-level performance is not guaranteed. |
|
|
|
The dataset has been generated automatically (cf. |
|
[dataset description](https://huggingface.co/datasets/BramVanroy/chatgpt-dutch-simplification)) and has not been |
|
manually verified. On top of that, this model has been fine-tuned and we did not scrutinize the parent model or its |
|
training data. Output of the current model is therefore subject to unexpected results (as most if not all neural |
|
networks). |
|
|
|
Because the dataset was generated with ChatGPT, this model cannot be used for commercial purposes. |
|
|
|
## Training procedure |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- learning_rate: 0.00026885245616406115 |
|
- train_batch_size: 12 |
|
- optimizer: Adafactor |
|
- num_epochs: 26 |
|
|
|
These hyperarameters were found through Bayesian hyperparameter search with `wandb`. This is described in the |
|
[repository](https://github.com/BramVanroy/mai-simplification-nl-2023#22-hyperparameter-sweep). |
|
|
|
### Training results |
|
|
|
`eval` results are on the evaluation set, `predict` results are on the test set. These were achieved with |
|
beam search (num_beams=3). |
|
|
|
```json |
|
{ |
|
"eval_gen_len": 21.206349206349206, |
|
"eval_loss": 2.598172903060913, |
|
"eval_rouge1": 41.5749, |
|
"eval_rouge2": 19.9, |
|
"eval_rougeL": 36.3204, |
|
"eval_rougeLsum": 36.2596, |
|
"eval_sari": 53.0091, |
|
|
|
"predict_gen_len": 22.40625, |
|
"predict_loss": 2.517918586730957, |
|
"predict_rouge1": 44.2877, |
|
"predict_rouge2": 20.8132, |
|
"predict_rougeL": 39.0951, |
|
"predict_rougeLsum": 39.2709, |
|
"predict_sari": 52.9621 |
|
} |
|
``` |
|
|
|
|
|
### Framework versions |
|
|
|
- Transformers 4.29.2 |
|
- Pytorch 2.0.1+cu117 |
|
- Datasets 2.12.0 |
|
- Tokenizers 0.13.3 |
|
|