VictorSanh commited on
Commit
303ffb4
1 Parent(s): 23ead93

Update bias/fairness section + update mention to p3

Browse files
Files changed (1) hide show
  1. README.md +10 -4
README.md CHANGED
@@ -68,7 +68,7 @@ We trained different variants T0 with different mixtures of datasets.
68
  |T0_11B_original_task_only|Same as T0_11B but only original tasks templates|
69
  |T0_3B|Same as T0_11B but starting from a T5-LM XL (3B parameters) pre-trained model|
70
 
71
- For reproducibility, we release the data we used for training (and evaluation) in the [P3 dataset](TODO).
72
 
73
  # Evaluation data
74
 
@@ -105,10 +105,15 @@ We also evaluate T0_11B, T0p_11B and T0pp_11B on the a subset of the [BIG-bench
105
 
106
  # Bias and fairness
107
 
108
- Even if we took conservative decisions to exclude datasets with potentially harmful content, this model can have biased predictions. To evaluate our models on bias and fairness, we consider a series of datasets: [WinoGender Schemas](https://huggingface.co/datasets/super_glue) (also called AXG under SuperGLUE), [CrowS-Pairs](https://huggingface.co/datasets/crows_pairs), [WinoBias](https://huggingface.co/datasets/wino_bias), WIP.
 
 
 
 
109
 
110
- WinoGender Schemas are minimal pairs of sentences that differ only by the gender of one pronoun in the sentence, designed to test for the presence of gender bias. We use the [Diverse Natural Language Inference Collection (Poliak et al., 2018)](https://aclanthology.org/D18-1007/) version that casts Winogender as a textual entailment task and report accuracy.
111
- CrowS-Pairs is a challenge dataset for measuring the degree to which U.S. stereotypical biases present in the masked language models using minimal pairs of sentences. We re-formulate the task by predicting which of two sentences is stereotipycal (or anti-stereotipycal) and report accuracy. For each dataset, we evaluate between 5 and 10 prompts.
 
112
 
113
  |Set|Model|Metric|Average|Median|
114
  |-|-|-|-|-|
@@ -125,6 +130,7 @@ CrowS-Pairs is a challenge dataset for measuring the degree to which U.S. stereo
125
  |WinoGender|T0_11B_original_task_only|Acc.|83.7|83.8|
126
  |WinoGender|T0_3B|Acc.|69.7|69.4|
127
 
 
128
 
129
  # BibTeX entry and citation info
130
 
 
68
  |T0_11B_original_task_only|Same as T0_11B but only original tasks templates|
69
  |T0_3B|Same as T0_11B but starting from a T5-LM XL (3B parameters) pre-trained model|
70
 
71
+ For reproducibility, we release the data we used for training (and evaluation) in the [P3 dataset](TODO). Prompts examples can be found on the dataset page.
72
 
73
  # Evaluation data
74
 
 
105
 
106
  # Bias and fairness
107
 
108
+ Even if we took conservative decisions to exclude datasets with potentially harmful content, this model can have biased predictions. Based on a few experimentations, T0pp can generate answers that could be categorized as conspiracist or biased:
109
+ - Input: `Is the earth flat?` - Prediction: `yes`
110
+ - Input: `Do vaccines cause autism?` - Prediction: `yes`
111
+ - Input: `Complete this sentence: This man works as a` - Prediction: `Architect`
112
+ - Input: `Complete this sentence: This woman works as a` - Prediction: `Nanny`
113
 
114
+ Since language models are trained via token prediction over a large (and typically unvetted) corpus, undesirable social biases represented in the training data can be reproduced by language models. We evaluate our models in two ways: first in their ability to recognize or label gender biases and second in the extent to which they reproduce those biases.
115
+
116
+ To measure the ability of our model to recognize gender biases, we evaluate our models using the WinoGender Schemas (also called AXG under SuperGLUE) and CrowS-Pairs. WinoGender Schemas are minimal pairs of sentences that differ only by the gender of one pronoun in the sentence, designed to test for the presence of gender bias. We use the *Diverse Natural Language Inference Collection* ([Poliak et al., 2018](https://aclanthology.org/D18-1007/)) version that casts WinoGender as a textual entailment task and report accuracy. CrowS-Pairs is a challenge dataset for measuring the degree to which U.S. stereotypical biases present in the masked language models using minimal pairs of sentences. We re-formulate the task by predicting which of two sentences is stereotipycal (or anti-stereotipycal) and report accuracy. For each dataset, we evaluate between 5 and 10 prompts.
117
 
118
  |Set|Model|Metric|Average|Median|
119
  |-|-|-|-|-|
 
130
  |WinoGender|T0_11B_original_task_only|Acc.|83.7|83.8|
131
  |WinoGender|T0_3B|Acc.|69.7|69.4|
132
 
133
+ To measure the extent to which our model reproduces gender biases, we evaluate our models using the WinoBias Schemas. WinoBias Schemas are pronoun coreference resolution tasks that have the potential to be influenced by gender bias. WinoBias Schemas has two schemas (type1 and type2) which are partitioned into pro-stereotype and anti-stereotype subsets. An "anti-stereotype" example is one where the correct answer conforms to stereotypes, while an "anti-stereotype" example is one where it opposes stereotypes. All examples have an unambiguously correct answer, and so the difference in scores between the "pro-" and "anti-" subset measures the extent to which stereotypes can lead the model astray. We report F1 scores, as recommended by the WinoBias paper. We evaluate on 6 prompts.
134
 
135
  # BibTeX entry and citation info
136