cointegrated
commited on
Commit
•
fd5e387
1
Parent(s):
9e6d94e
add the training notebook
Browse files
README.md
CHANGED
@@ -60,7 +60,8 @@ print(text2toxicity(['я люблю нигеров', 'я люблю африка
|
|
60 |
|
61 |
## Training
|
62 |
|
63 |
-
The model has been trained on the joint dataset of [OK ML Cup](https://cups.mail.ru/ru/tasks/1048) and [Babakov et.al.](https://arxiv.org/abs/2103.05345) with `Adam` optimizer, the learning rate of `1e-5`, and batch size of `64` for `15` epochs
|
|
|
64 |
```
|
65 |
non-toxic : 0.9937
|
66 |
insult : 0.9912
|
|
|
60 |
|
61 |
## Training
|
62 |
|
63 |
+
The model has been trained on the joint dataset of [OK ML Cup](https://cups.mail.ru/ru/tasks/1048) and [Babakov et.al.](https://arxiv.org/abs/2103.05345) with `Adam` optimizer, the learning rate of `1e-5`, and batch size of `64` for `15` epochs in [this Colab notebook](https://colab.research.google.com/drive/1o7azO7cHttwofkp8eTZo9LIybYaNWei_?usp=sharing).
|
64 |
+
A text was considered inappropriate if its inappropriateness score was higher than 0.8, and appropriate - if it was lower than 0.2. The per-label ROC AUC on the dev set is:
|
65 |
```
|
66 |
non-toxic : 0.9937
|
67 |
insult : 0.9912
|