cartesinus
commited on
Commit
•
7d8eaf0
1
Parent(s):
3410079
Update README.md
Browse files
README.md
CHANGED
@@ -4,6 +4,7 @@ tags:
|
|
4 |
- generated_from_trainer
|
5 |
- nlu
|
6 |
- domain-classificatoin
|
|
|
7 |
metrics:
|
8 |
- accuracy
|
9 |
- f1
|
@@ -25,7 +26,6 @@ datasets:
|
|
25 |
- AmazonScience/massive
|
26 |
language:
|
27 |
- en
|
28 |
-
arxiv: 2310.16609
|
29 |
---
|
30 |
|
31 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
@@ -80,4 +80,16 @@ The following hyperparameters were used during training:
|
|
80 |
- Transformers 4.22.1
|
81 |
- Pytorch 1.12.1+cu113
|
82 |
- Datasets 2.5.1
|
83 |
-
- Tokenizers 0.12.1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
- generated_from_trainer
|
5 |
- nlu
|
6 |
- domain-classificatoin
|
7 |
+
- 'arxiv: 2310.16609'
|
8 |
metrics:
|
9 |
- accuracy
|
10 |
- f1
|
|
|
26 |
- AmazonScience/massive
|
27 |
language:
|
28 |
- en
|
|
|
29 |
---
|
30 |
|
31 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
80 |
- Transformers 4.22.1
|
81 |
- Pytorch 1.12.1+cu113
|
82 |
- Datasets 2.5.1
|
83 |
+
- Tokenizers 0.12.1
|
84 |
+
|
85 |
+
## Citation
|
86 |
+
```bibtex
|
87 |
+
@article{kubis2023back,
|
88 |
+
title={Back Transcription as a Method for Evaluating Robustness of Natural Language Understanding Models to Speech Recognition Errors},
|
89 |
+
author={Kubis, Marek and Sk{\'o}rzewski, Pawe{\l} and Sowa{\'n}ski, Marcin and Zi{\k{e}}tkiewicz, Tomasz},
|
90 |
+
journal={arXiv preprint arXiv:2310.16609},
|
91 |
+
year={2023}
|
92 |
+
eprint={2310.16609},
|
93 |
+
archivePrefix={arXiv},
|
94 |
+
}
|
95 |
+
```
|