File size: 1,027 Bytes
a8a1fd4
 
215f100
 
a8a1fd4
215f100
6aa607d
215f100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
license: mit
language:
- en
---

# BERT-Mini (uncased)
This is one of 24 smaller BERT models (English only, uncased, trained with WordPiece masking)
released by [google-research/bert](https://github.com/google-research/bert).

These BERT models was released as TensorFlow checkpoints, however, this is the converted version to PyTorch.
More information can be found in [google-research/bert](https://github.com/google-research/bert) or [lyeoni/convert-tf-to-pytorch](https://github.com/lyeoni/convert-tf-to-pytorch).

## Evaluation
Here are the evaluation scores (F1/Accuracy) for the MPRC task.
|Model|MRPC|
|-|:-:|
|BERT-Tiny|81.22/68.38|
|BERT-Mini|81.43/69.36|
|BERT-Small|81.41/70.34|
|BERT-Medium|83.33/73.53|
|BERT-Base|85.62/78.19|

### References
```
@article{turc2019,
  title={Well-Read Students Learn Better: On the Importance of Pre-training Compact Models},
  author={Turc, Iulia and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
  journal={arXiv preprint arXiv:1908.08962v2 },
  year={2019}
}
```