fgaim commited on
Commit
80a682e
1 Parent(s): 7534720

Update readme

Browse files
Files changed (1) hide show
  1. README.md +6 -9
README.md CHANGED
@@ -6,19 +6,16 @@ widget:
6
 
7
  # BERT Base for Tigrinya Language
8
 
9
- We pretrain a BERT base-uncased model on a relatively small dataset for Tigrinya (34M tokens) for 40 epochs.
10
-
11
- Contained in this card is a PyTorch model exported from the original model that was trained on a TPU v3.8 with Flax.
12
 
 
13
 
14
  ## Hyperparameters
15
 
16
  The hyperparameters corresponding to model sizes mentioned above are as follows:
17
 
18
- | Model Size | L | AH | HS | FFN | P |
19
- |------------|----|----|-----|------|------|
20
- | BASE | 12 | 12 | 768 | 3072 | 110M |
21
-
22
- (L = number of layers; AH = number of attention heads; HS = hidden size; FFN = feedforward network dimension; P = number of parameters.)
23
-
24
 
 
 
6
 
7
  # BERT Base for Tigrinya Language
8
 
9
+ We pretrain a BERT base-uncased model for Tigrinya on a dataset of 40 million tokens trained for 40 epochs.
 
 
10
 
11
+ Contained in this repo are the original pretrained Flax model that was trained on a TPU v3.8 and it's correponding PyTorch version.
12
 
13
  ## Hyperparameters
14
 
15
  The hyperparameters corresponding to model sizes mentioned above are as follows:
16
 
17
+ | Model Size | L | AH | HS | FFN | P | Seq |
18
+ |------------|----|----|-----|------|------|------|
19
+ | BASE | 12 | 12 | 768 | 3072 | 110M | 128 |
 
 
 
20
 
21
+ (L = number of layers; AH = number of attention heads; HS = hidden size; FFN = feedforward network dimension; P = number of parameters; Seq = maximum sequence length.)