Added additional config options
Browse files- configuration_bert.py +14 -1
configuration_bert.py
CHANGED
@@ -75,8 +75,16 @@ class JinaBertConfig(PretrainedConfig):
|
|
75 |
pad_token_id=0,
|
76 |
window_size=(-1, -1),
|
77 |
dense_seq_output=False,
|
|
|
|
|
|
|
|
|
|
|
|
|
78 |
**kwargs,
|
79 |
):
|
|
|
|
|
80 |
super().__init__(pad_token_id=pad_token_id, **kwargs)
|
81 |
|
82 |
self.vocab_size = vocab_size
|
@@ -92,4 +100,9 @@ class JinaBertConfig(PretrainedConfig):
|
|
92 |
self.layer_norm_eps = layer_norm_eps
|
93 |
self.window_size = window_size
|
94 |
self.dense_seq_output = dense_seq_output
|
95 |
-
|
|
|
|
|
|
|
|
|
|
|
|
75 |
pad_token_id=0,
|
76 |
window_size=(-1, -1),
|
77 |
dense_seq_output=False,
|
78 |
+
fused_mlp=False,
|
79 |
+
mlp_checkpoint_lvl=0,
|
80 |
+
last_layer_subset=False,
|
81 |
+
fused_dropout_add_ln=False,
|
82 |
+
fused_bias_fc=False,
|
83 |
+
pad_vocab_size_multiple=1,
|
84 |
**kwargs,
|
85 |
):
|
86 |
+
assert 'position_embedding_type' not in kwargs
|
87 |
+
assert 'max_position_embeddings' not in kwargs
|
88 |
super().__init__(pad_token_id=pad_token_id, **kwargs)
|
89 |
|
90 |
self.vocab_size = vocab_size
|
|
|
100 |
self.layer_norm_eps = layer_norm_eps
|
101 |
self.window_size = window_size
|
102 |
self.dense_seq_output = dense_seq_output
|
103 |
+
self.fused_mlp = fused_mlp
|
104 |
+
self.mlp_checkpoint_lvl = mlp_checkpoint_lvl
|
105 |
+
self.last_layer_subset = last_layer_subset
|
106 |
+
self.fused_dropout_add_ln = fused_dropout_add_ln
|
107 |
+
self.fused_bias_fc = fused_bias_fc
|
108 |
+
self.pad_vocab_size_multiple = pad_vocab_size_multiple
|