Upload configuration_mplug_owl2.py with huggingface_hub
Browse files
configuration_mplug_owl2.py
CHANGED
@@ -117,6 +117,7 @@ class LlamaConfig(PretrainedConfig):
|
|
117 |
rope_theta=10000.0,
|
118 |
rope_scaling=None,
|
119 |
attention_bias=False,
|
|
|
120 |
**kwargs,
|
121 |
):
|
122 |
self.vocab_size = vocab_size
|
@@ -140,6 +141,8 @@ class LlamaConfig(PretrainedConfig):
|
|
140 |
self.rope_scaling = rope_scaling
|
141 |
self._rope_scaling_validation()
|
142 |
self.attention_bias = attention_bias
|
|
|
|
|
143 |
|
144 |
super().__init__(
|
145 |
pad_token_id=pad_token_id,
|
|
|
117 |
rope_theta=10000.0,
|
118 |
rope_scaling=None,
|
119 |
attention_bias=False,
|
120 |
+
attention_dropout=0.0,
|
121 |
**kwargs,
|
122 |
):
|
123 |
self.vocab_size = vocab_size
|
|
|
141 |
self.rope_scaling = rope_scaling
|
142 |
self._rope_scaling_validation()
|
143 |
self.attention_bias = attention_bias
|
144 |
+
self.attention_dropout = attention_dropout
|
145 |
+
self._attn_implementation = "flash_attention_2"
|
146 |
|
147 |
super().__init__(
|
148 |
pad_token_id=pad_token_id,
|