Upload HymbaForCausalLM
Browse files- model.safetensors +3 -0
- modeling_hymba.py +8 -0
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83e5b3b0f41d82964e0c22809786ff0eb10afc116d43cbbe53325ebf6cba85f1
|
3 |
+
size 3045665048
|
modeling_hymba.py
CHANGED
@@ -543,6 +543,14 @@ class HymbaAttention(nn.Module):
|
|
543 |
|
544 |
if self.config.rope:
|
545 |
self._init_rope()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
546 |
|
547 |
|
548 |
def _init_rope(self):
|
|
|
543 |
|
544 |
if self.config.rope:
|
545 |
self._init_rope()
|
546 |
+
|
547 |
+
|
548 |
+
def set_rope(self, rope_type, orig_max_position_embeddings, max_position_embeddings):
|
549 |
+
self.config.rope_type = rope_type
|
550 |
+
self.config.orig_max_position_embeddings = orig_max_position_embeddings
|
551 |
+
self.config.max_position_embeddings = max_position_embeddings
|
552 |
+
|
553 |
+
self._init_rope()
|
554 |
|
555 |
|
556 |
def _init_rope(self):
|