Upload HymbaForCausalLM
Browse files- model.safetensors +3 -0
- modeling_hymba.py +8 -0
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fdff92c3753ca4b58702b3bf835c36f5210a919e9bfa2383f35dc6ab79e792eb
|
| 3 |
+
size 3045665048
|
modeling_hymba.py
CHANGED
|
@@ -543,6 +543,14 @@ class HymbaAttention(nn.Module):
|
|
| 543 |
|
| 544 |
if self.config.rope:
|
| 545 |
self._init_rope()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 546 |
|
| 547 |
|
| 548 |
def _init_rope(self):
|
|
|
|
| 543 |
|
| 544 |
if self.config.rope:
|
| 545 |
self._init_rope()
|
| 546 |
+
|
| 547 |
+
|
| 548 |
+
def set_rope(self, rope_type, orig_max_position_embeddings, max_position_embeddings):
|
| 549 |
+
self.config.rope_type = rope_type
|
| 550 |
+
self.config.orig_max_position_embeddings = orig_max_position_embeddings
|
| 551 |
+
self.config.max_position_embeddings = max_position_embeddings
|
| 552 |
+
|
| 553 |
+
self._init_rope()
|
| 554 |
|
| 555 |
|
| 556 |
def _init_rope(self):
|