tomer-deci
commited on
Commit
•
e790e33
1
Parent(s):
89289b0
Fixed unnecessary flash_attn dependency
Browse files
transformers_v4_35_2__modeling_llama.py
CHANGED
@@ -45,9 +45,12 @@ from .transformers_v4_35_2__configuration_llama import LlamaConfig
|
|
45 |
|
46 |
|
47 |
if is_flash_attn_2_available():
|
48 |
-
|
49 |
-
|
|
|
|
|
50 |
|
|
|
51 |
|
52 |
# This makes `_prepare_4d_causal_attention_mask` a leaf function in the FX graph.
|
53 |
# It means that the function will not be traced through and simply appear as a node in the graph.
|
|
|
45 |
|
46 |
|
47 |
if is_flash_attn_2_available():
|
48 |
+
def import_flash_attn():
|
49 |
+
from flash_attn import flash_attn_func, flash_attn_varlen_func
|
50 |
+
from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
|
51 |
+
return flash_attn_func, flash_attn_varlen_func, index_first_axis, pad_input, unpad_input
|
52 |
|
53 |
+
flash_attn_func, flash_attn_varlen_func, index_first_axis, pad_input, unpad_input = import_flash_attn()
|
54 |
|
55 |
# This makes `_prepare_4d_causal_attention_mask` a leaf function in the FX graph.
|
56 |
# It means that the function will not be traced through and simply appear as a node in the graph.
|