Skip to content

Commit 1dd4c12

Browse files
zzhx1Kurumi5210
andcommitted
UT fix
Co-authored-by: chenxiao <[email protected]> Signed-off-by: zzhx1 <[email protected]>
1 parent 4612e8f commit 1dd4c12

File tree

1 file changed

+8
-0
lines changed

1 file changed

+8
-0
lines changed

tests/ut/attention/test_mla_v1.py

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -6,6 +6,7 @@
66
from vllm.model_executor.layers.linear import LinearBase
77

88
from tests.ut.base import TestBase
9+
from vllm_ascend.ascend_config import init_ascend_config
910
from vllm_ascend.attention.attention_v1 import AscendAttentionState
1011
from vllm_ascend.attention.mla_v1 import (AscendMLABackend,
1112
AscendMLADecodeMetadata,
@@ -723,6 +724,13 @@ def setUp(self, ascend_config, get_current_vllm_config, mock_get_tp_size,
723724
model_config.dtype = torch.float16
724725
vllm_config.model_config = model_config
725726
get_current_vllm_config.return_value = vllm_config
727+
vllm_config.additional_config = {
728+
"torchair_graph_config": {
729+
"enabled": False,
730+
},
731+
"refresh": True
732+
}
733+
init_ascend_config(vllm_config)
726734

727735
num_heads = 256
728736
head_size = 1024

0 commit comments

Comments
 (0)