Skip to content

Commit 11857a0

Browse files
[Attention] Add ROCM_AITER_MLA_SPARSE to attention backend registry (vllm-project#29103)
Signed-off-by: Matthew Bonanni <[email protected]>
1 parent 8c25f9c commit 11857a0

File tree

2 files changed

+4
-4
lines changed

2 files changed

+4
-4
lines changed

vllm/attention/backends/registry.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -52,6 +52,9 @@ class AttentionBackendEnum(Enum, metaclass=_AttentionBackendEnumMeta):
5252
ROCM_AITER_FA = (
5353
"vllm.v1.attention.backends.rocm_aiter_fa.AiterFlashAttentionBackend"
5454
)
55+
ROCM_AITER_MLA_SPARSE = (
56+
"vllm.v1.attention.backends.mla.rocm_aiter_mla_sparse.ROCMAiterMLASparseBackend"
57+
)
5558
TORCH_SDPA = "" # this tag is only used for ViT
5659
FLASHINFER = "vllm.v1.attention.backends.flashinfer.FlashInferBackend"
5760
FLASHINFER_MLA = (

vllm/platforms/rocm.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -233,10 +233,7 @@ def get_attn_backend_cls(
233233
"Sparse MLA backend on ROCm only supports block size 1 for now."
234234
)
235235
logger.info_once("Using Sparse MLA backend on V1 engine.")
236-
return (
237-
"vllm.v1.attention.backends.mla.rocm_aiter_mla_sparse."
238-
"ROCMAiterMLASparseBackend"
239-
)
236+
return AttentionBackendEnum.ROCM_AITER_MLA_SPARSE.get_path()
240237

241238
if use_mla:
242239
if selected_backend is None:

0 commit comments

Comments
 (0)