Skip to content

Commit 9830f85

Browse files
authored
[CI] Fix test_mla_v1 (#3570)
### What this PR does / why we need it? Remove test cases containing CPU incompatible operators ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 Signed-off-by: wangli <[email protected]>
1 parent 4a849df commit 9830f85

File tree

1 file changed

+0
-15
lines changed

1 file changed

+0
-15
lines changed

tests/ut/attention/test_mla_v1.py

Lines changed: 0 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -343,21 +343,6 @@ def test_init(self):
343343
self.assertEqual(self.impl.num_queries_per_kv, 32)
344344
self.assertEqual(self.impl.tp_size, 2)
345345

346-
def test_v_up_proj(self):
347-
batch_size = 4
348-
x = torch.randn(batch_size, self.impl.num_heads,
349-
self.impl.kv_lora_rank)
350-
351-
if not hasattr(self.impl, 'W_UV') or self.impl.W_UV is None:
352-
self.impl.W_UV = torch.randn(self.impl.num_heads,
353-
self.impl.kv_lora_rank,
354-
self.impl.v_head_dim)
355-
result = self.impl._v_up_proj(x)
356-
357-
self.assertEqual(result.shape[0], batch_size)
358-
self.assertEqual(result.shape[1],
359-
self.impl.num_heads * self.impl.v_head_dim)
360-
361346
def test_q_proj_and_k_up_proj(self):
362347
batch_size = 4
363348
x = torch.randn(batch_size, self.impl.num_heads, self.impl.qk_head_dim)

0 commit comments

Comments
 (0)