Skip to content

Commit 29891f2

Browse files
Update moe_mlp.py
1 parent 3c9e6d4 commit 29891f2

File tree

1 file changed

+2
-3
lines changed

1 file changed

+2
-3
lines changed

vllm_ascend/ops/fused_moe/moe_mlp.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -127,16 +127,14 @@ def quant_apply_mlp(hidden_states: torch.Tensor,
127127
if quantized_hidden_states is not None:
128128
dispose_tensor(quantized_hidden_states)
129129
# act_fn: swiglu
130-
group_diff = torch.diff(group_list)
131-
new_group = torch.cat([group_diff[0].unsqueeze(0), group_diff], dim=0)
132130
hidden_states, swiglu_out_scale = torch_npu.npu_dequant_swiglu_quant(
133131
x=hidden_states,
134132
weight_scale=w1_scale,
135133
activation_scale=pertoken_scale,
136134
bias=None,
137135
quant_scale=None,
138136
quant_offset=None,
139-
group_index=new_group,
137+
group_index=group_list,
140138
activate_left=True,
141139
quant_mode=1,
142140
)
@@ -298,3 +296,4 @@ def unified_apply_mlp(hidden_states: torch.Tensor,
298296
group_list_type=group_list_type,
299297
topk_scales=topk_scales,
300298
need_trans=need_trans)
299+

0 commit comments

Comments
 (0)