Skip to content

Commit 0d49bbb

Browse files
committed
[Quantization] Remove unused import.
Signed-off-by: menogrey <[email protected]>
1 parent 8913a0c commit 0d49bbb

File tree

1 file changed

+4
-17
lines changed
  • vllm_ascend/quantization/awq

1 file changed

+4
-17
lines changed

vllm_ascend/quantization/awq/awq.py

Lines changed: 4 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -1,38 +1,25 @@
1-
from types import MappingProxyType
2-
from typing import Any, Callable, Dict, List, Mapping, Optional, Union
1+
from typing import Any, Callable, List, Optional, Union
32

43
import torch
5-
from torch.nn.modules import Module
64
import torch_npu
7-
from vllm.config import get_current_vllm_config
8-
from vllm.distributed import get_tensor_model_parallel_rank, get_tp_group
95
from vllm.model_executor.layers.fused_moe import (FusedMoE, FusedMoEMethodBase,
106
FusedMoeWeightScaleSupported)
11-
from vllm.model_executor.layers.fused_moe.config import (FusedMoEConfig, FusedMoEQuantConfig,
12-
int4_w4a16_moe_quant_config,
13-
int8_w8a16_moe_quant_config,)
14-
from vllm.model_executor.layers.linear import (LinearBase, LinearMethodBase,
15-
RowParallelLinear, UnquantizedLinearMethod)
7+
from vllm.model_executor.layers.fused_moe.config import (FusedMoEConfig, FusedMoEQuantConfig)
8+
from vllm.model_executor.layers.linear import (LinearBase, LinearMethodBase)
169
from vllm.model_executor.layers.quantization import \
1710
QUANTIZATION_METHODS, register_quantization_config
1811
from vllm.model_executor.layers.quantization.base_config import (
1912
QuantizationConfig, QuantizeMethodBase)
20-
from vllm.model_executor.layers.quantization.kv_cache import BaseKVCacheMethod
2113
from vllm.model_executor.layers.quantization.utils.quant_utils import is_layer_skipped
2214
from vllm.model_executor.layers.quantization.awq import AWQLinearMethod
23-
from vllm.model_executor.layers.quantization.awq_marlin import AWQMarlinConfig, AWQMoEMethod
24-
from vllm.model_executor.layers.quantization.moe_wna16 import MoeWNA16Method
25-
from vllm.model_executor.layers.vocab_parallel_embedding import (
26-
UnquantizedEmbeddingMethod, VocabParallelEmbedding)
27-
from vllm.model_executor.parameter import PerTensorScaleParameter
2815
from vllm.model_executor.utils import set_weight_attrs
2916

30-
from vllm_ascend.ascend_config import get_ascend_config
3117
from vllm_ascend.ops.linear import AscendUnquantizedLinearMethod
3218
from vllm_ascend.utils import (AWQ_QUANTIZATION_METHOD)
3319
from vllm_ascend.ops.fused_moe.experts_selector import select_experts
3420
from vllm_ascend.ops.fused_moe.fused_moe import AscendUnquantizedFusedMoEMethod
3521

22+
3623
def remove_quantization_method():
3724
if AWQ_QUANTIZATION_METHOD in QUANTIZATION_METHODS:
3825
QUANTIZATION_METHODS.remove(AWQ_QUANTIZATION_METHOD)

0 commit comments

Comments
 (0)