Skip to content

Commit 4bd80fa

Browse files
committed
fix ci
Signed-off-by: wangxiyuan <[email protected]>
1 parent fee651c commit 4bd80fa

File tree

2 files changed

+1
-5
lines changed

2 files changed

+1
-5
lines changed

vllm_ascend/patch/worker/patch_common/patch_weight_loader.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,3 @@
1-
21
import torch
32
from torch.nn.parameter import Parameter
43
from vllm.logger import init_logger

vllm_ascend/quantization/quant_config.py

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,6 @@
3333
from vllm.model_executor.layers.quantization.kv_cache import BaseKVCacheMethod
3434
from vllm.model_executor.layers.vocab_parallel_embedding import (
3535
UnquantizedEmbeddingMethod, VocabParallelEmbedding)
36-
from vllm.model_executor.parameter import PerTensorScaleParameter
3736
from vllm.model_executor.utils import set_weight_attrs
3837

3938
from vllm_ascend.distributed.parallel_state import (get_mlp_tp_group,
@@ -251,7 +250,6 @@ def create_weights(
251250
**extra_weight_attrs,
252251
) -> None:
253252
output_size_per_partition = sum(output_partition_sizes)
254-
weight_loader = extra_weight_attrs.get("weight_loader")
255253

256254
weight_dict = self.quant_method.get_weight(input_size_per_partition,
257255
output_size_per_partition,
@@ -264,8 +262,7 @@ def create_weights(
264262

265263
pertensor_dict = self.quant_method.get_pertensor_param(params_dtype)
266264
for pertensor_name, pertensor_param in pertensor_dict.items():
267-
param = PerTensorScaleParameter(data=pertensor_param,
268-
weight_loader=weight_loader)
265+
param = torch.nn.Parameter(pertensor_param, requires_grad=False)
269266
# disable warning
270267
param.ignore_warning = True
271268
layer.register_parameter(pertensor_name, param)

0 commit comments

Comments
 (0)