Skip to content

Commit

Permalink
fix bugs
Browse files Browse the repository at this point in the history
  • Loading branch information
wenhuach21 committed Jun 4, 2024
1 parent 625f5e9 commit f537e06
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 3 deletions.
2 changes: 1 addition & 1 deletion auto_round/autoround.py
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
from torch import autocast

from .calib_dataset import get_dataloader
from .quantizer import WrapperMultiblock, wrapper_block, unwrapper_block
from .quantizer import WrapperMultiblock, wrapper_block, unwrapper_block, WrapperLinear, unwrapper_layer
from .special_model_handler import check_hidden_state_dim, check_share_attention_mask
from .utils import (
CpuInfo,
Expand Down
3 changes: 1 addition & 2 deletions auto_round/quantizer.py
Original file line number Diff line number Diff line change
Expand Up @@ -135,8 +135,7 @@ def quant_weight_actor(weight, num_bits, sym, v, min_scale, max_scale, scale_dty


def quant_weight(
weight, num_bits=4, group_size=-1, sym=False, v=0, min_scale=1.0, max_scale=1.0, scale_dtype=torch.float16,
**kwargs
weight, num_bits=4, group_size=-1, sym=False, v=0, min_scale=1.0, max_scale=1.0, scale_dtype=torch.float16
):
"""Quantizes and dequantizes weight, handing the group size issue .
Expand Down

0 comments on commit f537e06

Please sign in to comment.