File tree Expand file tree Collapse file tree 3 files changed +6
-3
lines changed
examples/models/llama/source_transformation Expand file tree Collapse file tree 3 files changed +6
-3
lines changed Original file line number Diff line number Diff line change 13
13
14
14
import torch
15
15
from torch import nn
16
- from torchao .quantization .GPTQ import Int8DynActInt4WeightLinear
16
+ from torchao .quantization .linear_quant_modules import Int8DynActInt4WeightLinear
17
17
from torchao .quantization .quant_api import _replace_with_custom_fn_if_matches_filter
18
18
19
19
Original file line number Diff line number Diff line change 13
13
import torch
14
14
from torch import nn
15
15
16
- from torchao .quantization .GPTQ import _check_linear_int4_k , Int8DynActInt4WeightLinear
16
+ from torchao .quantization .linear_quant_modules import (
17
+ _check_linear_int4_k ,
18
+ Int8DynActInt4WeightLinear ,
19
+ )
17
20
from torchao .quantization .quant_api import _replace_with_custom_fn_if_matches_filter
18
21
19
22
from .quantize import Int8DynActInt8WeightLinear , QuantizedGroupEmbedding
Original file line number Diff line number Diff line change @@ -847,7 +847,7 @@ def set_8da4w_computation_dtype(
847
847
module : nn .Module , computation_dtype : torch .dtype
848
848
) -> nn .Module :
849
849
850
- from torchao .quantization .GPTQ import Int8DynActInt4WeightLinear
850
+ from torchao .quantization .linear_quant_modules import Int8DynActInt4WeightLinear
851
851
852
852
def _set_8da4w_computation_dtype (module : nn .Module , dtype : torch .dtype ) -> None :
853
853
"""
You can’t perform that action at this time.
0 commit comments