30 lines
833 B
Python
30 lines
833 B
Python
|
# flake8: noqa: F401
|
||
|
r"""
|
||
|
Utils shared by different modes of quantization (eager/graph)
|
||
|
|
||
|
This file is in the process of migration to `torch/ao/quantization`, and
|
||
|
is kept here for compatibility while the migration process is ongoing.
|
||
|
If you are adding a new entry/functionality, please, add it to the
|
||
|
`torch/ao/quantization/utils.py`, while adding an import statement
|
||
|
here.
|
||
|
"""
|
||
|
|
||
|
from torch.ao.quantization.utils import (
|
||
|
activation_dtype,
|
||
|
activation_is_int8_quantized,
|
||
|
activation_is_statically_quantized,
|
||
|
calculate_qmin_qmax,
|
||
|
check_min_max_valid,
|
||
|
get_combined_dict,
|
||
|
get_qconfig_dtypes,
|
||
|
get_qparam_dict,
|
||
|
get_quant_type,
|
||
|
get_swapped_custom_module_class,
|
||
|
getattr_from_fqn,
|
||
|
is_per_channel,
|
||
|
is_per_tensor,
|
||
|
weight_dtype,
|
||
|
weight_is_quantized,
|
||
|
weight_is_statically_quantized,
|
||
|
)
|