U
    Mh                     @   sd   d Z ddlmZmZmZmZmZmZmZm	Z	m
Z
mZmZmZmZmZmZmZmZmZmZmZmZ dS )a%  
This file is in the process of migration to `torch/ao/quantization`, and
is kept here for compatibility while the migration process is ongoing.
If you are adding a new entry/functionality, please, add it to the
`torch/ao/quantization/fake_quantize.py`, while adding an import statement
here.
    )_is_fake_quant_script_module_is_per_channel_is_per_tensor_is_symmetric_quantdefault_fake_quant+default_fixed_qparams_range_0to1_fake_quant.default_fixed_qparams_range_neg1to1_fake_quantdefault_fused_act_fake_quant'default_fused_per_channel_wt_fake_quantdefault_fused_wt_fake_quantdefault_histogram_fake_quant%default_per_channel_weight_fake_quantdefault_weight_fake_quantdisable_fake_quantdisable_observerenable_fake_quantenable_observerFakeQuantizeFakeQuantizeBaseFixedQParamsFakeQuantizeFusedMovingAvgObsFakeQuantizeN)__doc__Z#torch.ao.quantization.fake_quantizer   r   r   r   r   r   r   r	   r
   r   r   r   r   r   r   r   r   r   r   r   r    r   r   R/var/www/html/venv/lib/python3.8/site-packages/torch/quantization/fake_quantize.py<module>   s   