
    |j                    v    d dl mZ d dlZd dlmZ d dlmZ ddlmZ erd dl	m
Z
 ddlmZ  G d	 d
e          ZdS )    )annotationsN)TYPE_CHECKING)fleet   )Quantization)Layer)QuantConfigc                  4     e Zd ZdZd fdZd ZdddZ xZS )PTQz;
    Applying post training quantization to the model.
    configr	   returnNonec                J    t                                          |           d S )N)super__init__)selfr   	__class__s     g/lsinfo/ai/hellotax_ai/data_center/backend/venv/lib/python3.11/site-packages/paddle/quantization/ptq.pyr   zPTQ.__init__"   s!             c                \    	 t          j                    dk    rdS dS # t          $ r Y dS w xY w)N   TF)r   
worker_num	Exception)r   s    r   _is_parallel_trainingzPTQ._is_parallel_training%   sG    	!!A%%tu 	 	 	55	s    
++Fmodelr   inplaceboolc                \   |}|sF|                                  r
J d            t          j        |          }|                                 |j        r
J d            | j                            |           |                     || j                   |                     || j                   |S )a9  
        Create a model for post-training quantization.

        The quantization configuration will be propagated in the model.
        And it will insert observers into the model to collect and compute
        quantization parameters.

        Args:
            model(Layer): The model to be quantized.
            inplace(bool): Whether to modify the model in-place.

        Return: The prepared model for post-training quantization.

        Examples:
            .. code-block:: python

                >>> from paddle.quantization import PTQ, QuantConfig
                >>> from paddle.quantization.observers import AbsmaxObserver
                >>> from paddle.vision.models import LeNet

                >>> observer = AbsmaxObserver()
                >>> q_config = QuantConfig(activation=observer, weight=observer)
                >>> ptq = PTQ(q_config)
                >>> model = LeNet()
                >>> model.eval()
                >>> quant_model = ptq.quantize(model)
                >>> print(quant_model)
                LeNet(
                  (features): Sequential(
                    (0): QuantedConv2D(
                      (weight_quanter): AbsmaxObserverLayer()
                      (activation_quanter): AbsmaxObserverLayer()
                    )
                    (1): ObserveWrapper(
                      (_observer): AbsmaxObserverLayer()
                      (_observed): ReLU()
                    )
                    (2): ObserveWrapper(
                      (_observer): AbsmaxObserverLayer()
                      (_observed): MaxPool2D(kernel_size=2, stride=2, padding=0)
                    )
                    (3): QuantedConv2D(
                      (weight_quanter): AbsmaxObserverLayer()
                      (activation_quanter): AbsmaxObserverLayer()
                    )
                    (4): ObserveWrapper(
                      (_observer): AbsmaxObserverLayer()
                      (_observed): ReLU()
                    )
                    (5): ObserveWrapper(
                      (_observer): AbsmaxObserverLayer()
                      (_observed): MaxPool2D(kernel_size=2, stride=2, padding=0)
                    )
                  )
                  (fc): Sequential(
                    (0): QuantedLinear(
                      (weight_quanter): AbsmaxObserverLayer()
                      (activation_quanter): AbsmaxObserverLayer()
                    )
                    (1): QuantedLinear(
                      (weight_quanter): AbsmaxObserverLayer()
                      (activation_quanter): AbsmaxObserverLayer()
                    )
                    (2): QuantedLinear(
                      (weight_quanter): AbsmaxObserverLayer()
                      (activation_quanter): AbsmaxObserverLayer()
                    )
                  )
                )
        z3'inplace' is not compatible with parallel training.zjPost-Training Quantization should not work on training models. Please set evaluation mode by model.eval().)	r   copydeepcopyevaltraining_config_specify_convert_to_quant_layers_insert_activation_observers)r   r   r   _models       r   quantizezPTQ.quantize.   s    N  	1133  E 3 ]5))FKKMMM> 	
 	
x	
 	
! 	f%%%%%fdl;;;))&$,???r   )r   r	   r   r   )F)r   r   r   r   r   r   )__name__
__module____qualname____doc__r   r   r(   __classcell__)r   s   @r   r   r      sx         ! ! ! ! ! !  T T T T T T T T Tr   r   )
__future__r   r   typingr   paddle.distributedr   r(   r   	paddle.nnr   r   r	   r    r   r   <module>r3      s    # " " " " "              $ $ $ $ $ $ " " " " " " $######e e e e e, e e e e er   