o
    ={c?                     @   s   d Z ddlm  mZ ddlmZ ddlmZ ddl	m
Z
 ddlmZ ddlmZ ddlmZ dd	lmZ d
ZdZe Zedd							d!ddZ	d"ddZedd#ddZedd$ddZejjdejejd e_ ejj e_ dS )%zInception V3 model for Keras.

Reference:
  - [Rethinking the Inception Architecture for Computer Vision](
      http://arxiv.org/abs/1512.00567) (CVPR 2016)
    N)backend)imagenet_utils)training)VersionAwareLayers)
data_utils)layer_utils)keras_exportz|https://storage.googleapis.com/tensorflow/keras-applications/inception_v3/inception_v3_weights_tf_dim_ordering_tf_kernels.h5zhttps://storage.googleapis.com/tensorflow/keras-applications/inception_v3/inception_v3_weights_tf_dim_ordering_tf_kernels_notop.h5z+keras.applications.inception_v3.InceptionV3zkeras.applications.InceptionV3Timagenet  softmaxc                 C   s  |dv st jj|std| |dkr#| r#|dkr#td| tj|ddt | |d}|d	u r;t	j
|d
}nt|sHt	j
||d}n|}t dkrSd}nd}t|dddddd}	t|	ddddd}	t|	ddd}	t	jddd|	}	t|	ddddd}	t|	ddddd}	t	jddd|	}	t|	ddd}
t|	ddd}t|ddd}t|	ddd}t|ddd}t|ddd}t	jdddd|	}t|ddd}t	j|
|||g|dd}	t|	ddd}
t|	ddd}t|ddd}t|	ddd}t|ddd}t|ddd}t	jdddd|	}t|ddd}t	j|
|||g|d d}	t|	ddd}
t|	ddd}t|ddd}t|	ddd}t|ddd}t|ddd}t	jdddd|	}t|ddd}t	j|
|||g|d!d}	t|	d"ddddd}t|	ddd}t|ddd}t|dddddd}t	jddd|	}t	j|||g|d#d}	t|	ddd}
t|	d$dd}t|d$dd%}t|dd%d}t|	d$dd}t|d$d%d}t|d$dd%}t|d$d%d}t|ddd%}t	jdddd|	}t|ddd}t	j|
|||g|d&d}	td'D ]e}t|	ddd}
t|	d(dd}t|d(dd%}t|dd%d}t|	d(dd}t|d(d%d}t|d(dd%}t|d(d%d}t|ddd%}t	jdddd|	}t|ddd}t	j|
|||g|d)td|  d}	qt|	ddd}
t|	ddd}t|ddd%}t|dd%d}t|	ddd}t|dd%d}t|ddd%}t|dd%d}t|ddd%}t	jdddd|	}t|ddd}t	j|
|||g|d*d}	t|	ddd}t|d+ddddd}t|	ddd}t|ddd%}t|dd%d}t|dddddd}t	jddd|	}t	j|||g|d,d}	td'D ]u}t|	d+dd}
t|	d"dd}t|d"dd}t|d"dd}t	j||g|d-t| d}t|	d.dd}t|d"dd}t|d"dd}t|d"dd}t	j||g|d/}t	jdddd|	}t|ddd}t	j|
|||g|d)td0|  d}	q| rt	jd1d2|	}	t|| t	j||d3d4|	}	n|d5krt	 |	}	n|d6krt	 |	}	|d	urt|}n|}tj||	d7d2}|dkr| rtjd8td9d:d;}n	tjd<td9d=d;}|| |S |d	ur|| |S )>a  Instantiates the Inception v3 architecture.

    Reference:
    - [Rethinking the Inception Architecture for Computer Vision](
        http://arxiv.org/abs/1512.00567) (CVPR 2016)

    This function returns a Keras image classification model,
    optionally loaded with weights pre-trained on ImageNet.

    For image classification use cases, see
    [this page for detailed examples](
      https://keras.io/api/applications/#usage-examples-for-image-classification-models).

    For transfer learning use cases, make sure to read the
    [guide to transfer learning & fine-tuning](
      https://keras.io/guides/transfer_learning/).

    Note: each Keras Application expects a specific kind of input preprocessing.
    For `InceptionV3`, call
    `tf.keras.applications.inception_v3.preprocess_input` on your inputs before
    passing them to the model. `inception_v3.preprocess_input` will scale input
    pixels between -1 and 1.

    Args:
      include_top: Boolean, whether to include the fully-connected
        layer at the top, as the last layer of the network. Default to `True`.
      weights: One of `None` (random initialization),
        `imagenet` (pre-training on ImageNet),
        or the path to the weights file to be loaded. Default to `imagenet`.
      input_tensor: Optional Keras tensor (i.e. output of `layers.Input()`)
        to use as image input for the model. `input_tensor` is useful for
        sharing inputs between multiple different networks. Default to None.
      input_shape: Optional shape tuple, only to be specified
        if `include_top` is False (otherwise the input shape
        has to be `(299, 299, 3)` (with `channels_last` data format)
        or `(3, 299, 299)` (with `channels_first` data format).
        It should have exactly 3 inputs channels,
        and width and height should be no smaller than 75.
        E.g. `(150, 150, 3)` would be one valid value.
        `input_shape` will be ignored if the `input_tensor` is provided.
      pooling: Optional pooling mode for feature extraction
        when `include_top` is `False`.
        - `None` (default) means that the output of the model will be
            the 4D tensor output of the last convolutional block.
        - `avg` means that global average pooling
            will be applied to the output of the
            last convolutional block, and thus
            the output of the model will be a 2D tensor.
        - `max` means that global max pooling will be applied.
      classes: optional number of classes to classify images
        into, only to be specified if `include_top` is True, and
        if no `weights` argument is specified. Default to 1000.
      classifier_activation: A `str` or callable. The activation function to use
        on the "top" layer. Ignored unless `include_top=True`. Set
        `classifier_activation=None` to return the logits of the "top" layer.
        When loading pretrained weights, `classifier_activation` can only
        be `None` or `"softmax"`.

    Returns:
      A `keras.Model` instance.
    >   r	   NzThe `weights` argument should be either `None` (random initialization), `imagenet` (pre-training on ImageNet), or the path to the weights file to be loaded; Received: weights=r	   r
   zjIf using `weights` as `"imagenet"` with `include_top` as true, `classes` should be 1000; Received classes=i+  K   )Zdefault_sizeZmin_sizedata_formatZrequire_flattenweightsN)shape)Ztensorr   channels_first          )   r   Zvalid)stridespadding)r   @   )r   r   )r   P      0      `   r   r   sameZmixed0)axisnameZmixed1Zmixed2i  Zmixed3      Zmixed4r      ZmixedZmixed7i@  Zmixed8Zmixed9_i  )r   	   Zavg_poolr    Zpredictions)Z
activationr    ZavgmaxZinception_v3z2inception_v3_weights_tf_dim_ordering_tf_kernels.h5ZmodelsZ 9a0d58056eeedaa3f26cb7ebd46da564)Zcache_subdirZ	file_hashz8inception_v3_weights_tf_dim_ordering_tf_kernels_notop.h5Z bcbd6486424b2319ff4ef7d526e38f63)tfioZgfileexists
ValueErrorr   Zobtain_input_shaper   image_data_formatlayersZInputZis_keras_tensor	conv2d_bnZMaxPooling2DZAveragePooling2DZconcatenaterangestrZGlobalAveragePooling2DZvalidate_activationZDenseZGlobalMaxPooling2Dr   Zget_source_inputsr   ZModelr   Zget_fileWEIGHTS_PATHWEIGHTS_PATH_NO_TOPZload_weights)Zinclude_topr   Zinput_tensorZinput_shapeZpoolingZclassesZclassifier_activationZ	img_inputZchannel_axisxZ	branch1x1Z	branch5x5Zbranch3x3dblZbranch_poolZ	branch3x3Z	branch7x7Zbranch7x7dbliZbranch7x7x3Zbranch3x3_1Zbranch3x3_2Zbranch3x3dbl_1Zbranch3x3dbl_2ZinputsZmodelZweights_path r4   ?lib/python3.10/site-packages/keras/applications/inception_v3.pyInceptionV3/   s  J	

















r6   r   r   c           
      C   s   |dur|d }|d }nd}d}t  dkrd}	nd}	tj|||f||d|d| } tj|	d|d	| } tjd
|d| } | S )a  Utility function to apply conv + BN.

    Args:
      x: input tensor.
      filters: filters in `Conv2D`.
      num_row: height of the convolution kernel.
      num_col: width of the convolution kernel.
      padding: padding mode in `Conv2D`.
      strides: strides in `Conv2D`.
      name: name of the ops; will become `name + '_conv'`
        for the convolution and `name + '_bn'` for the
        batch norm layer.

    Returns:
      Output tensor after applying `Conv2D` and `BatchNormalization`.
    NZ_bnZ_convr   r   r   F)r   r   Zuse_biasr    )r   Zscaler    Zrelur%   )r   r+   r,   ZConv2DZBatchNormalizationZ
Activation)
r2   filtersZnum_rowZnum_colr   r   r    Zbn_nameZ	conv_nameZbn_axisr4   r4   r5   r-     s*   
r-   z0keras.applications.inception_v3.preprocess_inputc                 C   s   t j| |ddS )Nr'   )r   mode)r   preprocess_input)r2   r   r4   r4   r5   r9     s   r9   z2keras.applications.inception_v3.decode_predictionsr   c                 C   s   t j| |dS )N)top)r   decode_predictions)Zpredsr:   r4   r4   r5   r;     s   r;    )r8   reterror)Tr	   NNNr
   r   )r   r   N)N)r   )__doc__Ztensorflow.compat.v2compatZv2r'   Zkerasr   Zkeras.applicationsr   Zkeras.enginer   Zkeras.layersr   Zkeras.utilsr   r   Z tensorflow.python.util.tf_exportr   r0   r1   r,   r6   r-   r9   r;   ZPREPROCESS_INPUT_DOCformatZPREPROCESS_INPUT_RET_DOC_TFZPREPROCESS_INPUT_ERROR_DOCr4   r4   r4   r5   <module>   sP     d
*