
    jP                     D   d dl Z d dlmZmZmZ d dlmZ ed         Zd7dej	        defdZ
	 	 	 	 d8d
ej	        dej	        deej	                 dedededej	        fdZ	 	 	 d9dej	        dej	        deej	                 dededej	        fdZ	 d:dej	        dej	        dedej	        fdZ	 d:dej	        dej	        dedej	        fdZ	 d;dej	        dej	        dededej	        f
dZ	 	 	 d<dej	        dej	        dej	        dedededej	        fdZ	 d;dej	        dej	        dededej	        f
d Z	 	 d=dej	        dej	        d"ededej	        f
d#Z	 	 	 	 	 d>d%ej	        d&ej	        d'ej	        ded(ed)edededej	        fd*Z	 d7dej	        dej	        dedej	        fd+Z	 	 d?dej	        dej	        d,ededej	        f
d-Z	 d7dej	        dej	        dedej	        fd.Z	 	 	 d@d1ej	        d2ej	        dedededej	        fd3Z	 	 dAd4ej	        d5ej	        dej	        d)ededej	        fd6ZdS )B    N)LiteralOptionalget_args)nonemeansumr   loss	reductionc                     |t          t                    vr%t          dt          t                     d          |dk    rt          j        |           S |dk    rt          j        |           S |dk    r| S d S )Nz"Invalid reduction. Must be one of .r   r   r   )r   	Reduction
ValueErrormxr   r   )r	   r
   s     W/lsinfo/ai/hellotax_ai/base_platform/venv/lib/python3.11/site-packages/mlx/nn/losses.py_reducer      s    ++++Thy>Q>QTTTUUUFwt}}	e		vd||	f		 
	            logitstargetsweightsaxislabel_smoothingreturnc                    |dk     s|dk    rt          d| d          |j        | j        k    }d }|r|j        | j        k    s|s:|j         || j        |          k    r t          d|j         d| j         d          |rt          j        | |z  |          }n<t          j        | t          j        ||          |                              |          }t          j        | |          }	|dk    r-d|z
  |z  }
| 	                    |          }| |z  }|	|
z
  |z   }n|	|z
  }|5|j        |j        k    r t          d
|j         d|j         d          ||z  }t          ||          S )aa  
    Computes the cross entropy loss.

    Args:
        logits (array): The unnormalized logits.
        targets (array): The ground truth values. These can be class indices or
            probabilities for each class. If the ``targets`` are class indices,
            then ``targets`` shape should match the ``logits`` shape with
            the ``axis`` dimension removed. If the ``targets`` are probabilities
            (or one-hot encoded), then the ``targets`` shape should be the same as
            the ``logits`` shape.
        weights (array, optional): Optional weights for each target. Default: ``None``.
        axis (int, optional): The axis over which to compute softmax. Default: ``-1``.
        label_smoothing (float, optional): Label smoothing factor. Default: ``0``.
        reduction (str, optional): Specifies the reduction to apply to the output:
            ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: The computed cross entropy loss.

    Examples:
        >>> import mlx.core as mx
        >>> import mlx.nn as nn
        >>>
        >>> # Class indices as targets
        >>> logits = mx.array([[2.0, -1.0], [-1.0, 2.0]])
        >>> targets = mx.array([0, 1])
        >>> nn.losses.cross_entropy(logits, targets)
        array([0.0485873, 0.0485873], dtype=float32)
        >>>
        >>> # Probabilities (or one-hot vectors) as targets
        >>> logits = mx.array([[2.0, -1.0], [-1.0, 2.0]])
        >>> targets = mx.array([[0.9, 0.1], [0.1, 0.9]])
        >>> nn.losses.cross_entropy(logits, targets)
        array([0.348587, 0.348587], dtype=float32)
    r      z$Label smoothing must in [0, 1), got r   c                 h    t          |           } |                     |           t          |           S )N)listpoptuple)shaper   s     r   	_drop_dimz cross_entropy.<locals>._drop_dimI   s(    U		$U||r   zTargets shape z does not match logits shape r   NWeights with shape + is not the same as output loss with shape )r   ndimr!   r   r   take_along_axisexpand_dimssqueeze	logsumexpr   r   )r   r   r   r   r   r
   targets_as_probsr"   scorelogsumexp_logitsadjusted_scoremean_logitssmoothed_lossr	   s                 r   cross_entropyr1      s   X o22RRRRSSS |v{2   	 
W]fl:: ;!())FL$2O2O!O!OXW]XXXXX
 
 	
  
v'd333"62>'4+H+H$OOWW
 
 |F666o-6 kktk,,$6  .0=@%' =DJ&&8gm 8 8*.*8 8 8   	 4###r   Tr   inputswith_logitsc                    | j         |j         k    r t          d| j          d|j          d          |rt          j        d|           | |z  z
  }ndt          j        t          j        |           dd          }t          j        t          j        d| z
            dd          }||z  d|z
  |z  z    }|5|j         |j         k    r t          d	|j          d
|j          d          ||z  }t          ||          S )a  
    Computes the binary cross entropy loss.

    By default, this function takes the pre-sigmoid logits, which results in a faster
    and more precise loss. For improved numerical stability when ``with_logits=False``,
    the loss calculation clips the input probabilities (in log-space) to a minimum value
    of ``-100``.

    Args:
        inputs (array): The predicted values. If ``with_logits`` is ``True``, then
            ``inputs`` are unnormalized logits. Otherwise, ``inputs`` are probabilities.
        targets (array): The binary target values in {0, 1}.
        with_logits (bool, optional): Whether ``inputs`` are logits. Default: ``True``.
        weights (array, optional): Optional weights for each target. Default: ``None``.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'mean'``.

    Returns:
        array: The computed binary cross entropy loss.
    Examples:
        >>> import mlx.core as mx
        >>> import mlx.nn as nn

        >>> logits = mx.array([0.105361, 0.223144, 1.20397, 0.916291])
        >>> targets = mx.array([0, 0, 1, 1])
        >>> loss = nn.losses.binary_cross_entropy(logits, targets, reduction="mean")
        >>> loss
        array(0.539245, dtype=float32)

        >>> probs = mx.array([0.1, 0.1, 0.4, 0.4])
        >>> targets = mx.array([0, 0, 1, 1])
        >>> loss = nn.losses.binary_cross_entropy(probs, targets, with_logits=False, reduction="mean")
        >>> loss
        array(0.510826, dtype=float32)
    Inputs shape  does not match targets shape r   r   iN)a_mina_maxr   r$   r%   )r!   r   r   	logaddexpcliplogr   )r2   r   r   r3   r
   r	   log_inputs_cliplog_inputs_inv_clips           r   binary_cross_entropyr>   x   s2   T |w}$$XFLXXXXX
 
 	
  R|C((6G+;;'"&..DIII gbfQZ&8&8DQQQ?*a'k=P-PPQ =DJ&&8gm 8 8*.*8 8 8   	4###r   predictionsc                     | j         |j         k    r t          d| j          d|j          d          t          j        | |z
            }t	          ||          S )aS  
    Computes the L1 loss.

    Args:
        predictions (array): The predicted values.
        targets (array): The target values.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'mean'``.

    Returns:
        array: The computed L1 loss.
    Predictions shape r6   r   )r!   r   r   absr   r?   r   r
   r	   s       r   l1_lossrD      st     GM)).!2 . .$]. . .
 
 	
 6+'((D4###r   c                     | j         |j         k    r t          d| j          d|j          d          t          j        | |z
            }t	          ||          S )as  
    Computes the mean squared error loss.

    Args:
        predictions (array): The predicted values.
        targets (array): The target values.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'mean'``.

    Returns:
        array: The computed mean squared error loss.
    rA   r6   r   )r!   r   r   squarer   rC   s       r   mse_lossrG      st     GM)).!2 . .$]. . .
 
 	

 9[7*++D4###r   c                     t          j        | |d         |                              d           }t          ||          S )a  
    Computes the negative log likelihood loss.

    Args:
        inputs (array): The predicted distribution in log space.
        targets (array): The target values.
        axis (int, optional): The distribution axis. Default: ``-1``.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: The computed NLL loss.
    ).Nr   )r   r'   r)   r   r2   r   r   r
   r	   s        r   nll_lossrJ      s>      vwy'94@@HHLLLD4###r   Fư>varsfullepsc                    | j         |j         k    r t          d| j          d|j          d          | j         |j         k    r t          d| j          d|j          d          t          j        ||          }dt          j        |          t          j        || z
            |z  z   z  }|r'|dt          j        dt          j        z            z  z  }t          ||          S )aj  
    Computes the negative log likelihood loss for a Gaussian distribution.

    The loss is given by:

    .. math::
        \frac{1}{2}\left(\log\left(\max\left(\text{vars},
        \ \epsilon\right)\right) + \frac{\left(\text{inputs} - \text{targets} \right)^2}
        {\max\left(\text{vars}, \ \epsilon \right)}\right) + \text{const.}

    where ``inputs`` are the predicted means and ``vars`` are the the
    predicted variances.

    Args:
        inputs (array): The predicted expectation of the Gaussian distribution.
        targets (array): The target values (samples from the Gaussian distribution).
        vars (array): The predicted variance of the Gaussian distribution.
        full (bool, optional): Whether to include the constant term in the loss calculation.
            Default: ``False``.
        eps (float, optional): Small positive constant for numerical stability.
            Default: ``1e-6``.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: The Gaussian NLL loss.
    r5   r6   r   z does not match vars shape       ?   )	r!   r   r   maximumr;   rF   mathpir   )r2   r   rL   rM   rN   r
   r	   s          r   gaussian_nll_lossrU     s    F |w}$$XFLXXXXX
 
 	
 |tz!!RFLRRTZRRR
 
 	

 :dC  D"&,,7V+;!<!<t!CCDD ,dhq47{++++4###r   c                 |    t          j        t          j        |          || z
  z  |          }t          ||          S )a  
    Computes the Kullback-Leibler divergence loss.

    Computes the following when ``reduction == 'none'``:

    .. code-block:: python

        mx.exp(targets) * (targets - inputs).sum(axis)

    Args:
        inputs (array): Log probabilities for the predicted distribution.
        targets (array): Log probabilities for the target distribution.
        axis (int, optional): The distribution axis. Default: ``-1``.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: The computed Kullback-Leibler divergence loss.
    )r   r   expr   rI   s        r   kl_div_lossrX   8  s7    , 6"&//Wv%56==D4###r         ?betac                 D   | j         |j         k    r t          d| j          d|j          d          t          j        | |z
            }t          j        ||k     dt          j        |          z  |z  t          j        |          d|z  z
            }t          ||          S )a  
    Computes the smooth L1 loss.

    The smooth L1 loss is a variant of the L1 loss which replaces the absolute
    difference with a squared difference when the absolute difference is less
    than ``beta``.

    The formula for the smooth L1 Loss is:

    .. math::

      l = \begin{cases}
            0.5 (x - y)^2 / \beta, & \text{if } |x - y| < \beta \\
            |x - y| - 0.5 \beta, & \text{otherwise}
          \end{cases}

    Args:
        predictions (array): Predicted values.
        targets (array): Ground truth values.
        beta (float, optional): The threshold after which the loss changes
          from the squared to the absolute difference. Default: ``1.0``.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'mean'``.

    Returns:
        array: The computed smooth L1 loss.
    rA   r6   r   rP   )r!   r   r   rB   whererF   r   )r?   r   rZ   r
   diffr	   s         r   smooth_l1_lossr^   S  s    B GM)).!2 . .$]. . .
 
 	

 6+'((D8tS29T??*T126$<<#*3L D 4###r   rQ   anchors	positives	negativespmarginc           	      P   t          j        t          j        t          j        | |z
  |                              |          |z             t          j        t          j        | |z
  |                              |          |z             z
  |z   d          }t          ||          S )a1  
    Computes the triplet loss for a set of anchor, positive, and negative samples.
    Margin is represented with alpha in the math section.

    .. math::

       \max\left(\|A - P\|_p - \|A - N\|_p + \alpha, 0\right)

    Args:
        anchors (array): The anchor samples.
        positives (array): The positive samples.
        negatives (array): The negative samples.
        axis (int, optional): The distribution axis. Default: ``-1``.
        p (int, optional): The norm degree for pairwise distance. Default: ``2``.
        margin (float, optional): Margin for the triplet loss. Defaults to ``1.0``.
        eps (float, optional): Small positive constant to prevent numerical instability. Defaults to ``1e-6``.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: Computed triplet loss. If reduction is "none", returns a tensor of the same shape as input;
                  if reduction is "mean" or "sum", returns a scalar tensor.
    r   )r   rR   sqrtpowerr   r   )	r_   r`   ra   r   rb   rc   rN   r
   r	   s	            r   triplet_lossrg     s    B :
9,a0044T::S@AA
'"(7Y.2266t<<sB
C
C	D
	 	
	 D 4###r   c                 X    t          j        d| |z  z
  d          }t          ||          S )a  
    Computes the hinge loss between inputs and targets.

    .. math::

       \text{hinge}(y, y_{\text{pred}}) = \max(0, 1 - y \cdot y_{\text{pred}})


    Args:
        inputs (array): The predicted values.
        targets (array): The target values. They should be -1 or 1.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: The computed hinge loss.
    r   r   )r   rR   r   )r2   r   r
   r	   s       r   
hinge_lossri     s/    ( :a&7**A..D4###r   deltac                     | |z
  }t          j        |          }t          j        ||          }||z
  }d|dz  z  ||z  z   }t          ||          S )a  
    Computes the Huber loss between inputs and targets.

    .. math::

        l_{\delta}(a) =
        \left\{ \begin{array}{ll}
            \frac{1}{2} a^2 & \text{for } |a| \leq \delta, \\
            \delta \left( |a| - \frac{1}{2} \delta \right) & \text{otherwise.}
        \end{array} \right.

    Args:
        inputs (array): The predicted values.
        targets (array): The target values.
        delta (float, optional): The threshold at which to change between L1 and L2 loss.
          Default: ``1.0``.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: The computed Huber loss.
    rP   rQ   )r   rB   minimumr   )	r2   r   rj   r
   errors
abs_errors	quadraticlinearr	   s	            r   
huber_lossrq     s]    8 gFJ
:u--I)#FA.D4###r   c                     | |z
  }t          j        ||           t          j        d          z
  }t	          ||          S )a  
    Computes the log cosh loss between inputs and targets.

    Logcosh acts like L2 loss for small errors, ensuring stable gradients,
    and like the L1 loss for large errors, reducing sensitivity to outliers. This
    dual behavior offers a balanced, robust approach for regression tasks.

    .. math::

       \text{logcosh}(y_{\text{true}}, y_{\text{pred}}) =
            \frac{1}{n} \sum_{i=1}^{n}
            \log(\cosh(y_{\text{pred}}^{(i)} - y_{\text{true}}^{(i)}))


    Args:
        inputs (array): The predicted values.
        targets (array): The target values.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: The computed log cosh loss.
    rQ   )r   r9   rS   r;   r   )r2   r   r
   rm   r	   s        r   log_cosh_lossrs     s>    4 gF<((48A;;6D4###r   r   :0yE>x1x2c                 
   t           j                            | |          }t           j                            ||          }t          j        | |z  |          t          j        ||z  |          z  }t          ||          S )a  
    Computes the cosine similarity between the two inputs.

    The cosine similarity loss is given by

    .. math::

        \frac{x_1 \cdot x_2}{\max(\|x_1\|  \cdot \|x_2\|, \epsilon)}

    Args:
        x1 (mx.array): The first set of inputs.
        x2 (mx.array): The second set of inputs.
        axis (int, optional): The embedding axis. Default: ``1``.
        eps (float, optional): The minimum value of the denominator used for
          numerical stability. Default: ``1e-8``.
        reduction (str, optional): Specifies the reduction to apply to the output:
          ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        mx.array: The computed cosine similarity loss.
    r#   )r   linalgnormr   rR   r   )ru   rv   r   rN   r
   x1_normx2_normr	   s           r   cosine_similarity_lossr|   
  sp    8 innRdn++GinnRdn++G6"r'%%%
7W3Dc(J(JJD4###r   inputs1inputs2c           	          | j         |j         cxk    r|j         k    s*n t          d| j          d|j          d|j          d          | |z
  }t          j        d| |z  |z             }t	          ||          S )aW  
    Calculate the margin ranking loss that loss given inputs :math:`x_1`, :math:`x_2` and a label
    :math:`y` (containing 1 or -1).

    The loss is given by:

    .. math::
        \text{loss} = \max (0, -y * (x_1 - x_2) + \text{margin})

    Where :math:`y` represents ``targets``, :math:`x_1` represents ``inputs1`` and :math:`x_2`
    represents ``inputs2``.

    Args:
        inputs1 (array): Scores for the first input.
        inputs2 (array): Scores for the second input.
        targets (array): Labels indicating whether samples in ``inputs1`` should be ranked higher
            than samples in ``inputs2``. Values should be 1 or -1.
        margin (float, optional): The margin by which the scores should be separated.
            Default: ``0.0``.
        reduction (str, optional): Specifies the reduction to apply to the output:
            ``'none'`` | ``'mean'`` | ``'sum'``. Default: ``'none'``.

    Returns:
        array: The computed margin ranking loss.

    Examples:
        >>> import mlx.core as mx
        >>> import mlx.nn as nn
        >>> targets = mx.array([1, 1, -1])
        >>> inputs1 = mx.array([-0.573409, -0.765166, -0.0638])
        >>> inputs2 = mx.array([0.75596, 0.225763, 0.256995])
        >>> loss = nn.losses.margin_ranking_loss(inputs1, inputs2, targets)
        >>> loss
        array(0.773433, dtype=float32)
    zPThe shapes of the arguments do not match. The provided shapes are inputs1.shape=z, inputs2.shape=z, and targets.shape=r   r   )r!   r   r   rR   r   )r}   r~   r   rc   r
   differencesr	   s          r   margin_ranking_lossr   .  s    T MW];;;;gm;;;;.$]. .<CM. .$]. . .
 
 	
 G#K:a'K/&899D4###r   )r   )Nr   r   r   )NTr   )r   )r   r   )FrK   r   )rY   r   )r   rQ   rY   rK   r   )rY   r   )r   rt   r   )r   r   )rS   typingr   r   r   mlx.corecorer   r   arrayr   intfloatr1   boolr>   rD   rG   rJ   rU   rX   r^   rg   ri   rq   rs   r|   r    r   r   <module>r      s    . . . . . . . . . .      )*		 	"( 	y 	 	 	 	 #' !^$ ^$H^$X^$ bh^$ 	^$
 ^$ ^$ X^$ ^$ ^$ ^$H #'!?$ ?$H?$X?$ bh?$ 	?$
 ?$ X?$ ?$ ?$ ?$F FL$ $$$&H$9B$X$ $ $ $4 FL$ $$$&H$9B$X$ $ $ $4 QW$ $H$!x$/2$DM$X$ $ $ $2 !4$ 4$H4$X4$ (4$ 	4$
 
4$ 4$ X4$ 4$ 4$ 4$p QW$ $H$!x$/2$DM$X$ $ $ $< !	,$ ,$,$X,$ ,$ 	,$
 X,$ ,$ ,$ ,$f !'$ '$X'$x'$ x'$ 	'$
 '$ '$ 
'$ '$ X'$ '$ '$ '$V AG$ $H$!x$4=$X$ $ $ $8 !	"$ "$H"$X"$ "$ 	"$
 X"$ "$ "$ "$L AG$ $H$!x$4=$X$ $ $ $F !!$ !$
!$
!$ !$ 
	!$
 !$ X!$ !$ !$ !$P !4$ 4$X4$X4$ X4$ 	4$
 4$ X4$ 4$ 4$ 4$ 4$ 4$r   