This is not an officially supported Google product.
Visualization of practical properties of the Bi-Tempered logistic loss.
A replacement for standard logistic loss function: tf.losses.softmax_cross_entropy
is available here
def bi_tempered_logistic_loss(activations,
labels,
t1,
t2,
label_smoothing=0.0,
num_iters=5):
"""Bi-Tempered Logistic Loss with custom gradient.
Args:
activations: A multi-dimensional tensor with last dimension `num_classes`.
labels: A tensor with shape and dtype as activations.
t1: Temperature 1 (< 1.0 for boundedness).
t2: Temperature 2 (> 1.0 for tail heaviness).
label_smoothing: Label smoothing parameter between [0, 1).
num_iters: Number of iterations to run the method.
Returns:
A loss tensor.
"""
When referencing Bi-Tempered loss, cite this paper:
@article{bitemperedloss,
author = {Ehsan Amid and
Manfred K. Warmuth and
Rohan Anil and
Tomer Koren},
title = {Robust Bi-Tempered Logistic Loss Based on Bregman Divergences},
journal = {CoRR},
volume = {abs/1906.03361},
year = {2019},
url = {http://arxiv.org/abs/1906.03361},
}
We are eager to collaborate with you too! Please send us a pull request or open a github issue. Please see this doc for further details