arXiv Analytics

Sign in

arXiv:1909.08079 [stat.ML]AbstractReferencesReviewsResources

Relaxed Softmax for learning from Positive and Unlabeled data

Ugo Tanielian, Flavian Vasile

Published 2019-09-17Version 1

In recent years, the softmax model and its fast approximations have become the de-facto loss functions for deep neural networks when dealing with multi-class prediction. This loss has been extended to language modeling and recommendation, two fields that fall into the framework of learning from Positive and Unlabeled data. In this paper, we stress the different drawbacks of the current family of softmax losses and sampling schemes when applied in a Positive and Unlabeled learning setup. We propose both a Relaxed Softmax loss (RS) and a new negative sampling scheme based on Boltzmann formulation. We show that the new training objective is better suited for the tasks of density estimation, item similarity and next-event prediction by driving uplifts in performance on textual and recommendation datasets against classical softmax.

Comments: 9 pages, 5 figures, 2 tables, published at RecSys 2019
Journal: RecSys 2019 Proceedings of the 13th ACM Conference on Recommender Systems
Categories: stat.ML, cs.CL, cs.LG
Related articles: Most relevant | Search more
arXiv:1402.1869 [stat.ML] (Published 2014-02-08, updated 2014-06-07)
On the Number of Linear Regions of Deep Neural Networks
arXiv:1905.10634 [stat.ML] (Published 2019-05-25)
Adaptive, Distribution-Free Prediction Intervals for Deep Neural Networks
arXiv:1901.02182 [stat.ML] (Published 2019-01-08)
Comments on "Deep Neural Networks with Random Gaussian Weights: A Universal Classification Strategy?"