arXiv Analytics

Sign in

arXiv:2106.02676 [math.NA]AbstractReferencesReviewsResources

A novel multi-scale loss function for classification problems in machine learning

Leonid Berlyand, Robert Creese, Pierre-Emmanuel Jabin

Published 2021-06-04Version 1

We introduce two-scale loss functions for use in various gradient descent algorithms applied to classification problems via deep neural networks. This new method is generic in the sense that it can be applied to a wide range of machine learning architectures, from deep neural networks to support vector machines for example. These two-scale loss functions allow to focus the training onto objects in the training set which are not well classified. This leads to an increase in several measures of performance for appropriately-defined two-scale loss functions with respect to the more classical cross-entropy when tested on traditional deep neural networks on the MNIST, CIFAR10, and CIFAR100 data-sets.

Related articles: Most relevant | Search more
arXiv:2009.14596 [math.NA] (Published 2020-09-23)
Machine Learning and Computational Mathematics
arXiv:2004.01138 [math.NA] (Published 2020-04-02)
Numerical analysis of least squares and perceptron learning for classification problems
arXiv:2410.09666 [math.NA] (Published 2024-10-12)
A forward scheme with machine learning for forward-backward SDEs with jumps by decoupling jumps