arXiv Analytics

Sign in

arXiv:1805.05814 [stat.ML]AbstractReferencesReviewsResources

SHADE: Information-Based Regularization for Deep Learning

Michael Blot, Thomas Robert, Nicolas Thome, Matthieu Cord

Published 2018-05-14Version 1

Regularization is a big issue for training deep neural networks. In this paper, we propose a new information-theory-based regularization scheme named SHADE for SHAnnon DEcay. The originality of the approach is to define a prior based on conditional entropy, which explicitly decouples the learning of invariant representations in the regularizer and the learning of correlations between inputs and labels in the data fitting term. Our second contribution is to derive a stochastic version of the regularizer compatible with deep learning, resulting in a tractable training scheme. We empirically validate the efficiency of our approach to improve classification performances compared to standard regularization schemes on several standard architectures.

Comments: IEEE International Conference on Image Processing (ICIP) 2018. arXiv admin note: substantial text overlap with arXiv:1804.10988
Categories: stat.ML, cs.LG
Related articles: Most relevant | Search more
arXiv:1804.10988 [stat.ML] (Published 2018-04-29)
SHADE: Information-Based Regularization for Deep Learning
arXiv:1805.09317 [stat.ML] (Published 2018-05-23)
Communication Algorithms via Deep Learning
arXiv:2012.06969 [stat.ML] (Published 2020-12-13, updated 2020-12-16)
Predicting Generalization in Deep Learning via Local Measures of Distortion