arXiv Analytics

Sign in

arXiv:2007.06737 [cs.LG]AbstractReferencesReviewsResources

Representation Transfer by Optimal Transport

Xuhong Li, Yves Grandvalet, Rémi Flamary, Nicolas Courty, Dejing Dou

Published 2020-07-13Version 1

Deep learning currently provides the best representations of complex objects for a wide variety of tasks. However, learning these representations is an expensive process that requires very large training samples and significant computing resources. Thankfully, sharing these representations is a common practice, enabling to solve new tasks with relatively little training data and few computing resources; the transfer of representations is nowadays an essential ingredient in numerous real-world applications of deep learning. Transferring representations commonly relies on the parameterized form of the features making up the representation, as encoded by the computational graph of these features. In this paper, we propose to use a novel non-parametric metric between representations. It is based on a functional view of features, and takes into account certain invariances of representations, such as the permutation of their features, by relying on optimal transport. This distance is used as a regularization term promoting similarity between two representations. We show the relevance of this approach in two representation transfer settings, where the representation of a trained reference model is transferred to another one, for solving a new related task (inductive transfer learning), or for distilling knowledge to a simpler model (model compression).

Related articles: Most relevant | Search more
arXiv:1905.11885 [cs.LG] (Published 2019-05-28)
Differentiable Sorting using Optimal Transport:The Sinkhorn CDF and Quantile Operator
arXiv:1906.09218 [cs.LG] (Published 2019-06-21)
FlipTest: Fairness Auditing via Optimal Transport
arXiv:1903.03850 [cs.LG] (Published 2019-03-09)
Stochastic Incremental Algorithms for Optimal Transport with SON Regularizer