arXiv Analytics

Sign in

arXiv:1805.06523 [cs.LG]AbstractReferencesReviewsResources

End-to-end Learning of a Convolutional Neural Network via Deep Tensor Decomposition

Samet Oymak, Mahdi Soltanolkotabi

Published 2018-05-16Version 1

In this paper we study the problem of learning the weights of a deep convolutional neural network. We consider a network where convolutions are carried out over non-overlapping patches with a single kernel in each layer. We develop an algorithm for simultaneously learning all the kernels from the training data. Our approach dubbed Deep Tensor Decomposition (DeepTD) is based on a rank-1 tensor decomposition. We theoretically investigate DeepTD under a realizable model for the training data where the inputs are chosen i.i.d. from a Gaussian distribution and the labels are generated according to planted convolutional kernels. We show that DeepTD is data-efficient and provably works as soon as the sample size exceeds the total number of convolutional weights in the network. We carry out a variety of numerical experiments to investigate the effectiveness of DeepTD and verify our theoretical findings.

Related articles: Most relevant | Search more
arXiv:2002.05707 [cs.LG] (Published 2020-02-13)
A Framework for End-to-End Learning on Semantic Tree-Structured Data
arXiv:1906.03986 [cs.LG] (Published 2019-06-10)
Unit Impulse Response as an Explainer of Redundancy in a Deep Convolutional Neural Network
arXiv:1805.09898 [cs.LG] (Published 2018-05-24)
Generative Model: Membership Attack,Generalization and Diversity