arXiv Analytics

Sign in

arXiv:1705.02302 [cs.LG]AbstractReferencesReviewsResources

Analysis and Design of Convolutional Networks via Hierarchical Tensor Decompositions

Nadav Cohen, Or Sharir, Ronen Tamari, David Yakira, Yoav Levine, Amnon Shashua

Published 2017-05-05Version 1

The driving force behind convolutional networks - the most successful deep learning architecture to date, is their expressive power. Despite its wide acceptance and vast empirical evidence, formal analyses supporting this belief are scarce. The primary notions for formally reasoning about expressiveness are efficiency and inductive bias. Efficiency refers to the ability of a network architecture to realize functions that require an alternative architecture to be much larger. Inductive bias refers to the prioritization of some functions over others given prior knowledge regarding a task at hand. In this paper we provide a high-level overview of a series of works written by the authors, that through an equivalence to hierarchical tensor decompositions, analyze the expressive efficiency and inductive bias of various architectural features in convolutional networks (depth, width, convolution strides and more). The results presented shed light on the demonstrated effectiveness of convolutional networks, and in addition, provide new tools for network design.

Related articles: Most relevant | Search more
arXiv:1506.08230 [cs.LG] (Published 2015-06-26)
Scale-invariant learning and convolutional networks
arXiv:2106.04693 [cs.LG] (Published 2021-06-08)
On the Evolution of Neuron Communities in a Deep Learning Architecture
arXiv:2408.16686 [cs.LG] (Published 2024-08-29)
CW-CNN & CW-AN: Convolutional Networks and Attention Networks for CW-Complexes