arXiv Analytics

Sign in

arXiv:1905.12882 [cs.LG]AbstractReferencesReviewsResources

Function approximation by deep networks

H. N. Mhaskar, T. Poggio

Published 2019-05-30Version 1

We show that deep networks are better than shallow networks at approximating functions that can be expressed as a composition of functions described by a directed acyclic graph, because the deep networks can be designed to have the same compositional structure, while a shallow network cannot exploit this knowledge. Thus, the blessing of compositionality mitigates the curse of dimensionality. On the other hand, a theorem called good propagation of errors allows to `lift' theorems about shallow networks to those about deep networks with an appropriate choice of norms, smoothness, etc. We illustrate this in three contexts where each channel in the deep network calculates a spherical polynomial, a non-smooth ReLU network, or another zonal function network related closely with the ReLU network.

Related articles: Most relevant | Search more
arXiv:2402.12687 [cs.LG] (Published 2024-02-20, updated 2024-08-18)
Learning on manifolds without manifold learning
arXiv:1911.04384 [cs.LG] (Published 2019-11-11)
Provably Convergent Off-Policy Actor-Critic with Function Approximation
arXiv:2310.19861 [cs.LG] (Published 2023-10-30)
Posterior Sampling for Competitive RL: Function Approximation and Partial Observation