arXiv Analytics

Sign in

arXiv:1812.05836 [cs.LG]AbstractReferencesReviewsResources

Rethinking Layer-wise Feature Amounts in Convolutional Neural Network Architectures

Martin Mundt, Sagnik Majumder, Tobias Weis, Visvanathan Ramesh

Published 2018-12-14Version 1

We characterize convolutional neural networks with respect to the relative amount of features per layer. Using a skew normal distribution as a parametrized framework, we investigate the common assumption of monotonously increasing feature-counts with higher layers of architecture designs. Our evaluation on models with VGG-type layers on the MNIST, Fashion-MNIST and CIFAR-10 image classification benchmarks provides evidence that motivates rethinking of our common assumption: architectures that favor larger early layers seem to yield better accuracy.

Comments: Accepted at the Critiquing and Correcting Trends in Machine Learning (CRACT) Workshop at the 32nd Conference on Neural Information Processing Systems (NeurIPS 2018)
Categories: cs.LG, cs.CV, stat.ML
Related articles: Most relevant | Search more
arXiv:2204.08624 [cs.LG] (Published 2022-04-19)
Topology and geometry of data manifold in deep learning
arXiv:2010.12046 [cs.LG] (Published 2020-10-22)
Using Deep Image Priors to Generate Counterfactual Explanations
arXiv:1909.09432 [cs.LG] (Published 2019-09-20)
Genetic Neural Architecture Search for automatic assessment of human sperm images