arXiv Analytics

Sign in

arXiv:2205.08836 [cs.CV]AbstractReferencesReviewsResources

Large Neural Networks Learning from Scratch with Very Few Data and without Regularization

Christoph Linse, Thomas Martinetz

Published 2022-05-18Version 1

Recent findings have shown that Neural Networks generalize also in over-parametrized regimes with zero training error. This is surprising, since it is completely against traditional machine learning wisdom. In our empirical study we fortify these findings in the domain of fine-grained image classification. We show that very large Convolutional Neural Networks with millions of weights do learn with only a handful of training samples and without image augmentation, explicit regularization or pretraining. We train the architectures ResNet018, ResNet101 and VGG19 on subsets of the difficult benchmark datasets Caltech101, CUB_200_2011, FGVCAircraft, Flowers102 and StanfordCars with 100 classes and more, perform a comprehensive comparative study and draw implications for the practical application of CNNs. Finally, we show that VGG19 with 140 million weights learns to distinguish airplanes and motorbikes up to 95% accuracy with only 20 samples per class.

Related articles: Most relevant | Search more
arXiv:2311.15658 [cs.CV] (Published 2023-11-27)
Regularization by Texts for Latent Diffusion Inverse Solvers
arXiv:1502.06105 [cs.CV] (Published 2015-02-21)
Regularization and Kernelization of the Maximin Correlation Approach
arXiv:2211.10948 [cs.CV] (Published 2022-11-20, updated 2023-09-18)
FedDCT: Federated Learning of Large Convolutional Neural Networks on Resource Constrained Devices using Divide and Collaborative Training