arXiv Analytics

Sign in

arXiv:1805.00325 [cs.CV]AbstractReferencesReviewsResources

Study of Residual Networks for Image Recognition

Mohammad Sadegh Ebrahimi, Hossein Karkeh Abadi

Published 2018-04-21Version 1

Deep neural networks demonstrate to have a high performance on image classification tasks while being more difficult to train. Due to the complexity and vanishing gradient problem, it normally takes a lot of time and more computational power to train deeper neural networks. Deep residual networks (ResNets) can make the training process faster and attain more accuracy compared to their equivalent neural networks. ResNets achieve this improvement by adding a simple skip connection parallel to the layers of convolutional neural networks. In this project we first design a ResNet model that can perform the image classification task on the Tiny ImageNet dataset with a high accuracy, then we compare the performance of this ResNet model with its equivalent Convolutional Network (ConvNet). Our findings illustrate that ResNets are more prone to overfitting despite their higher accuracy. Several methods to prevent overfitting such as adding dropout layers and stochastic augmentation of the training dataset has been studied in this work.

Related articles: Most relevant | Search more
arXiv:2105.01883 [cs.CV] (Published 2021-05-05)
RepMLP: Re-parameterizing Convolutions into Fully-connected Layers for Image Recognition
arXiv:1802.02207 [cs.CV] (Published 2018-01-22)
Automated dataset generation for image recognition using the example of taxonomy
arXiv:2211.06163 [cs.CV] (Published 2022-11-11)
Dual Complementary Dynamic Convolution for Image Recognition