arXiv Analytics

Sign in

arXiv:1609.01596 [stat.ML]AbstractReferencesReviewsResources

Direct Feedback Alignment Provides Learning in Deep Neural Networks

Arild Nøkland

Published 2016-09-06Version 1

Artificial neural networks are most commonly trained with the back-propagation algorithm, where the gradient for learning is provided by back-propagating the error, layer by layer, from the output layer to the hidden layers. A recently discovered method called feedback-alignment shows that the weights used for propagating the error backward don't have to be symmetric with the weights used for propagation the activation forward. In fact, random feedback weights work evenly well, because the network learns how to make the feedback useful. In this work, the feedback alignment principle is used for training hidden layers more independently from the rest of the network, and from a zero initial condition. The error is propagated through fixed random feedback connections directly from the output layer to each hidden layer. This simple method is able to achieve zero training error even in convolutional networks and very deep networks, completely without error back-propagation. The method is a step towards biologically plausible machine learning because the error signal is almost local, and no symmetric or reciprocal weights are required. Experiments show that the test performance on MNIST and CIFAR is almost as good as those obtained with back-propagation for fully connected networks. If combined with dropout, the method achieves 1.45% error on the permutation invariant MNIST task.

Comments: Accepted for publication at NIPS 2016
Categories: stat.ML, cs.LG
Related articles: Most relevant | Search more
arXiv:1805.10965 [stat.ML] (Published 2018-05-28)
Lipschitz regularity of deep neural networks: analysis and efficient estimation
arXiv:1402.1869 [stat.ML] (Published 2014-02-08, updated 2014-06-07)
On the Number of Linear Regions of Deep Neural Networks
arXiv:1712.09482 [stat.ML] (Published 2017-12-27)
Robust Loss Functions under Label Noise for Deep Neural Networks