arXiv Analytics

Sign in

arXiv:1902.05312 [stat.ML]AbstractReferencesReviewsResources

Generalisation in fully-connected neural networks for time series forecasting

Anastasia Borovykh, Cornelis W. Oosterlee, Sander M. Bohte

Published 2019-02-14Version 1

In this paper we study the generalisation capabilities of fully-connected neural networks trained in the context of time series forecasting. Time series do not satisfy the typical assumption in statistical learning theory of the data being i.i.d. samples from some data-generating distribution. We use the input and weight Hessians, that is the smoothness of the learned function with respect to the input and the width of the minimum in weight space, to quantify a network's ability to generalise to unseen data. While such generalisation metrics have been studied extensively in the i.i.d. setting of for example image recognition, here we empirically validate their use in the task of time series forecasting. Furthermore we discuss how one can control the generalisation capability of the network by means of the training process using the learning rate, batch size and the number of training iterations as controls. Using these hyperparameters one can efficiently control the complexity of the output function without imposing explicit constraints.

Related articles: Most relevant | Search more
arXiv:1909.13316 [stat.ML] (Published 2019-09-29)
Machine Learning vs Statistical Methods for Time Series Forecasting: Size Matters
arXiv:2304.03877 [stat.ML] (Published 2023-04-08)
OFTER: An Online Pipeline for Time Series Forecasting
arXiv:1803.09704 [stat.ML] (Published 2018-03-26, updated 2018-08-11)
MOrdReD: Memory-based Ordinal Regression Deep Neural Networks for Time Series Forecasting