arXiv Analytics

Sign in

arXiv:cs/0611011 [cs.LG]AbstractReferencesReviewsResources

Hedging predictions in machine learning

Alexander Gammerman, Vladimir Vovk

Published 2006-11-02Version 1

Recent advances in machine learning make it possible to design efficient prediction algorithms for data sets with huge numbers of parameters. This paper describes a new technique for "hedging" the predictions output by many such algorithms, including support vector machines, kernel ridge regression, kernel nearest neighbours, and by many other state-of-the-art methods. The hedged predictions for the labels of new objects include quantitative measures of their own accuracy and reliability. These measures are provably valid under the assumption of randomness, traditional in machine learning: the objects and their labels are assumed to be generated independently from the same probability distribution. In particular, it becomes possible to control (up to statistical fluctuations) the number of erroneous predictions by selecting a suitable confidence level. Validity being achieved automatically, the remaining goal of hedged prediction is efficiency: taking full account of the new objects' features and other available information to produce as accurate predictions as possible. This can be done successfully using the powerful machinery of modern machine learning.

Comments: 24 pages; 9 figures; 2 tables; a version of this paper (with discussion and rejoinder) is to appear in "The Computer Journal"
Journal: Computer Journal, 50:151-177, 2007
Categories: cs.LG
Related articles: Most relevant | Search more
arXiv:1506.00976 [cs.LG] (Published 2015-06-02)
Toward a generic representation of random variables for machine learning
arXiv:1808.00931 [cs.LG] (Published 2018-08-02)
Machine Learning of Space-Fractional Differential Equations
arXiv:1606.04275 [cs.LG] (Published 2016-06-14)
Efficient Pairwise Learning Using Kernel Ridge Regression: an Exact Two-Step Method