arXiv:1402.2300 [cs.LG]AbstractReferencesReviewsResources
Feature and Variable Selection in Classification
Published 2014-02-10Version 1
The amount of information in the form of features and variables avail- able to machine learning algorithms is ever increasing. This can lead to classifiers that are prone to overfitting in high dimensions, high di- mensional models do not lend themselves to interpretable results, and the CPU and memory resources necessary to run on high-dimensional datasets severly limit the applications of the approaches. Variable and feature selection aim to remedy this by finding a subset of features that in some way captures the information provided best. In this paper we present the general methodology and highlight some specific approaches.
Comments: Part of master seminar in document analysis held by Marcus Eichenberger-Liwicki
Related articles: Most relevant | Search more
arXiv:1902.00045 [cs.LG] (Published 2019-01-31)
Gaussian Conditional Random Fields for Classification
arXiv:1708.08591 [cs.LG] (Published 2017-08-29)
EC3: Combining Clustering and Classification for Ensemble Learning
arXiv:1703.08816 [cs.LG] (Published 2017-03-26)
Uncertainty Quantification in the Classification of High Dimensional Data