arXiv Analytics

Sign in

arXiv:1402.2300 [cs.LG]AbstractReferencesReviewsResources

Feature and Variable Selection in Classification

Aaron Karper

Published 2014-02-10Version 1

The amount of information in the form of features and variables avail- able to machine learning algorithms is ever increasing. This can lead to classifiers that are prone to overfitting in high dimensions, high di- mensional models do not lend themselves to interpretable results, and the CPU and memory resources necessary to run on high-dimensional datasets severly limit the applications of the approaches. Variable and feature selection aim to remedy this by finding a subset of features that in some way captures the information provided best. In this paper we present the general methodology and highlight some specific approaches.

Comments: Part of master seminar in document analysis held by Marcus Eichenberger-Liwicki
Categories: cs.LG, cs.AI, stat.ML
Related articles: Most relevant | Search more
arXiv:1902.00045 [cs.LG] (Published 2019-01-31)
Gaussian Conditional Random Fields for Classification
arXiv:1708.08591 [cs.LG] (Published 2017-08-29)
EC3: Combining Clustering and Classification for Ensemble Learning
arXiv:1703.08816 [cs.LG] (Published 2017-03-26)
Uncertainty Quantification in the Classification of High Dimensional Data