Online Feature Selection by Adaptive Sub-gradient Methods

Autor: Hao Wang, Yang Gao, Tingting Zhai, Frédéric Koriche
Přispěvatelé: Faculty of Physics and Electronic Science (PES), Hubei University of Science and Technology, Centre de Recherche en Informatique de Lens (CRIL), Université d'Artois (UA)-Centre National de la Recherche Scientifique (CNRS), China Information Technology Security Evaluation Center
Rok vydání: 2019
Předmět:
Zdroj: Machine Learning and Knowledge Discovery in Databases ISBN: 9783030109271
ECML/PKDD (2)
Machine Learning and Knowledge Discovery in Databases-European Conference (ECML-PKDD)
Machine Learning and Knowledge Discovery in Databases-European Conference (ECML-PKDD), 2018, Unknown, Ireland
DOI: 10.1007/978-3-030-10928-8_26
Popis: The overall goal of online feature selection is to iteratively select, from high-dimensional streaming data, a small, “budgeted” number of features for constructing accurate predictors. In this paper, we address the online feature selection problem using novel truncation techniques for two online sub-gradient methods: Adaptive Regularized Dual Averaging (ARDA) and Adaptive Mirror Descent (AMD). The corresponding truncation-based algorithms are called B-ARDA and B-AMD, respectively. The key aspect of our truncation techniques is to take into account the magnitude of feature values in the current predictor, together with their frequency in the history of predictions. A detailed regret analysis for both algorithms is provided. Experiments on six high-dimensional datasets indicate that both B-ARDA and B-AMD outperform two advanced online feature selection algorithms, OFS and SOFS, especially when the number of selected features is small. Compared to sparse online learning algorithms that use \(\ell _1\) regularization, B-ARDA is superior to \(\ell _1\)-ARDA, and B-AMD is superior to Ada-Fobos. Code related to this paper is available at: https://github.com/LUCKY-ting/online-feature-selection.
Databáze: OpenAIRE