Exploiting feature relationships towards stable feature selection


Autoria(s): Kamkar, Iman; Gupta, Sunil; Phung, Dinh; Venkatesh, Svetha
Contribuinte(s)

Gaussier, Eric

Cao, Longbing

Gallinari, Patrick

Kwok, James

Pasi, Gabriella

Zaiane, Osmar

Data(s)

01/01/2015

Resumo

Feature selection is an important step in building predictive models for most real-world problems. One of the popular methods in feature selection is Lasso. However, it shows instability in selecting features when dealing with correlated features. In this work, we propose a new method that aims to increase the stability of Lasso by encouraging similarities between features based on their relatedness, which is captured via a feature covariance matrix. Besides modeling positive feature correlations, our method can also identify negative correlations between features. We propose a convex formulation for our model along with an alternating optimization algorithm that can learn the weights of the features as well as the relationship between them. Using both synthetic and real-world data, we show that the proposed method is more stable than Lasso and many state-of-the-art shrinkage and feature selection methods. Also, its predictive performance is comparable to other methods.

Identificador

http://hdl.handle.net/10536/DRO/DU:30081985

Idioma(s)

eng

Publicador

IEEE

Relação

http://dro.deakin.edu.au/eserv/DU:30081985/kamkar-exploitingfeature-2015.pdf

http://dro.deakin.edu.au/eserv/DU:30081985/kamkar-exploitingfeature-evid-2015.pdf

http://www.dx.doi.org/10.1109/DSAA.2015.7344859

Direitos

2015, IEEE

Palavras-Chave #stability #Lasso #correlated features #prediction
Tipo

Conference Paper