TY - GEN
T1 - Linear dimensionality reduction for multi-label classification
AU - Ji, Shuiwang
AU - Ye, Jieping
N1 - Copyright:
Copyright 2020 Elsevier B.V., All rights reserved.
PY - 2009
Y1 - 2009
N2 - Dimensionality reduction is an essential step in high-dimensional data analysis. Many dimensionality reduction algorithms have been applied successfully to multi-class and multi-label problems. They are commonly applied as a separate data preprocessing step before classification algorithms. In this paper, we study a joint learning framework in which we perform dimensionality reduction and multi-label classification simultaneously. We show that when the least squares loss is used in classification, this joint learning decouples into two separate components, i.e., dimensionality reduction followed by multi-label classification. This analysis partially justifies the current practice of a separate application of dimensionality reduction for classification problems. We extend our analysis using other loss functions, including the hinge loss and the squared hinge loss. We further extend the formulation to the more general case where the input data for different class labels may differ, overcoming the limitation of traditional dimensionality reduction algorithms. Experiments on benchmark data sets have been conducted to evaluate the proposed joint formulations.
AB - Dimensionality reduction is an essential step in high-dimensional data analysis. Many dimensionality reduction algorithms have been applied successfully to multi-class and multi-label problems. They are commonly applied as a separate data preprocessing step before classification algorithms. In this paper, we study a joint learning framework in which we perform dimensionality reduction and multi-label classification simultaneously. We show that when the least squares loss is used in classification, this joint learning decouples into two separate components, i.e., dimensionality reduction followed by multi-label classification. This analysis partially justifies the current practice of a separate application of dimensionality reduction for classification problems. We extend our analysis using other loss functions, including the hinge loss and the squared hinge loss. We further extend the formulation to the more general case where the input data for different class labels may differ, overcoming the limitation of traditional dimensionality reduction algorithms. Experiments on benchmark data sets have been conducted to evaluate the proposed joint formulations.
UR - http://www.scopus.com/inward/record.url?scp=77953179079&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=77953179079&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:77953179079
SN - 9781577354260
T3 - IJCAI International Joint Conference on Artificial Intelligence
SP - 1077
EP - 1082
BT - IJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence
PB - International Joint Conferences on Artificial Intelligence
T2 - 21st International Joint Conference on Artificial Intelligence, IJCAI 2009
Y2 - 11 July 2009 through 16 July 2009
ER -