Multiclass Semi-Supervised Boosting and Similarity Learning

Authors
Publication date 2013
Host editors
  • H. Xiong
  • G. Karypis
  • B. Thuraisingham
  • D. Cook
  • X. Wu
Book title IEEE 13th International Conference on Data Mining: ICDM 2013: 7-10 December 2013, Dallas, Texas: proceedings
ISBN
  • 9780768551081
Event IEEE International Conference on Data Mining (ICDM 2013)
Pages (from-to) 1205-1210
Publisher Los Alamitos, CA: IEEE Computer Society
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
In this paper, we consider the multiclass semi-supervised classification problem. A boosting algorithm is proposed to solve the multiclass problem directly. The proposed multiclass approach uses a new multiclass loss function, which includes two terms. The first term is the cost of the multiclass margin and the second term is a regularization term on unlabeled data. The regularization term is used to minimize the inconsistency between the pair wise similarity and the classifier predictions. It assigns the soft labels weighted with the similarity between unlabeled and labeled examples. We then derive a boosting algorithm, named CD-MSSBoost, from the proposed loss function using coordinate gradient descent. The derived algorithm is further used for learning optimal similarity function for a given data. Our experiments on a number of UCI datasets show that CD-MSSBoost outperforms the state-of-the-art methods to multiclass semi-supervised learning.
Document type Conference contribution
Language English
Published at https://doi.org/10.1109/ICDM.2013.108
Permalink to this page
Back