• DocumentCode
    457255
  • Title

    A New Data Selection Principle for Semi-Supervised Incremental Learning

  • Author

    Zhang, Rong ; Rudnicky, Alexander I.

  • Author_Institution
    Language Technol. Inst., Carnegie Mellon Univ., Pittsburgh, PA
  • Volume
    2
  • fYear
    0
  • fDate
    0-0 0
  • Firstpage
    780
  • Lastpage
    783
  • Abstract
    Current semi-supervised incremental learning approaches select unlabeled examples with predicted high confidence for model re-training. We show that for many applications this data selection strategy is not correct. This is because the confidence score is primarily a metric to measure the classification correctness on a particular example, rather than one to measure the example´s contribution to the training of an improved model, especially in the case that the information used in the confidence annotator is correlated with that generated by the classifier. To address this problem, we propose a performance-driven principle for unlabeled data selection in which only the unlabeled examples that help to improve classification accuracy are selected for semi-supervised learning. Encouraging results are presented for a variety of public benchmark datasets
  • Keywords
    learning (artificial intelligence); pattern classification; confidence annotator; confidence score; data selection principle; performance-driven principle; semisupervised incremental learning; unlabeled data selection; Bridges; Degradation; Image retrieval; Information analysis; Information retrieval; Particle measurements; Predictive models; Semisupervised learning; Speech recognition; Variable structure systems;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Pattern Recognition, 2006. ICPR 2006. 18th International Conference on
  • Conference_Location
    Hong Kong
  • ISSN
    1051-4651
  • Print_ISBN
    0-7695-2521-0
  • Type

    conf

  • DOI
    10.1109/ICPR.2006.115
  • Filename
    1699321