• DocumentCode
    786194
  • Title

    Best-case results for nearest-neighbor learning

  • Author

    Salzberg, Steven ; Delcher, Arthur L. ; Heath, David ; Kasif, Simon

  • Author_Institution
    Dept. of Comput. Sci., Johns Hopkins Univ., Baltimore, MD, USA
  • Volume
    17
  • Issue
    6
  • fYear
    1995
  • fDate
    6/1/1995 12:00:00 AM
  • Firstpage
    599
  • Lastpage
    608
  • Abstract
    Proposes a theoretical model for analysis of classification methods, in which the teacher knows the classification algorithm and chooses examples in the best way possible. The authors apply this model using the nearest-neighbor learning algorithm, and develop upper and lower bounds on sample complexity for several different concept classes. For some concept classes, the sample complexity turns out to be exponential even using this best-case model, which implies that the concept class is inherently difficult for the NN algorithm. The authors identify several geometric properties that make learning certain concepts relatively easy. Finally the authors discuss the relation of their work to helpful teacher models, its application to decision tree learning algorithms, and some of its implications for experimental work
  • Keywords
    geometry; learning (artificial intelligence); neural nets; pattern classification; best-case results; classification methods analysis; concept classes; decision tree learning; geometric properties; nearest-neighbor learning; sample complexity; Algorithm design and analysis; Character recognition; Classification algorithms; Computer Society; Computer science; Decision trees; Intelligent systems; Neural networks; Pattern recognition; Speech processing;
  • fLanguage
    English
  • Journal_Title
    Pattern Analysis and Machine Intelligence, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0162-8828
  • Type

    jour

  • DOI
    10.1109/34.387506
  • Filename
    387506