• DocumentCode
    2138122
  • Title

    Exploring error-sensitive attributes and branches of decision trees

  • Author

    Wu, Wenchuan

  • Author_Institution
    Fac. of Eng. & Inf. Technol., Univ. of Technol., Sydney, NSW, Australia
  • fYear
    2013
  • fDate
    23-25 July 2013
  • Firstpage
    929
  • Lastpage
    934
  • Abstract
    Decision trees have a reputation of being efficient and illustrative in classification learning, and majority of the research effort has been focused on making classification improvement in a head-on style with wide-range research topics, such as tree algorithm development and refinement, attribute selection and prioritization, sampling technique improvement, and the addition of cost matrix and other performance-enhancing factors. One less commonly studied topic is about the characteristics of classification errors and how they may be associated with specific attributes due to correlation or causation, and within what value ranges on such attributes when pattern are most likely. This research intends to study this dim area in a sort-of reverse and forensic style as part of post-classification analysis, to analyze the patterns and relationship between errors and attributes, to explore how attributes´ risk level in error may play a role in leading to more risky, more error-prone decision tree branches or decision paths. Possible benefits from this study would include raising data stakeholders´ awareness of such specific error-sensitive attributes and decision paths, to facilitate better understanding of possible causes and impact of errors and the development of more effective error-reduction measures customized to suit the specific patterns and individual datasets. Such emphasis on highlighting the specific error-sensitive attributes and decision branches within individual datasets is a reflection of our observation which shared by others - “additional domain-specific knowledge, external to the training set, must be employed to estimate the noise level (... and) the underlying model´s complexity ... (because) knowledge-poor tree induction algorithms do not exploit such information.” [2].
  • Keywords
    decision trees; learning (artificial intelligence); matrix algebra; pattern classification; attribute selection; classification learning; cost matrix; decision path; decision trees; error-reduction measures; error-sensitive attribute; forensic style; knowledge-poor tree induction algorithm; noise level; post-classification analysis; reverse style; sampling technique improvement; tree algorithm development; tree algorithm refinement; Accuracy; Data models; Decision trees; Error analysis; Magnesium; Measurement uncertainty; Radiation detectors; decision tree; error-sensitive attribute; error-sensitive tree branch; feature selection; post-classification analysis;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Natural Computation (ICNC), 2013 Ninth International Conference on
  • Conference_Location
    Shenyang
  • Type

    conf

  • DOI
    10.1109/ICNC.2013.6818109
  • Filename
    6818109