Learning characteristic decision trees

Paul Davidsson

Research output: Chapter in Book/Report/Conference proceedingPaper in conference proceedingpeer-review

Abstract

Decision trees constructed by ID3-like algorithms suffer from an inability of detecting instances of categories not present in the set of training examples, i.e., they are discriminative representations. Instead, such instances are assigned to one of the classes actually present in the training set, resulting in undesired misclassifications. Two methods of reducing this problem by learning characteristic representations are presented. The central idea behind
both methods is to augment each leaf of the decision tree with a subtree containing additional information concerning each feature’s values in that leaf. This is done by computing two limits (lower and upper) for every feature from the training instances belonging to the
leaf. A subtree is then constructed from these limits that tests every feature; if the value is below the lower limit or above the upper limit for some feature, the instancewill be rejected,
i.e., regarded as belonging to a novel class. This subtree is then appended to the leaf. The first method presented corresponds to creating a maximumspecific description, whereas the second is a novel method that makes use of the information about the statistical distribution of the feature values that can be extracted from the training examples. An important property of the novel method is that the degree of generalization can be controlled. The methods
are evaluated empirically in two different domains, the Iris classification problem and a novel coin classification problem. It is concluded that the dynamical properties of the second method makes it preferable in most applications. Finally, we argue that this method is
very general in that it, in principle, can be applied to any empirical learning algorithm.
Original languageEnglish
Title of host publicationAI, 1995 : Proceedings of the Eighth Australian Joint Conference on Artificial Intelligence, Canberra, Australia 13 - 17 November 1995
Publication statusPublished - 1995
Externally publishedYes

Subject classification (UKÄ)

  • Computer Sciences

Free keywords

  • decision trees
  • algorithms

Fingerprint

Dive into the research topics of 'Learning characteristic decision trees'. Together they form a unique fingerprint.

Cite this