Top-Down Induction of Regression Trees

Name Top-Down Induction of Regression Trees
Description

The Top-Down Induction of Regression Trees addresses the task of function approximation. Nevertheless the method is quite similar to top-down induction of decision trees.

Given:

  • a set E of instances of the examples language in attribute-value representation
  • for each instance x ∈ E its function value f(x)

The method:

A regression tree is constructed top-down.

  • In each step a test for the actual node is chosen (starting with the root node). The search for good tests is guided by variably defined quality measures, which reward tests that induce partitions which subsume instances with similar function values.
  • As in decision trees in regression trees a test is a mapping from the set of all instances of the underlying example language to a finite set of possible results. There is a successor of the actual node for each possible result of the test.
  • The given set of examples is split with respect to the chosen test.
  • For each successor that does not meet a given acceptance criterion, the procedure is called recursively.
  • Each leaf of a regression tree is marked by a real value. The value is, of course, chosen such, that the actual quality measure for the leaf is maximized. Alternatively a linear function may be attached to leaf nodes, if the attributes describing examples are numerical as well.

When the method of regression trees is invoked, there are often numerical attribute values, and tests may compare the value of certain attributes to appropriate constants.

The aim of regression trees usually is to find a good generalization of the training data, which means good predictions of yet unseen values of the actual function. Another possible aim is to analyze which variables are well suited to predict a target value.
Specialization CART
Dm Step Function Approximation
Method Type Method