Quasi-Ordering of Hypothesis Languages

Description:

In the concept learning/ILP setting we are looking for a concept, more general than any positive example, and more specific than any concept, covering positive as well as negative examples. So it is a good idea, to structure the hypotheses by their generality, and to generate more general or more specific hypotheses stepwise. The search in a structured search space of all hypotheses can be pruned, which is not possible in an unstructured search space.

In his book "Machine Learning" Tom Mitchell proposed a quasi-ordering for attribute-value hypothesis languages (partial ordering of the set of all hypotheses) based on the more-special-than (or more-general-than) relation:

A concept c1 is more special than a concept c2, if and only if every example covered by c1 is covered by c2 as well. A concept c1 is strictly more special than c2, if it is more special but distinct from c2, which means that there is at least one instance covered by c2, that is not covered by c1 as well.

A concept c1 is more general than c2, if c2 is more special than c1. Concept c1 is strictly more general than c2, if c2 is strictly more specific than c1.

If concepts are characterized by attribute values, which can be arranged in a hierarchy according to a more-special-than relation, the search space (which can be seen as a cartesian product of the attribute value sets) can always be (partially) ordered.

Publications: Mitchell/97b: Machine Learning