|
Bayesian Learning
Publication |
Mitchell/97b: Machine Learning
|
Name |
Bayesian Learning |
Description |
Bayesian Learning constitutes a probabilistic view of learning,
based on Bayes Theorem. The underlying assumption is, that
there is a set of hypotheses, each having a certain probability
of being correct. Receiving more information changes the probabilities
from a learner's point of view. For instance an observation might
contradict a hypothesis, or strengthen the belief in it.
The aim in this setting is to be able to find a hypothesis with highest
probability of being correct, given a specific set of data / piece of
information.
Bayes Theorem:
Let
- ... P be a probability distribution.
- ... hi, i ∈ {1, ..., n} denote a set of hypotheses.
- ... P(hi) denote the probability of hi being
correct in the general case, that is without being given any additional
information. P(hi) is called the prior probability of
hi.
- ... D denote a set of data.
- ... P(D) be the probability of the information denoted by D being
correct in the general case.
- ... P(D | hi) denote the probability of the information D
being correct, given the correctness of hypothesis hi.
- ... P(hi | D) denote the probability of the correctness of
hypothesis hi, given the additional information D.
This is called the posterior probability of hypothesis hi.
Bayes Theorem:
P(h | D) = P(D | h) * P(h) / P(D)
This theorem allows us to find a hypothesis h'with maximum posterior probability,
given the prior probabilities of all hypotheses and the probabilities of D
being correct under the assumption of each single hypothesis being correct:
h' := | max | [ P(D | hi) * P(hi) ] |
| hi | |
|
Dm Step |
Concept Learning
|
Methods |
Naive Bayes
|
|
|