Bayesian learning methods calculate explicit probabilities for hypotheses. They allow each training example to incrementally increase or decrease a hypothesis' probability. Prior knowledge can be combined with data to determine final probabilities. Bayesian methods can accommodate probabilistic hypotheses and classify instances by combining hypothesis predictions weighted by probabilities. The minimum description length principle provides a Bayesian perspective on Occam's razor, preferring shorter hypotheses that minimize description length.