This document discusses and compares several different probabilistic models for sequence labeling tasks, including Hidden Markov Models (HMMs), Maximum Entropy Markov Models (MEMMs), and Conditional Random Fields (CRFs).
It provides mathematical formulations of HMMs, describing how to calculate the most likely label sequence using the Viterbi algorithm. It then introduces MEMMs, which address some limitations of HMMs by incorporating arbitrary, overlapping features. CRFs are presented as an improvement over MEMMs that models the conditional probability of labels given observations, avoiding the label bias problem of MEMMs. The document concludes by describing how to train CRF models using generalized iterative scaling.