This document compares the performance of Pointwise Mutual Information (PMI) and Latent Semantic Analysis (LSA) on several semantic similarity tasks when trained on different sized corpora. It finds that PMI trained on a large Wikipedia corpus outperforms LSA trained on a smaller subset, showing that simpler models can perform better when given more data. It also compares PMI trained on Wikipedia to other publicly available measures of semantic relatedness, finding PMI performs competitively. The document concludes that simple, scalable models that leverage large amounts of data show promise for modeling semantics.