This document discusses research on automated essay evaluation using natural language processing. It provides background on previous systems for automated essay scoring like Project Essay Grader (PEG) from the 1960s and more recent systems like e-Rater, IntelliMetric, and Intelligent Essay Assessors. The researchers extracted features from essays like word count, sentence count, spelling, and part-of-speech to train machine learning models. They achieved correlation scores between 0.86-0.87 when comparing predicted scores to human scores, showing the models can perform at similar reliability levels to human graders. The researchers conclude the models could be improved by incorporating features like parse trees and accounting for different essay prompts.