This document reviews evaluation metrics for data classification. It discusses accuracy and other alternative metrics like precision, recall, F-measure, ROC curves, and cost curves. Accuracy is widely used but has weaknesses when dealing with imbalanced data. The paper also discusses metrics designed specifically for discriminating optimal solutions during generative classifier training like OAERP and a hybrid evaluation metric. It suggests five aspects to consider when constructing a new evaluation metric: informativeness, discriminability, bias handling, scalability and interpretability. The review covers over 35 references on evaluation metrics for data classification.