The document discusses evaluating search system performance. It begins with an overview of the Cranfield paradigm for evaluation, which involves test collections of documents, queries and relevance judgments. Common evaluation metrics are then explained, such as precision, recall, F-measure, mean average precision and normalized discounted cumulative gain. Several evaluation initiatives are also described, including TREC, NTCIR and CLEF. The document notes that the Cranfield paradigm focuses on system-oriented measures and effectiveness, but that users should be more central to evaluation. It suggests that relevance is subjective and user experience is also important.