This document discusses web crawling and text classification. It describes how web crawlers copy pages for search engines to index and introduces a project to crawl the Figshare website. Requirements for the project include Python, Anaconda, NLTK, and Beautiful Soup. Beautiful Soup is then explained as a tool to pull data from HTML and XML. The document also covers what text classification is and how support vector classifiers can perform multi-class text classification tasks with good accuracy. It presents that support vector classifiers achieved an accuracy of 85% on this project's classification model involving data preparation, review analysis, and classification techniques.