This document provides an introduction to Hadoop and MapReduce. It describes how Hadoop is composed of MapReduce and HDFS. MapReduce is a programming model that allows processing of large datasets in a distributed, parallel manner. It works by breaking the processing into mappers that perform filtering and sorting, and reducers that perform summarization. HDFS provides a distributed file system that stores data across clusters of machines. An example of word count using MapReduce is described to illustrate how it works.