This document provides an introduction to Hadoop, including its programming model, distributed file system (HDFS), and MapReduce framework. It discusses how Hadoop uses a distributed storage model across clusters, racks, and data nodes to store large datasets in a fault-tolerant manner. It also summarizes the core components of Hadoop, including HDFS for storage, MapReduce for processing, and YARN for resource management. Hadoop provides a scalable platform for distributed processing and analysis of big data.