This document provides an overview of HDFS (Hadoop Distributed File System). It discusses key concepts such as the client-server model, storing large datasets across commodity hardware, and how HDFS is designed to be fault tolerant. The document also describes the main daemons (NameNode, DataNode, Secondary NameNode), how files are split into blocks and replicated across DataNodes, and the roles of the NameNode and Secondary NameNode in managing metadata and checkpoints.