HDFS is a distributed file system designed for large data sets and high throughput. It has a master/slave architecture with a NameNode master that controls file system operations and maps data blocks to DataNodes slaves that store file blocks. HDFS is fault tolerant through data replication across DataNodes and can continue operating if nodes fail since data is still available elsewhere. Future changes may include file append, user quotas, and standby nodes to improve the single point of failure of the NameNode master.