This document provides an overview of data intensive computing and some of the challenges it presents. It discusses how data intensive applications deal with large datasets ranging from terabytes to petabytes in size across various domains. Some of the key challenges in data intensive computing include developing scalable algorithms, metadata management technologies, and distributed file systems that can handle petabytes of data efficiently. Frameworks like MapReduce and cloud computing technologies help address these challenges by providing computation and storage at large scales.