This document discusses information theory and coding techniques. It provides examples and explanations of key concepts such as: - Shannon's theorem, which states that error-free communication is possible if transmission rate does not exceed channel capacity. - Measuring the information content of messages based on probability, with less probable messages containing more information. - Entropy, which is the average information contained in symbols from an information source. - Source coding techniques like fixed-length and variable-length codes to represent symbols with binary codewords. Instantaneous codes ensure unique decodability. - Shannon-Fano and Huffman coding algorithms for designing optimal variable-length codes based on symbol probabilities.