This document discusses number representation systems used in computers, including binary, decimal, octal, and hexadecimal. It provides examples of converting between these different bases. Specifically, it covers:
1) Converting between decimal, binary, octal, and hexadecimal using positional notation.
2) Signed integer representation in binary, including sign-magnitude, one's complement, and two's complement. Examples are given of converting positive and negative decimals to these binary representations.
3) Storing integer, character, and floating point numbers in binary. Twos complement is described as the most common method for signed integer representation.