Given bits, with bits numbered , we assume that bit corresponding to the least significant bit. Then the sequence of bits on the left is the binary number equivalent of the decimal number represented by the summation on the right. Since each bit can only take values from the alphabet {0,1}, a string of bits so numbered can represent up to unique patterns. The type representation adopted for this string of bits determines how the string will be interpreted, as described below.