You are given a data stream that has been compressed to a length of 100,000 bits, and told that it is the result of running an "ideal" entropy coder on a sequence of data. You are also told that the original data consists of samples of a continuous waveform, quantized to 2 bits per sample. The probabilities of the uncompressed values are as follows: 00 1/2 01 3/8 10 1/16 11 1/16. What (approximately) was the length of the uncompressed signal?