Answer:
[tex]1001\times 10^{-7}\%[/tex]
Step-by-step explanation:
The receiver makes a decoding error only if three bits or two bits of a 3-bit string are sent wrongly.
Let's call
p = probability that one bit was sent incorrectly.
Since bit errors occur random and independently of each other, the probability that 3 bits are sent incorrectly is
[tex]p\times p\times p=p^3[/tex]
Similarly, the probability that 2 bits are sent incorrectly is
[tex]p\times p=p^2[/tex]
The probability that 3 or 2 bits are sent incorrectly is
[tex]p^3+p^2[/tex]
So, all we have to do now is compute p.
Let x be the number of bits incorrectly transmitted per unit of time.
Since the channel operates at 3 Mbps (3,000,000 bits per second) and has a bit error rate of 0.001, then
[tex]BitErrorRate=\frac{errors}{total bits}=\frac{x}{3,000,000}=0.001[/tex]
and
x = 3000
This means that for every 3 million bits transmitted, 3000 are wrong.
So, the probability p that one bit is incorrect when transmitted is
[tex]p=\frac{3000}{3,000,000}=0.001[/tex]
(Remark: When the probability is measured as a number between 0 and 1, it can be shown that the bit error rate and the probability of sending one bit incorrectly are the same)
Hence the probability that the receiver makes a decoding error is
[tex]0.001^3+0.001^2=1001\times 10^{-9}[/tex]
or in % notation
[tex]1001\times 10^{-7}\%[/tex]