- Chebyshev's inequality
-
also called Bienaymé-Chebyshev inequalityin probability theory, a theorem that characterizes the dispersion of data away from its mean (average). The general theorem is attributed to the 19th-century Russian mathematician Pafnuty Chebyshev (Chebyshev, Pafnuty Lvovich), though credit for it should be shared with the French mathematician Irénée-Jules Bienaymé, whose (less general) 1853 proof predated Chebyshev's by 14 years.Chebyshev's inequality puts an upper bound on the probability that an observation should be far from its mean. It requires only two minimal conditions: (1) that the underlying distribution (distribution function) have a mean and (2) that the average size of the deviations away from this mean (as gauged by the standard deviation) not be infinite. Chebyshev's inequality then states that the probability that an observation will be more than k standard deviations from the mean is at most 1/k2. Chebyshev used the inequality to prove his version of the law of large numbers.Unfortunately, with virtually no restriction on the shape of an underlying distribution, the inequality is so weak as to be virtually useless to anyone looking for a precise statement on the probability of a large deviation. To achieve this goal, people usually try to justify a specific error distribution, such as the normal distribution as proposed by the German mathematician Carl Friedrich Gauss (Gauss, Carl Friedrich). Gauss also developed a tighter bound, 4/9k2 (for k > 2/√3), on the probability of a large deviation by imposing the natural restriction that the error distribution decline symmetrically from a maximum at 0.The difference between these values is substantial. According to Chebyshev's inequality, the probability that a value will be more than two standard deviations from the mean (k = 2) cannot exceed 25 percent. Gauss's bound is 11 percent, and the value for the normal distribution is just under 5 percent. Thus, it is apparent that Chebyshev's inequality is useful only as a theoretical tool for proving generally applicable theorems, not for generating tight probability bounds.Richard Routledge
* * *
Universalium. 2010.