Chebyshev's Inequality

Let X be a random variable which may be either discrete type or continuous type. If the probability distribution of the random variable X is known, then we can compute its mean E (X) and variance V (X) if these exist and hence we can make the probability statements regarding the E (X) and V (X). The probabilities are given by the area under a frequency curve of the distribution. If the frequency curve is approximately normal curve i.e. bell shaped curve, then the approximate probabilities over certain intervals such asμ±σ,μ± 2σ andμ± 3σ can be obtained as, $$P (\mu \ - \ \sigma \ < \ X \ < \ \mu \ + \ \sigma) \ = \ 0.6826$$ $$P \ (\mu \ - \ 2 \sigma \ < \ X \ < \ \mu \ + \ 2 \sigma) \ = \ 0.9544$$ $$P \ (\mu \ - \ 3 \sigma \ < \ X \ < \ \mu \ + \ 3 \sigma) \ = \ 0.9974$$ $$Where, \ E \ (X) \ = \ \mu \ and \ V \ (X) \ = \ \sigma^2$$ But in many situations, probability or frequency curve differ markedly from a bell - shape. In this case, even if E (X) and V (X) are known, we can not compute the probability such as P ( | X - E (X) |≤ k ). However, we can give a very useful upper or lower bound to such probabilities. Such upper or lower bounds to the probabilities are given by an important theorem in probabilities known as Tchebyshev (1821 - 1894). In fact, the Chebyshev's inequality indicates how the variance V (X) measures the degree of concentration of probability near E (X) =μ. In other words,Chebyshev's inequality provides an upper bound (or lower bound) to the probability of getting a value that deviated from the mean by more than (or less than) k times standard deviations. Statement : Let X be a random variable of discrete type with probability mass function (pmf) p(x) or continious type with probability density function (pdf) f(x) with finite meanμ = E (X) and variance V (X) =σ2. Then for any positive constant K > 0, $$ P \ ( \ | \ X \ - \ \mu \ | \ ≥ \ k \sigma \ ) \ ≤ \ \frac{1}{k^2}$$ $$or, \ P \ ( \ | \ X \ - \ \mu \ | \ < \ k \sigma \ ) \ ≥ \ 1 \ - \ \frac{1}{k^2}$$ $$Where \ E(X) \ = \ \mu \ is \ finite \ and \ variance \ V(X) \ = \ \sigma^2 \ > \ 0 \ and \ < \ \infty.$$

Summary

Let X be a random variable which may be either discrete type or continuous type. If the probability distribution of the random variable X is known, then we can compute its mean E (X) and variance V (X) if these exist and hence we can make the probability statements regarding the E (X) and V (X). The probabilities are given by the area under a frequency curve of the distribution. If the frequency curve is approximately normal curve i.e. bell shaped curve, then the approximate probabilities over certain intervals such asμ±σ,μ± 2σ andμ± 3σ can be obtained as, $$P (\mu \ - \ \sigma \ < \ X \ < \ \mu \ + \ \sigma) \ = \ 0.6826$$ $$P \ (\mu \ - \ 2 \sigma \ < \ X \ < \ \mu \ + \ 2 \sigma) \ = \ 0.9544$$ $$P \ (\mu \ - \ 3 \sigma \ < \ X \ < \ \mu \ + \ 3 \sigma) \ = \ 0.9974$$ $$Where, \ E \ (X) \ = \ \mu \ and \ V \ (X) \ = \ \sigma^2$$ But in many situations, probability or frequency curve differ markedly from a bell - shape. In this case, even if E (X) and V (X) are known, we can not compute the probability such as P ( | X - E (X) |≤ k ). However, we can give a very useful upper or lower bound to such probabilities. Such upper or lower bounds to the probabilities are given by an important theorem in probabilities known as Tchebyshev (1821 - 1894). In fact, the Chebyshev's inequality indicates how the variance V (X) measures the degree of concentration of probability near E (X) =μ. In other words,Chebyshev's inequality provides an upper bound (or lower bound) to the probability of getting a value that deviated from the mean by more than (or less than) k times standard deviations. Statement : Let X be a random variable of discrete type with probability mass function (pmf) p(x) or continious type with probability density function (pdf) f(x) with finite meanμ = E (X) and variance V (X) =σ2. Then for any positive constant K > 0, $$ P \ ( \ | \ X \ - \ \mu \ | \ ≥ \ k \sigma \ ) \ ≤ \ \frac{1}{k^2}$$ $$or, \ P \ ( \ | \ X \ - \ \mu \ | \ < \ k \sigma \ ) \ ≥ \ 1 \ - \ \frac{1}{k^2}$$ $$Where \ E(X) \ = \ \mu \ is \ finite \ and \ variance \ V(X) \ = \ \sigma^2 \ > \ 0 \ and \ < \ \infty.$$

Things to Remember

Let X be a random variable of discrete type with probability mass function (pmf) p(x) or continious type with probability density function (pdf) f(x) with finite meanμ = E (X) and variance V (X) =σ2. Then for any positive constant K > 0,

$$ P \ ( \ | \ X \ - \ \mu \ | \ ≥ \ k \sigma \ ) \ ≤ \ \frac{1}{k^2}$$

$$or, \ P \ ( \ | \ X \ - \ \mu \ | \ < \ k \sigma \ ) \ ≥ \ 1 \ - \ \frac{1}{k^2}$$

$$Where \ E(X) \ = \ \mu \ is \ finite \ and \ variance \ V(X) \ = \ \sigma^2 \ > \ 0 \ and \ < \ \infty.$$

MCQs

No MCQs found.

Subjective Questions

No subjective questions found.

Videos

No videos found.

Chebyshev's Inequality

Chebyshev's Inequality

Introduction :

Let X be a random variable which may be either discrete type or continuous type. If the probability distribution of the random variable X is known, then we can compute its mean E (X) and variance V (X) if these exist and hence we can make the probability statements regarding the E (X) and V (X). The probabilities are given by the area under a frequency curve of the distribution. If the frequency curve is approximately normal curve i.e. bell shaped curve, then the approximate probabilities over certain intervals such asμ±σ,μ± 2σ andμ± 3σ can be obtained as,

$$P (\mu \ - \ \sigma \ < \ X \ < \ \mu \ + \ \sigma) \ = \ 0.6826$$

$$P \ (\mu \ - \ 2 \sigma \ < \ X \ < \ \mu \ + \ 2 \sigma) \ = \ 0.9544$$

$$P \ (\mu \ - \ 3 \sigma \ < \ X \ < \ \mu \ + \ 3 \sigma) \ = \ 0.9974$$

$$Where, \ E \ (X) \ = \ \mu \ and \ V \ (X) \ = \ \sigma^2$$

But in many situations, probability or frequency curve differ markedly from a bell - shape. In this case, even if E (X) and V (X) are known, we can not compute the probability such as P ( | X - E (X) |≤ k ). However, we can give a very useful upper or lower bound to such probabilities. Such upper or lower bounds to the probabilities are given by an important theorem in probabilities known as Tchebyshev (1821 - 1894). In fact, the Chebyshev's inequality indicates how the variance V (X) measures the degree of concentration of probability near E (X) =μ. In other words,Chebyshev's inequality provides an upper bound (or lower bound) to the probability of getting a value that deviated from the mean by more than (or less than) k times standard deviations.

Statement :

Let X be a random variable of discrete type with probability mass function (pmf) p(x) or continious type with probability density function (pdf) f(x) with finite meanμ = E (X) and variance V (X) =σ2. Then for any positive constant K > 0,

$$ P \ ( \ | \ X \ - \ \mu \ | \ ≥ \ k \sigma \ ) \ ≤ \ \frac{1}{k^2}$$

$$or, \ P \ ( \ | \ X \ - \ \mu \ | \ < \ k \sigma \ ) \ ≥ \ 1 \ - \ \frac{1}{k^2}$$

$$Where \ E(X) \ = \ \mu \ is \ finite \ and \ variance \ V(X) \ = \ \sigma^2 \ > \ 0 \ and \ < \ \infty.$$

Proof :

Consider, X is a random variable.

By definition, we have

$$V \ (X) \ = \ \sigma^2 \ = \ E \ [ \ X \ - \ E(X) \ ]^2 \ = \ E \ ( \ X \ - \ \mu \ )^2$$

$$= \ \int_{-\infty}^{\infty} \ ( \ X \ - \ \mu \ )^2 \ f(x) \ dx$$

$$= \ \int_{-\infty}^{\mu \ - \ k\sigma} \ ( \ x \ - \ \mu \ )^2 \ f(x) \ dx \ + \ \int_{\mu \ - \ k\sigma}^{\mu \ + \ k\sigma} \ ( \ x \ - \ \mu \ )^2 \ f(x) \ dx \ + \ \int_{\mu \ + \ k\sigma}^{\infty} \ ( \ x \ - \ \mu \ )^2 \ f(x) \ dx$$

$$\therefore \ \sigma^2 \ ≥ \ \int_{-\infty}^{\mu \ - \ k\sigma} \ ( x \ - \ \mu \ )^2 \ f(x) \ dx \ +\ \int_{\mu \ + \ k\sigma}^{\infty} \ ( \ x \ - \ \mu \ )^2 \ f(x) \ dx$$

$$Because, \ the \ integral \ \int_{\mu \ - \ k\sigma}^{\mu \ + \ k\sigma} \ ( x \ - \ \mu \ )^2 \ f(x) \ dx \ ≥ \ 0$$

Now, for the first integral, x ≤ μ - kσ ⇒μ - x ≥ kσ and for the second integral, x ≥ μ + kσ ⇒ x - μ ≥ kσ. Therefore, (x-μ)2≥ k2σ2.

$$\therefore \ \sigma^2 \ ≥ \ k^2 \sigma^2 \ \left [ \ \int_{-\infty}^{\mu \ - \ k\sigma} \ f(x) \ dx \ + \ \int_{\mu \ + \ k\sigma}^{\infty} \ f(x) \ dx \ \right ]$$

$$or, \ \sigma^2 \ ≥ \ k^2 \sigma^2 \ \left [ \ P ( - \infty \ < \ X \ < \ \mu \ - \ k\sigma ) \ + \ P ( \mu \ + \ k\sigma \ < \ X \ < \ \infty ) \ \right ]$$

$$or, \ \sigma^2 \ ≥ \ k^2 \sigma^2 \ \left [ \ P ( X \ - \ \mu \ < \ - \ k\sigma ) \ + \ P ( X \ - \ \mu \ > \ k\sigma ) \ \right ]$$

$$or, \ \sigma^2 \ ≥ \ k^2 \sigma^2 \ \left [ \ P ( | X \ - \ \mu | \ ≥ \ k\sigma ) \ \right ]$$

$$\therefore \ P( | X \ - \ \mu | \ ≥ \ k \sigma ) \ ≤ \ \frac{1}{k^2}$$

$$Also, \ since, \ P( | X \ - \ \mu | \ ≥ \ k \sigma ) \ + \ P( | X \ - \ \mu | \ ≥ \ k\sigma ) \ = \ 1$$

$$We \ get, \ P( | X \ - \ \mu | \ ≥ \ k\sigma ) \ = \ 1 \ - \ P( | X \ - \ \mu | \ ≥ \ k\sigma ) \ ≥ \ 1 \ - \ \frac{1}{k^2}$$

Bibliography

Sukubhattu N.P. (2013). Probability & Inference - II. Asmita Books Publishers & Distributors (P) Ltd., Kathmandu.

Larson H.J. Introduction to Probability Theory and Statistical Inference. WileyInternational, New York.

Lesson

Inequalities

Subject

Statistics

Grade

Bachelor of Science

Recent Notes

No recent notes.

Related Notes

No related notes.