← Back to home

Markov'S Inequality

Alright. Let's get this over with. You want to know about Markov's inequality. It's a concept in probability theory, which is, for the most part, a rather dreary subject, but sometimes even the most mundane ideas can offer a sliver of… utility. This one gives you an upper bound on the probability that a non-negative random variable will exceed a certain positive constant. Think of it as a grim prediction, a ceiling on how bad things could get, based on the average. It’s tight, they say. Meaning, for any given constant you pick, there’s a scenario where the inequality becomes an exact equality. Because of course there is.

It’s named after Andrey Markov, a Russian mathematician. Though, apparently, his teacher, Pafnuty Chebyshev, touched on it first. Some analysts, in their infinite need to complicate things, call it the first Chebyshev's inequality, reserving the "second" for another, more familiar inequality. And then there’s Irénée-Jules Bienaymé, who apparently also had a hand in it. It’s a tangled web, as most things tend to be.

These inequalities, Markov's and its ilk, they bridge the gap between probabilities and expectations. They offer bounds for the cumulative distribution function of a random variable. Often, these bounds are loose. But useful. Sometimes. It can also be used to put an upper limit on the expectation of a non-negative random variable, based on its distribution. Which, I suppose, is something.

Statement

So, here’s the deal. If you have a non-negative random variable, let’s call it X, and a positive constant a. The probability that X will be greater than or equal to a is bounded. It’s at most the expectation of X divided by a.

P(Xa)E(X)a\operatorname {P} (X\geq a)\leq {\frac {\operatorname {E} (X)}{a}}

Now, if the expectation of X is positive, which it usually is if we're bothering with this, we can play with the constant. Let a be ã times the expectation of X, where ã is also positive. Then the inequality takes on a slightly different, perhaps more elegant, form:

P(Xa~E(X))1a~\operatorname {P} (X\geq {\tilde {a}}\cdot \operatorname {E} (X))\leq {\frac {1}{\tilde {a}}}

In the more abstract realm of measure theory, it’s a bit more formal. If you have a measure space (X, Σ, μ), and a measurable function f that maps to the extended real number line, and ε is a positive number, then:

μ({xX:f(x)ε})1εXfdμ\mu (\{x\in X:|f(x)|\geq \varepsilon \})\leq {\frac {1}{\varepsilon }}\int _{X}|f|\,d\mu

This version, the measure-theoretic one, is sometimes what they mean by Chebyshev's inequality. Go figure.

Extended version for nondecreasing functions

There’s an extension. If φ is a nondecreasing, non-negative function, and X is a random variable (it doesn't have to be non-negative here, though the original inequality does), and φ(a) is positive, then:

P(Xa)E(φ(X))φ(a)\operatorname {P} (X\geq a)\leq {\frac {\operatorname {E} (\varphi (X))}{\varphi (a)}}

A direct consequence of this is the one involving higher moments. For any positive integer n, if you’re looking at the n-th moment of X (assuming it’s defined and supported on non-zero values):

P(Xa)E(Xn)an\operatorname {P} (|X|\geq a)\leq {\frac {\operatorname {E} (|X|^{n})}{a^{n}}}

The uniformly randomized Markov's inequality

This is where it gets… interesting. If X is a non-negative random variable and a is a positive constant, and you introduce U, a random variable uniformly distributed on [0, 1] that's independent of X. Then:

P(XUa)E(X)a\operatorname {P} (X\geq Ua)\leq {\frac {\operatorname {E} (X)}{a}}

Since U is almost certainly less than one, this bound is tighter than the original Markov's inequality. And here’s the kicker: you can't replace U with any constant smaller than one and expect to maintain the general improvement. It means deterministic enhancements to Markov's inequality are generally not possible. While the original inequality holds with equality for distributions concentrated on {0, a}, this randomized version achieves equality for any distribution bounded on [0, a]. It’s a subtle, almost elegant, twist.

Proofs

Let’s look at how this is derived. We’ll separate the probability space case from the more general measure space one, because, frankly, most people find the former less… intimidating.

Intuition

Consider the expectation of X. It’s a weighted average of all possible values. We can break it down based on whether X is less than a or greater than or equal to a:

E(X)=P(X<a)E(XX<a)+P(Xa)E(XXa)\operatorname {E} (X)=\operatorname {P} (X<a)\cdot \operatorname {E} (X|X<a)+\operatorname {P} (X\geq a)\cdot \operatorname {E} (X|X\geq a)

Now, let's analyze the components.

  • Property 1: P(X<a)E(XX<a)0\operatorname {P} (X<a)\cdot \operatorname {E} (X\mid X<a)\geq 0 Since X is non-negative, its conditional expectation given X < a must also be non-negative. Probabilities are always non-negative, so their product is too. Simple.
  • Property 2: P(Xa)E(XXa)aP(Xa)\operatorname {P} (X\geq a)\cdot \operatorname {E} (X\mid X\geq a)\geq a\cdot \operatorname {P} (X\geq a) If we're conditioning on X ≥ a, then the expected value of X under this condition must be at least a. That is, E(X | X ≥ a) ≥ a. Multiply both sides by the probability P(X ≥ a), and you get this. It’s logical; if all the values you're considering are at least a, their average will also be at least a.

Putting it together: E(X)P(Xa)E(XXa)aP(Xa)\operatorname {E} (X)\geq \operatorname {P} (X\geq a)\cdot \operatorname {E} (X|X\geq a)\geq a\cdot \operatorname {P} (X\geq a) From this chain, the inequality emerges: P(Xa)E(X)a\operatorname {P} (X\geq a)\leq {\frac {\operatorname {E} (X)}{a}} It’s not exactly rocket science, but it holds.

Probability-theoretic proof

Method 1: We start with the definition of expectation for a probability density function f(x):

E(X)=xf(x)dx\operatorname {E} (X)=\int _{-\infty }^{\infty }xf(x)\,dx Since X is non-negative, the integral is only over the non-negative range:

E(X)=0xf(x)dx\operatorname {E} (X)=\int _{0}^{\infty }xf(x)\,dx Now, we can split this integral:

E(X)=0axf(x)dx+axf(x)dx\operatorname {E} (X)=\int _{0}^{a}xf(x)\,dx+\int _{a}^{\infty }xf(x)\,dx The first part, from 0 to a, is non-negative. The second part, from a to infinity, is where x is at least a. So, we can establish a lower bound:

E(X)axf(x)dx\operatorname {E} (X)\geq \int _{a}^{\infty }xf(x)\,dx And since x ≥ a in this integral, we can further bound it:

axf(x)dxaaf(x)dx\int _{a}^{\infty }xf(x)\,dx\geq \int _{a}^{\infty }af(x)\,dx Pulling out the constant a:

aaf(x)dxa\int _{a}^{\infty }f(x)\,dx And that integral, ∫a f(x) dx, is precisely P(X ≥ a). So, we have:

E(X)aP(Xa)\operatorname {E} (X)\geq a\operatorname {P} (X\geq a) Divide by a (since a > 0), and you get the inequality.

Method 2: This uses indicator random variables. Let IE be 1 if event E occurs, and 0 otherwise. The indicator for X ≥ a is I(X≥a). For a > 0, we have the relationship:

aI(Xa)XaI_{(X\geq a)}\leq X This is because if X < a, then I(X≥a) is 0, making the left side 0, which is less than or equal to X. If X ≥ a, then I(X≥a) is 1, making the left side a, which is less than or equal to X. Now, take the expectation of both sides. Since expectation is monotonic:

E(aI(Xa))E(X)\operatorname {E} (aI_{(X\geq a)})\leq \operatorname {E} (X) Using linearity of expectation on the left side:

aE(I(Xa))=a(1P(Xa)+0P(X<a))=aP(Xa)a\operatorname {E} (I_{(X\geq a)}) = a(1 \cdot \operatorname {P} (X\geq a) + 0 \cdot \operatorname {P} (X<a)) = a\operatorname {P} (X\geq a) So, we arrive at:

aP(Xa)E(X)a\operatorname {P} (X\geq a)\leq \operatorname {E} (X) And again, dividing by a yields the result.

Measure-theoretic proof

Assume f is non-negative. We can define a function s(x):

s(x)={ε,if f(x)ε0,if f(x)<εs(x)= \begin{cases} \varepsilon, & \text{if } f(x) \geq \varepsilon \\ 0, & \text{if } f(x) < \varepsilon \end{cases} Clearly, 0 ≤ s(x)f(x). By the properties of the Lebesgue integral:

Xf(x)dμXs(x)dμ\int _{X}f(x)\,d\mu \geq \int _{X}s(x)\,d\mu The integral of s(x) is simply ε times the measure of the set where f(x) is greater than or equal to ε:

Xs(x)dμ=εμ({xX:f(x)ε})\int _{X}s(x)\,d\mu = \varepsilon \mu (\{x\in X:\,f(x)\geq \varepsilon \}) So:

Xf(x)dμεμ({xX:f(x)ε})\int _{X}f(x)\,d\mu \geq \varepsilon \mu (\{x\in X:\,f(x)\geq \varepsilon \}) Since ε > 0, we can divide by it:

μ({xX:f(x)ε})1εXfdμ\mu (\{x\in X:\,f(x)\geq \varepsilon \})\leq {\frac {1}{\varepsilon }}\int _{X}f\,d\mu This covers the general case.

Discrete case

For a discrete random variable X taking non-negative integer values. Let a be a positive integer. Consider a * P(X > a):

aP(X>a)=aP(X=a+1)+aP(X=a+2)+aP(X=a+3)+a\operatorname {P} (X>a) = a\operatorname {P} (X=a+1)+a\operatorname {P} (X=a+2)+a\operatorname {P} (X=a+3)+\dots Now, we can establish an inequality: aP(X=a)+(a+1)P(X=a+1)+(a+2)P(X=a+2)+\leq a\operatorname {P} (X=a)+(a+1)\operatorname {P} (X=a+1)+(a+2)\operatorname {P} (X=a+2)+\dots This is because aa, a < a+1, a < a+2, and so on. And this sum is less than or equal to: 1P(X=1)+2P(X=2)+3P(X=3)++aP(X=a)+(a+1)P(X=a+1)+\leq 1\operatorname {P} (X=1)+2\operatorname {P} (X=2)+3\operatorname {P} (X=3)+\dots + a\operatorname {P} (X=a)+(a+1)\operatorname {P} (X=a+1)+\dots This entire sum is the definition of the expectation E(X). So, a P(X > a) ≤ E(X). Dividing by a gives the result.

Corollaries

Markov's inequality is the foundation for other, sometimes more useful, inequalities.

Chebyshev's inequality

Chebyshev's inequality uses the variance to bound the probability of a random variable straying far from its mean. It states that for any a > 0:

P(XE(X)a)Var(X)a2\operatorname {P} (|X-\operatorname {E} (X)|\geq a)\leq {\frac {\operatorname {Var} (X)}{a^{2}}} Here, Var(X) is the variance, defined as E[(X - E(X))2]. How does this follow from Markov's? Simple. Apply Markov's inequality to the random variable (X - E(X))2 and the constant a2.

P((XE(X))2a2)E((XE(X))2)a2\operatorname {P} ((X-\operatorname {E} (X))^{2}\geq a^{2})\leq {\frac {\operatorname {E} ((X-\operatorname {E} (X))^{2})}{a^{2}}} The left side is equivalent to P(|X - E(X)| ≥ a), and the numerator on the right is the variance. Thus:

P(XE(X)a)Var(X)a2\operatorname {P} (|X-\operatorname {E} (X)|\geq a) \leq \frac{\operatorname {Var} (X)}{a^{2}} It’s a neat trick, really.

Other corollaries

  • The "monotonic" result I mentioned earlier: P(Xa)=P(φ(X)φ(a))MIE(φ(X))φ(a)\operatorname {P} (|X|\geq a)=\operatorname {P} {\big (}\varphi (|X|)\geq \varphi (a){\big )}\,{\overset {\underset {\mathrm {MI} }{}}{\leq }}\,{\frac {\operatorname {E} (\varphi (|X|))}{\varphi (a)}} This just applies Markov's inequality using the function φ.

  • For a non-negative random variable X, the quantile function QX satisfies: QX(1p)E(X)pQ_{X}(1-p)\leq {\frac {\operatorname {E} (X)}{p}} The proof involves p ≤ P(X ≥ QX(1-p)), which then uses Markov's inequality.

  • For a self-adjoint matrix-valued random variable M and a positive definite matrix A (A ≻ 0), we have: P(MA)tr(E(M)A1)\operatorname {P} (M\npreceq A)\leq \operatorname {tr} (\operatorname {E} (M)A^{-1}) This is a more advanced extension, proved similarly.

Examples

Let's consider income. Assuming income is never negative, Markov's inequality tells us that no more than 10% of the population can earn more than 10 times the average income. If the average income is, say, 50,000,thenatmost1050,000, then at most 10% of people can earn over 500,000. It’s a rough estimate, but it’s something.

Another straightforward example: Suppose Andrew makes, on average, 4 mistakes on his Statistics tests. What's the best upper bound for the probability that he makes at least 10 mistakes? Using Markov's inequality:

P(X10)E(X)10=410=0.4\operatorname {P} (X\geq 10)\leq {\frac {\operatorname {E} (X)}{10}} = {\frac {4}{10}} = 0.4 So, the probability is at most 0.4. It doesn't mean he will make 10 mistakes, or even that the probability is that high, but it's the absolute ceiling based on the given information.

See also

  • Paley–Zygmund inequality: This provides a corresponding lower bound, which is often more difficult to establish.
  • Concentration inequality: A broader category of inequalities that bound the probability of a random variable deviating from its expected value. Markov's is a basic member of this family.