II. SIMPLE EXPLANATIONS OF KEY IDEAS IN ORDINAL OPTIMIZATION




II.3. THE CHERNOFF BOUND EXPLAINED

We are interested in finding the probability of a random variable x exceeding certain value V, i.e., P(x >= V). Let us define another random variable YV such that

YV = 1 if x >= V
YV = 0 if x < V

Notice that E[YV] = Prob(x >= V).
Then elementary argument (see figure below) shows that for all s > 0.

esx >= esVYV

Thus,

E[esx] >= esV E[YV] = esV Prob(x >= V)
=> Prob(x >= V) <= e-sV E[esx], for all s >= 0

which is known as the Chernoff Bound. Furthermore,

Prob(x >= V) <= mins>=0 e-sV E[esx]
Prob(x >= V) <= mins>=0 exp{-sV + log E[esx]}

and we can define the rate function

R(s) = sups>=0 {sV - log E[esx]}

finally yielding

Prob(x >= V) <= exp[-R(s)]

i.e., the probability that a random variable is outside some set decreases exponentially fast as a function of the size of the set (V in this case).



To Next Slide / To Table of Content