Pseudorandomness Chapter 6 learning notes.
IID-Bit Sources: A simple version of this question was already
considered by von Neumann. He looked at sources that consist of
boolean random variables X1,X2,…,Xn ∈ {0,1} that are independent
but biased. That is, for every i, Pr [Xi = 1] = δ for some unknown δ.
How can such a source be converted into a source of independent, unbiased bits? Von Neumann proposed the following extractor: Break all
the variables in pairs and for each pair output 0 if the outcome was
01, 1 if the outcome was 10, and skip the pair if the outcome was 00
or 11. This will yield an unbiased random bit after 1/(2δ(1 − δ)) pairs
on average.
Statistical Difference: For random variables X and Y taking values in U, their statistical difference (a.k.a. variation distance) is Δ(X,Y)≜T⊂Umax∣Pr[X∈T]−Pr[Y∈T]∣. We say that X and Y are ε-close if Δ(X,Y)≤ε.
Deterministic Extractors: Let C be a class of sources on {0,1}n. An ε-extractor for C is a function Ext:{0,1}n→{0,1}m such that for every X∈C, Ext(X) is ε-close to Um.
Properties of Statistical Difference: Let X,Y,Z,X1,X2,Y1,Y2 be random variables taking values in a universe U. Then,
(1) Δ(X,Y)≥0, with equality iff X and Y are identically distributed,
(2) Δ(X,Y)≤1, with equality iff X and Y have disjoint supports,
(3) Δ(X,Y)=Δ(Y,X),
(4) Δ(X,Z)≤Δ(X,Y)+Δ(Y,Z),
(5) for every function f, we have Δ(f(X),f(Y))≤Δ(X,Y),
(6) Δ((X1,X2),(Y1,Y2))≤Δ(X1,Y1)+Δ(X2,Y2) if X1 and X2, as well as Y1 and Y2, are independent, and
(7) Δ(X,Y)=21⋅∣X−Y∣1.
Proof: Properties 1∼4 are obviously correct.
(5) For every function f:U→U′, Δ(f(X),f(Y))=T′⊂U′max∣Pr[f(X)∈T′]−Pr[f(Y)∈T′]∣=T′⊂U′max∣Pr[X∈f−1(T′)]−Pr[Y∈f−1(T′)]∣≤T⊂Umax∣Pr[X∈T]−Pr[Y∈T]∣=Δ(X,Y).
(6)
(7) Δ(X,Y)=T⊂Umax∣Pr[X∈T]−Pr[Y∈T]∣=∑t∈S(Pr[X=t]−Pr[Y=t]), where S={t∈U∣Pr[X=t]≥Pr[Y=t]}.
Due to the symmetry, we also have Δ(X,Y)=∑t∈T(Pr[Y=t]−Pr[X=t]), where T={t∈U∣Pr[X=t]<Pr[Y=t]}.
Thus:
Δ(X,Y)=21⋅(t∈S∑(Pr[X=t]−Pr[Y=t])+t∈T∑(Pr[Y=t]−Pr[X=t]))=21⋅∣X−Y∣1
as desired.
□
Proposition: Let A(w;r) be a randomized algorithm such that A(w;Um) has error probability at most γ, and let Ext:{0,1}n→{0,1}m be an ε-extractor for a class C of sources on {0,1}n. Define A′(w;x)=A(w;Ext(x)). Then for every source X∈C, A′(w;X) has error probability at most γ+ϵ.
Proof: The error probability of A′(w;X) is Pr[Ext(X)∈S], where S⊂{0,1}m is the set of coin tosses sequences which makes A(w;Um) error. Note that Pr[Um∈S]≤γ.
Thus, Pr[Ext(X)∈S]≤Pr[Um∈S]+Δ(Ext(X),Um)≤γ+ϵ.
□
Unpredictable-Bit Sources (a.k.a. Santha–Vazirani Sources): UnpredBitsn,δ is the class of unpredictable-bit sources, i.e. for every i, every x1,…,xn∈{0,1} and some constant δ>0, satisfy δ≤Pr[Xi=1∣X1=x1,X2=x2,…,Xi−1=xi−1]≤1−δ.
Problem 6.6 (Extracting from Unpredictable-Bit Sources):
(1) Let X be a source taking values in {0,1}n such that for all x,y, Pr[X=x]/Pr[X=y]≤(1−δ)/δ. Show that X∈UnpredBitsn,δ.
(2) Prove that for every function Ext:{0,1}n→{0,1} and every δ>0, there exists a source X∈UnpredBitsn,δ with parameter δ such that Pr[Ext(X)=1]≤δ or Pr[Ext(X)=0]≥1−δ.
Proof:
□
Entropy Measures: Let X be a random variable. Then
-
the Shannon entropy of X is:
HSh(X)=x←RXE[log2Pr[X=x]1]
-
the Rényi entropy of X is:
H2(X)=log2⎝⎜⎛x←RXE[Pr[X=x]]1⎠⎟⎞=log2CP(X)1
-
the min-entropy of X is:
H∞(X)=xmin{log2Pr[X=x]1}