Misplaced Pages

Small-bias sample space

Article snapshot taken from[REDACTED] with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
This article includes a list of general references, but it lacks sufficient corresponding inline citations. Please help to improve this article by introducing more precise citations. (June 2020) (Learn how and when to remove this message)

In theoretical computer science, a small-bias sample space (also known as ϵ {\displaystyle \epsilon } -biased sample space, ϵ {\displaystyle \epsilon } -biased generator, or small-bias probability space) is a probability distribution that fools parity functions. In other words, no parity function can distinguish between a small-bias sample space and the uniform distribution with high probability, and hence, small-bias sample spaces naturally give rise to pseudorandom generators for parity functions.

The main useful property of small-bias sample spaces is that they need far fewer truly random bits than the uniform distribution to fool parities. Efficient constructions of small-bias sample spaces have found many applications in computer science, some of which are derandomization, error-correcting codes, and probabilistically checkable proofs. The connection with error-correcting codes is in fact very strong since ϵ {\displaystyle \epsilon } -biased sample spaces are equivalent to ϵ {\displaystyle \epsilon } -balanced error-correcting codes.

Definition

Bias

Let X {\displaystyle X} be a probability distribution over { 0 , 1 } n {\displaystyle \{0,1\}^{n}} . The bias of X {\displaystyle X} with respect to a set of indices I { 1 , , n } {\displaystyle I\subseteq \{1,\dots ,n\}} is defined as

bias I ( X ) = | Pr x X ( i I x i = 0 ) Pr x X ( i I x i = 1 ) | = | 2 Pr x X ( i I x i = 0 ) 1 | , {\displaystyle {\text{bias}}_{I}(X)=\left|\Pr _{x\sim X}\left(\sum _{i\in I}x_{i}=0\right)-\Pr _{x\sim X}\left(\sum _{i\in I}x_{i}=1\right)\right|=\left|2\cdot \Pr _{x\sim X}\left(\sum _{i\in I}x_{i}=0\right)-1\right|\,,}

where the sum is taken over F 2 {\displaystyle \mathbb {F} _{2}} , the finite field with two elements. In other words, the sum i I x i {\displaystyle \sum _{i\in I}x_{i}} equals 0 {\displaystyle 0} if the number of ones in the sample x { 0 , 1 } n {\displaystyle x\in \{0,1\}^{n}} at the positions defined by I {\displaystyle I} is even, and otherwise, the sum equals 1 {\displaystyle 1} . For I = {\displaystyle I=\emptyset } , the empty sum is defined to be zero, and hence bias ( X ) = 1 {\displaystyle {\text{bias}}_{\emptyset }(X)=1} .

ϵ-biased sample space

A probability distribution X {\displaystyle X} over { 0 , 1 } n {\displaystyle \{0,1\}^{n}} is called an ϵ {\displaystyle \epsilon } -biased sample space if bias I ( X ) ϵ {\displaystyle {\text{bias}}_{I}(X)\leq \epsilon } holds for all non-empty subsets I { 1 , 2 , , n } {\displaystyle I\subseteq \{1,2,\ldots ,n\}} .

ϵ-biased set

An ϵ {\displaystyle \epsilon } -biased sample space X {\displaystyle X} that is generated by picking a uniform element from a multiset X { 0 , 1 } n {\displaystyle X\subseteq \{0,1\}^{n}} is called ϵ {\displaystyle \epsilon } -biased set. The size s {\displaystyle s} of an ϵ {\displaystyle \epsilon } -biased set X {\displaystyle X} is the size of the multiset that generates the sample space.

ϵ-biased generator

An ϵ {\displaystyle \epsilon } -biased generator G : { 0 , 1 } { 0 , 1 } n {\displaystyle G:\{0,1\}^{\ell }\to \{0,1\}^{n}} is a function that maps strings of length {\displaystyle \ell } to strings of length n {\displaystyle n} such that the multiset X G = { G ( y ) | y { 0 , 1 } } {\displaystyle X_{G}=\{G(y)\;\vert \;y\in \{0,1\}^{\ell }\}} is an ϵ {\displaystyle \epsilon } -biased set. The seed length of the generator is the number {\displaystyle \ell } and is related to the size of the ϵ {\displaystyle \epsilon } -biased set X G {\displaystyle X_{G}} via the equation s = 2 {\displaystyle s=2^{\ell }} .

Connection with epsilon-balanced error-correcting codes

There is a close connection between ϵ {\displaystyle \epsilon } -biased sets and ϵ {\displaystyle \epsilon } -balanced linear error-correcting codes. A linear code C : { 0 , 1 } n { 0 , 1 } s {\displaystyle C:\{0,1\}^{n}\to \{0,1\}^{s}} of message length n {\displaystyle n} and block length s {\displaystyle s} is ϵ {\displaystyle \epsilon } -balanced if the Hamming weight of every nonzero codeword C ( x ) {\displaystyle C(x)} is between ( 1 2 ϵ ) s {\displaystyle ({\frac {1}{2}}-\epsilon )s} and ( 1 2 + ϵ ) s {\displaystyle ({\frac {1}{2}}+\epsilon )s} . Since C {\displaystyle C} is a linear code, its generator matrix is an ( n × s ) {\displaystyle (n\times s)} -matrix A {\displaystyle A} over F 2 {\displaystyle \mathbb {F} _{2}} with C ( x ) = x A {\displaystyle C(x)=x\cdot A} .

Then it holds that a multiset X { 0 , 1 } n {\displaystyle X\subset \{0,1\}^{n}} is ϵ {\displaystyle \epsilon } -biased if and only if the linear code C X {\displaystyle C_{X}} , whose columns are exactly elements of X {\displaystyle X} , is ϵ {\displaystyle \epsilon } -balanced.

Constructions of small epsilon-biased sets

Usually the goal is to find ϵ {\displaystyle \epsilon } -biased sets that have a small size s {\displaystyle s} relative to the parameters n {\displaystyle n} and ϵ {\displaystyle \epsilon } . This is because a smaller size s {\displaystyle s} means that the amount of randomness needed to pick a random element from the set is smaller, and so the set can be used to fool parities using few random bits.

Theoretical bounds

The probabilistic method gives a non-explicit construction that achieves size s = O ( n / ϵ 2 ) {\displaystyle s=O(n/\epsilon ^{2})} . The construction is non-explicit in the sense that finding the ϵ {\displaystyle \epsilon } -biased set requires a lot of true randomness, which does not help towards the goal of reducing the overall randomness. However, this non-explicit construction is useful because it shows that these efficient codes exist. On the other hand, the best known lower bound for the size of ϵ {\displaystyle \epsilon } -biased sets is s = Ω ( n / ( ϵ 2 log ( 1 / ϵ ) ) {\displaystyle s=\Omega (n/(\epsilon ^{2}\log(1/\epsilon ))} , that is, in order for a set to be ϵ {\displaystyle \epsilon } -biased, it must be at least that big.

Explicit constructions

There are many explicit, i.e., deterministic constructions of ϵ {\displaystyle \epsilon } -biased sets with various parameter settings:

  • Naor & Naor (1990) achieve s = n poly ( ϵ ) {\displaystyle \displaystyle s={\frac {n}{{\text{poly}}(\epsilon )}}} . The construction makes use of Justesen codes (which is a concatenation of Reed–Solomon codes with the Wozencraft ensemble) as well as expander walk sampling.
  • Alon et al. (1992) achieve s = O ( n ϵ log ( n / ϵ ) ) 2 {\displaystyle \displaystyle s=O\left({\frac {n}{\epsilon \log(n/\epsilon )}}\right)^{2}} . One of their constructions is the concatenation of Reed–Solomon codes with the Hadamard code; this concatenation turns out to be an ϵ {\displaystyle \epsilon } -balanced code, which gives rise to an ϵ {\displaystyle \epsilon } -biased sample space via the connection mentioned above.
  • Concatenating Algebraic geometric codes with the Hadamard code gives an ϵ {\displaystyle \epsilon } -balanced code with s = O ( n ϵ 3 log ( 1 / ϵ ) ) {\displaystyle \displaystyle s=O\left({\frac {n}{\epsilon ^{3}\log(1/\epsilon )}}\right)} .
  • Ben-Aroya & Ta-Shma (2009) achieves s = O ( n ϵ 2 log ( 1 / ϵ ) ) 5 / 4 {\displaystyle \displaystyle s=O\left({\frac {n}{\epsilon ^{2}\log(1/\epsilon )}}\right)^{5/4}} .
  • Ta-Shma (2017) achieves s = O ( n ϵ 2 + o ( 1 ) ) {\displaystyle \displaystyle s=O\left({\frac {n}{\epsilon ^{2+o(1)}}}\right)} which is almost optimal because of the lower bound.

These bounds are mutually incomparable. In particular, none of these constructions yields the smallest ϵ {\displaystyle \epsilon } -biased sets for all settings of ϵ {\displaystyle \epsilon } and n {\displaystyle n} .

Application: almost k-wise independence

An important application of small-bias sets lies in the construction of almost k-wise independent sample spaces.

k-wise independent spaces

A random variable Y {\displaystyle Y} over { 0 , 1 } n {\displaystyle \{0,1\}^{n}} is a k-wise independent space if, for all index sets I { 1 , , n } {\displaystyle I\subseteq \{1,\dots ,n\}} of size k {\displaystyle k} , the marginal distribution Y | I {\displaystyle Y|_{I}} is exactly equal to the uniform distribution over { 0 , 1 } k {\displaystyle \{0,1\}^{k}} . That is, for all such I {\displaystyle I} and all strings z { 0 , 1 } k {\displaystyle z\in \{0,1\}^{k}} , the distribution Y {\displaystyle Y} satisfies Pr Y ( Y | I = z ) = 2 k {\displaystyle \Pr _{Y}(Y|_{I}=z)=2^{-k}} .

Constructions and bounds

k-wise independent spaces are fairly well understood.

  • A simple construction by Joffe (1974) achieves size n k {\displaystyle n^{k}} .
  • Alon, Babai & Itai (1986) construct a k-wise independent space whose size is n k / 2 {\displaystyle n^{k/2}} .
  • Chor et al. (1985) prove that no k-wise independent space can be significantly smaller than n k / 2 {\displaystyle n^{k/2}} .

Joffe's construction

Joffe (1974) constructs a k {\displaystyle k} -wise independent space Y {\displaystyle Y} over the finite field with some prime number n > k {\displaystyle n>k} of elements, i.e., Y {\displaystyle Y} is a distribution over F n n {\displaystyle \mathbb {F} _{n}^{n}} . The initial k {\displaystyle k} marginals of the distribution are drawn independently and uniformly at random:

( Y 0 , , Y k 1 ) F n k {\displaystyle (Y_{0},\dots ,Y_{k-1})\sim \mathbb {F} _{n}^{k}} .

For each i {\displaystyle i} with k i < n {\displaystyle k\leq i<n} , the marginal distribution of Y i {\displaystyle Y_{i}} is then defined as

Y i = Y 0 + Y 1 i + Y 2 i 2 + + Y k 1 i k 1 , {\displaystyle Y_{i}=Y_{0}+Y_{1}\cdot i+Y_{2}\cdot i^{2}+\dots +Y_{k-1}\cdot i^{k-1}\,,}

where the calculation is done in F n {\displaystyle \mathbb {F} _{n}} . Joffe (1974) proves that the distribution Y {\displaystyle Y} constructed in this way is k {\displaystyle k} -wise independent as a distribution over F n n {\displaystyle \mathbb {F} _{n}^{n}} . The distribution Y {\displaystyle Y} is uniform on its support, and hence, the support of Y {\displaystyle Y} forms a k {\displaystyle k} -wise independent set. It contains all n k {\displaystyle n^{k}} strings in F n k {\displaystyle \mathbb {F} _{n}^{k}} that have been extended to strings of length n {\displaystyle n} using the deterministic rule above.

Almost k-wise independent spaces

A random variable Y {\displaystyle Y} over { 0 , 1 } n {\displaystyle \{0,1\}^{n}} is a δ {\displaystyle \delta } -almost k-wise independent space if, for all index sets I { 1 , , n } {\displaystyle I\subseteq \{1,\dots ,n\}} of size k {\displaystyle k} , the restricted distribution Y | I {\displaystyle Y|_{I}} and the uniform distribution U k {\displaystyle U_{k}} on { 0 , 1 } k {\displaystyle \{0,1\}^{k}} are δ {\displaystyle \delta } -close in 1-norm, i.e., Y | I U k 1 δ {\displaystyle {\Big \|}Y|_{I}-U_{k}{\Big \|}_{1}\leq \delta } .

Constructions

Naor & Naor (1990) give a general framework for combining small k-wise independent spaces with small ϵ {\displaystyle \epsilon } -biased spaces to obtain δ {\displaystyle \delta } -almost k-wise independent spaces of even smaller size. In particular, let G 1 : { 0 , 1 } h { 0 , 1 } n {\displaystyle G_{1}:\{0,1\}^{h}\to \{0,1\}^{n}} be a linear mapping that generates a k-wise independent space and let G 2 : { 0 , 1 } { 0 , 1 } h {\displaystyle G_{2}:\{0,1\}^{\ell }\to \{0,1\}^{h}} be a generator of an ϵ {\displaystyle \epsilon } -biased set over { 0 , 1 } h {\displaystyle \{0,1\}^{h}} . That is, when given a uniformly random input, the output of G 1 {\displaystyle G_{1}} is a k-wise independent space, and the output of G 2 {\displaystyle G_{2}} is ϵ {\displaystyle \epsilon } -biased. Then G : { 0 , 1 } { 0 , 1 } n {\displaystyle G:\{0,1\}^{\ell }\to \{0,1\}^{n}} with G ( x ) = G 1 ( G 2 ( x ) ) {\displaystyle G(x)=G_{1}(G_{2}(x))} is a generator of an δ {\displaystyle \delta } -almost k {\displaystyle k} -wise independent space, where δ = 2 k / 2 ϵ {\displaystyle \delta =2^{k/2}\epsilon } .

As mentioned above, Alon, Babai & Itai (1986) construct a generator G 1 {\displaystyle G_{1}} with h = k 2 log n {\displaystyle h={\tfrac {k}{2}}\log n} , and Naor & Naor (1990) construct a generator G 2 {\displaystyle G_{2}} with = log s = log h + O ( log ( ϵ 1 ) ) {\displaystyle \ell =\log s=\log h+O(\log(\epsilon ^{-1}))} . Hence, the concatenation G {\displaystyle G} of G 1 {\displaystyle G_{1}} and G 2 {\displaystyle G_{2}} has seed length = log k + log log n + O ( log ( ϵ 1 ) ) {\displaystyle \ell =\log k+\log \log n+O(\log(\epsilon ^{-1}))} . In order for G {\displaystyle G} to yield a δ {\displaystyle \delta } -almost k-wise independent space, we need to set ϵ = δ 2 k / 2 {\displaystyle \epsilon =\delta 2^{-k/2}} , which leads to a seed length of = log log n + O ( k + log ( δ 1 ) ) {\displaystyle \ell =\log \log n+O(k+\log(\delta ^{-1}))} and a sample space of total size 2 log n poly ( 2 k δ 1 ) {\displaystyle 2^{\ell }\leq \log n\cdot {\text{poly}}(2^{k}\cdot \delta ^{-1})} .

Notes

  1. cf., e.g., Goldreich (2001)
  2. ^ cf., e.g., p. 2 of Ben-Aroya & Ta-Shma (2009)
  3. Section 4 in Naor & Naor (1990)

References

Categories:
Small-bias sample space Add topic