Furstenberg-Katznelson argument: Difference between revisions
New page: '''Theorem DHJ(3)'''. For every <math>\delta > 0</math> there exists n such that every subset <math>A \subset [3]^n</math> of density at least <math>\delta<\math> contains a combinatorial... |
No edit summary |
||
Line 1: | Line 1: | ||
'''Theorem DHJ(3)'''. For every <math>\delta > 0</math> there exists n such that every subset <math>A \subset [3]^n</math> of density at least <math>\delta< | '''Theorem DHJ(3)'''. For every <math>\delta > 0</math> there exists n such that every subset <math>A \subset [3]^n</math> of density at least <math>\delta</math> contains a combinatorial line. | ||
'''Proof'''. Suppose for contradiction that the claim failed; then there exists <math>\delta > 0</math>, a sequence of n going to infinity, and a sequence <math> | '''Proof'''. Suppose for contradiction that the claim failed; then there exists <math>\delta > 0</math>, a sequence of n going to infinity, and a sequence <math>A^{(n)} \subset [3]^n</math> of line-free sets, all with density at least <math>\delta</math>. | ||
== Step 1: Random sampling == | == Step 1: Random sampling == | ||
Let <math>m = m(n)</math> be a sufficiently slowly growing integer-valued function of n (e.g. one can take <math>m := \lfloor \log \log \log n \rfloor </math>). We can then construct a random embedding <math>\ | Let <math>m = m^{(n)}</math> be a sufficiently slowly growing integer-valued function of n (e.g. one can take <math>m := \lfloor \log \log \log n \rfloor </math>). We can then construct a random embedding <math>\phi^{(n)}: [3]^m \to [3]^n</math> as follows: | ||
# Pick a random base point <math>x \in [3]^n</math>. Note from the law of large numbers that the number of 0s, 1s, and 2s in x will be comparable to n/3 with probability 1-o(1) (here we use o(1) to denote anything that goes to zero in the limit <math>n \to \infty</math>, independently of the choice of m). | # Pick a random base point <math>x \in [3]^n</math>. Note from the law of large numbers that the number of 0s, 1s, and 2s in x will be comparable to n/3 with probability 1-o(1) (here we use o(1) to denote anything that goes to zero in the limit <math>n \to \infty</math>, independently of the choice of m). | ||
# Pick m random indices <math>i_1,\ldots,i_m</math> in the 0-set of x. These will be distinct with probability 1-o(1), if m is a sufficiently slowly growing function of n. | # Pick m random indices <math>i_1,\ldots,i_m</math> in the 0-set of x. These will be distinct with probability 1-o(1), if m is a sufficiently slowly growing function of n. | ||
# Define <math>\ | # Define <math>\phi^{(n)}(w_1 \ldots w_m)</math> to be the word in <math>[3]^n</math> formed by starting with x, and replacing the <math>i_j</math> coefficient of x (which is 0) with <math>w_i</math>, for each <math>i=1,\ldots,m</math>. | ||
For instance, if n=5, m=2, x=01002, and <math>i_1 = 1; i_2 = 4</math>, then <math>\phi(ab) = a10b2</math> for all <math>a,b=0,1,2</math>. | For instance, if n=5, m=2, x=01002, and <math>i_1 = 1; i_2 = 4</math>, then <math>\phi(ab) = a10b2</math> for all <math>a,b=0,1,2</math>. | ||
The above map is well-defined with probability 1-o(1), and maps combinatorial lines to combinatorial lines. Thus the random set <math>\ | The above map is well-defined with probability 1-o(1), and maps combinatorial lines to combinatorial lines. Thus the random set <math>B^{(n)}_m := (\phi^{(n)})^{-1}(A^{(n)}) \subset [3]^m</math> will be line-free. | ||
Given any <math>0 \leq j \leq m</math> and <math>w \in [3]^j</math>, let <math>B_{m-j,w} \subset [3]^{m-j}</math> denote the random set | Given any <math>0 \leq j \leq m</math> and <math>w \in [3]^j</math>, let <math>B_{m-j,w} \subset [3]^{m-j}</math> denote the random set | ||
: <math> | : <math>B^{(n)}_{m-j,w} := \{ v \in [3]^{m-j}: vw \in B^{(n)}_m \}</math> | ||
where vw is of course the concatenation of v and w. A convenient observation is that the <math> | where vw is of course the concatenation of v and w. A convenient observation is that the <math>B^{(n)}_{m-j,w}</math> are asymptotically ''stationary'' in the sense that the distribution are asymptotically independent of w: | ||
'''Stationarity lemma'''. Let <math>0 \leq j\leq m</math>, let <math>B \subset [3]^{m-j}</math>, and let <math>w, w' \in [3]^j</math>. Then | '''Stationarity lemma'''. Let <math>0 \leq j\leq m</math>, let <math>B \subset [3]^{m-j}</math>, and let <math>w, w' \in [3]^j</math>. Then | ||
: <math>{\Bbb P}( | : <math>{\Bbb P}( B^{(n)}_{m-j,w} = B ) = {\Bbb P}( B^{(n)}_{m-j,w'} = B ) + o(1).</math> | ||
'''Proof'''. Observe that <math>B_{m-j,w} = \ | '''Proof'''. Observe that <math>B_{m-j,w} = \phi^{(n)}_{m-j,w}^{-1}(A_n)</math>, where <math>\phi^{(n)}_{m-j,w}: [3]^{m-j} \to [3]^n</math> is defined like <math>\phi^{(n)}</math>, but replacing m with m-j, and also replacing x with x', defined by replacing x at j random locations <math>i_{m-j+1},\ldots,i_m</math> in the 0-set by the entries in w. But a direct computation using Stirling's formula (if m is sufficiently slowly growing) shows that the probability distribution of x' is within o(1) in total variation norm from the distribution of x (intuitively the j changes to x are a "drop in the ocean" compared with the noise level <math>\sqrt{n}</math> of x). In particular, replacing w with w' only affects the distribution of x' by o(1) in total variation, and the claim follows. | ||
<math>\Box</math> | <math>\Box</math> | ||
== Step 2: Taking limits == | == Step 2: Taking limits == | ||
For fixed n, and m = m(n), we can construct random sets <math> | For fixed n, and <math>m = m^{(n)}</math>, we can construct random sets <math>B^{(n)}_i \subset [3]^i</math> for <math>i=0,\ldots,m</math> by the formula <math>B^{(n)}_i := B^{(n)}_{i,0^{m-i}}</math>; these are all line-free. We then have the asymptotic stationarity property | ||
: <math>{\Bbb P}( | : <math>{\Bbb P}( B^{(n)}_{i,w} = B ) = {\Bbb P}( B^{(n)}_i = B ) + o(1)</math> | ||
for all <math>w \in [3]^j</math> with <math>i+j \leq m</math>, where | for all <math>w \in [3]^j</math> with <math>i+j \leq m</math>, where | ||
: <math> | : <math>B^{(n)}_{i,w} := \{ v \in [3]^i: vw \in B_{i+j}^{(n)} \}.</math> (1) | ||
Also, observe that <math>{\Bbb P}( | Also, observe that <math>{\Bbb P}(B^{(n)}_0 \neq \emptyset)</math> is the density of <math>A_n</math> and is thus at least <math>\delta</math>. | ||
Using the Helly selection principle we can take a subsequence limit as <math>n \to \infty</math> in the vague topology. This gives us a new sequence of random sets <math>B_i \subset [3]^i</math> for all <math>i \geq 0</math> obeying the exact stationarity property | Using the Helly selection principle we can take a subsequence limit as <math>n \to \infty</math> in the vague topology. This gives us a new sequence of random sets <math>B_i \subset [3]^i</math> for all <math>i \geq 0</math> obeying the exact stationarity property | ||
Line 42: | Line 42: | ||
: <math>{\Bbb P}( B_{i,w} = B ) = {\Bbb P}( B_i = B )</math> | : <math>{\Bbb P}( B_{i,w} = B ) = {\Bbb P}( B_i = B )</math> | ||
for all <math>w \in [3]^j</math>, where <math>B_{i,w}</math> is defined | for all <math>w \in [3]^j</math>, where <math>B_{i,w}</math> is defined similarly to (1), and such that <math>B_0 \neq \emptyset</math> with positive probability. Also, all the <math>B_i</math> are almost surely line-free; by removing an event of probability zero we may assume that they are ''surely'' line-free. | ||
== Step 3: Creating transformations == | == Step 3: Creating transformations == | ||
... | ... |
Revision as of 23:08, 13 February 2009
Theorem DHJ(3). For every [math]\displaystyle{ \delta \gt 0 }[/math] there exists n such that every subset [math]\displaystyle{ A \subset [3]^n }[/math] of density at least [math]\displaystyle{ \delta }[/math] contains a combinatorial line.
Proof. Suppose for contradiction that the claim failed; then there exists [math]\displaystyle{ \delta \gt 0 }[/math], a sequence of n going to infinity, and a sequence [math]\displaystyle{ A^{(n)} \subset [3]^n }[/math] of line-free sets, all with density at least [math]\displaystyle{ \delta }[/math].
Step 1: Random sampling
Let [math]\displaystyle{ m = m^{(n)} }[/math] be a sufficiently slowly growing integer-valued function of n (e.g. one can take [math]\displaystyle{ m := \lfloor \log \log \log n \rfloor }[/math]). We can then construct a random embedding [math]\displaystyle{ \phi^{(n)}: [3]^m \to [3]^n }[/math] as follows:
- Pick a random base point [math]\displaystyle{ x \in [3]^n }[/math]. Note from the law of large numbers that the number of 0s, 1s, and 2s in x will be comparable to n/3 with probability 1-o(1) (here we use o(1) to denote anything that goes to zero in the limit [math]\displaystyle{ n \to \infty }[/math], independently of the choice of m).
- Pick m random indices [math]\displaystyle{ i_1,\ldots,i_m }[/math] in the 0-set of x. These will be distinct with probability 1-o(1), if m is a sufficiently slowly growing function of n.
- Define [math]\displaystyle{ \phi^{(n)}(w_1 \ldots w_m) }[/math] to be the word in [math]\displaystyle{ [3]^n }[/math] formed by starting with x, and replacing the [math]\displaystyle{ i_j }[/math] coefficient of x (which is 0) with [math]\displaystyle{ w_i }[/math], for each [math]\displaystyle{ i=1,\ldots,m }[/math].
For instance, if n=5, m=2, x=01002, and [math]\displaystyle{ i_1 = 1; i_2 = 4 }[/math], then [math]\displaystyle{ \phi(ab) = a10b2 }[/math] for all [math]\displaystyle{ a,b=0,1,2 }[/math].
The above map is well-defined with probability 1-o(1), and maps combinatorial lines to combinatorial lines. Thus the random set [math]\displaystyle{ B^{(n)}_m := (\phi^{(n)})^{-1}(A^{(n)}) \subset [3]^m }[/math] will be line-free.
Given any [math]\displaystyle{ 0 \leq j \leq m }[/math] and [math]\displaystyle{ w \in [3]^j }[/math], let [math]\displaystyle{ B_{m-j,w} \subset [3]^{m-j} }[/math] denote the random set
- [math]\displaystyle{ B^{(n)}_{m-j,w} := \{ v \in [3]^{m-j}: vw \in B^{(n)}_m \} }[/math]
where vw is of course the concatenation of v and w. A convenient observation is that the [math]\displaystyle{ B^{(n)}_{m-j,w} }[/math] are asymptotically stationary in the sense that the distribution are asymptotically independent of w:
Stationarity lemma. Let [math]\displaystyle{ 0 \leq j\leq m }[/math], let [math]\displaystyle{ B \subset [3]^{m-j} }[/math], and let [math]\displaystyle{ w, w' \in [3]^j }[/math]. Then
- [math]\displaystyle{ {\Bbb P}( B^{(n)}_{m-j,w} = B ) = {\Bbb P}( B^{(n)}_{m-j,w'} = B ) + o(1). }[/math]
Proof. Observe that [math]\displaystyle{ B_{m-j,w} = \phi^{(n)}_{m-j,w}^{-1}(A_n) }[/math], where [math]\displaystyle{ \phi^{(n)}_{m-j,w}: [3]^{m-j} \to [3]^n }[/math] is defined like [math]\displaystyle{ \phi^{(n)} }[/math], but replacing m with m-j, and also replacing x with x', defined by replacing x at j random locations [math]\displaystyle{ i_{m-j+1},\ldots,i_m }[/math] in the 0-set by the entries in w. But a direct computation using Stirling's formula (if m is sufficiently slowly growing) shows that the probability distribution of x' is within o(1) in total variation norm from the distribution of x (intuitively the j changes to x are a "drop in the ocean" compared with the noise level [math]\displaystyle{ \sqrt{n} }[/math] of x). In particular, replacing w with w' only affects the distribution of x' by o(1) in total variation, and the claim follows. [math]\displaystyle{ \Box }[/math]
Step 2: Taking limits
For fixed n, and [math]\displaystyle{ m = m^{(n)} }[/math], we can construct random sets [math]\displaystyle{ B^{(n)}_i \subset [3]^i }[/math] for [math]\displaystyle{ i=0,\ldots,m }[/math] by the formula [math]\displaystyle{ B^{(n)}_i := B^{(n)}_{i,0^{m-i}} }[/math]; these are all line-free. We then have the asymptotic stationarity property
- [math]\displaystyle{ {\Bbb P}( B^{(n)}_{i,w} = B ) = {\Bbb P}( B^{(n)}_i = B ) + o(1) }[/math]
for all [math]\displaystyle{ w \in [3]^j }[/math] with [math]\displaystyle{ i+j \leq m }[/math], where
- [math]\displaystyle{ B^{(n)}_{i,w} := \{ v \in [3]^i: vw \in B_{i+j}^{(n)} \}. }[/math] (1)
Also, observe that [math]\displaystyle{ {\Bbb P}(B^{(n)}_0 \neq \emptyset) }[/math] is the density of [math]\displaystyle{ A_n }[/math] and is thus at least [math]\displaystyle{ \delta }[/math].
Using the Helly selection principle we can take a subsequence limit as [math]\displaystyle{ n \to \infty }[/math] in the vague topology. This gives us a new sequence of random sets [math]\displaystyle{ B_i \subset [3]^i }[/math] for all [math]\displaystyle{ i \geq 0 }[/math] obeying the exact stationarity property
- [math]\displaystyle{ {\Bbb P}( B_{i,w} = B ) = {\Bbb P}( B_i = B ) }[/math]
for all [math]\displaystyle{ w \in [3]^j }[/math], where [math]\displaystyle{ B_{i,w} }[/math] is defined similarly to (1), and such that [math]\displaystyle{ B_0 \neq \emptyset }[/math] with positive probability. Also, all the [math]\displaystyle{ B_i }[/math] are almost surely line-free; by removing an event of probability zero we may assume that they are surely line-free.
Step 3: Creating transformations
...