Cutting out a circle using circles

1.5k Views Asked by At

Let $X_0$ be the unit disc, and consider the process of "cutting out circles", where to construct $X_n$ you select a uniform random point $x \in X_{n-1}$, and cut out the largest circle with center $x$. To illustrate this process, we have the following graphic:

cutting out circles

where the graphs are respectively showing one sample of $X_1,X_2,X_3,X_{100}$ (the orange parts have been cut out).

Can we prove we eventually cut everything out? Formally, is the following true $$\text{lim}_{n \to \infty} \mathbb{E}[\text{Area}(X_n)] = 0$$

where $\mathbb{E}$ denotes we are taking the expectation value. Doing simulations, this seems true, in fact $\mathbb{E}[\text{Area}($X_n$)]$ seems to decay with some power law, but after 4 years I still don't really know how to prove this :(. The main thing you need to rule out is that $X_n$ doesn't get too skinny too quickly, it seems.

4

There are 4 best solutions below

10
On

New to this, so not sure about the rigor, but here goes.

Let $A_k$ be the $k$th circle. Assume the area of $\bigcup_{k=1}^n A_k$ does not approach the total area of the circle $A_T$ as $n$ tends towards infinity. Then there must be some area $K$ which is not covered yet cannot harbor a new circle. Let $C = \bigcup_{k=1}^\infty A_k$. Consider a point $P$ in such that $d(P,K)=0$ and $d(P,C)>0$. If no such point exists, then $K \subset C$, as $C$ is a clearly a closed set of points. If such a point does exist, then another circle with center $P$ and nonzero area can be made to cover part of $K$, and the same logic applies to all possible $K$. Therefore there is no area $K$ which cannot contain a new circle, and by consequence $$\lim_{n\to\infty}\Bigg[\bigcup_{k=1}^n A_k\Bigg] = \big[A_T\big]$$ Since the size of circles is continuous, there must be a set of circles $\{A_k\}_{k=1}^\infty$ such that $\big[A_k\big]=E(\big[A_k\big])$ for each $k \in \mathbb{N}$, and therefore $$\lim_{n\to\infty} E(\big[A_k\big]) = \big[A_k\big] $$

EDIT: This proof is wrong becuase I'm bad at probability, working on a new one.

0
On

This is not a complete answer but is too lengthy for a comment. This problem is similar to the problems of statistical geometry. There, it is proposed that you can fill an arbitrary region on the plane with another shape (or shapes) that decrease in size according to a power law. The foundations of statistical geometry are laid out in a book by its developer, John Shier: Fractalize That! : A Visual Essay On Statistical Geometry, World Scientific, 2018 (available at Amazon).

I, myself, have made many such calculations. Several examples can found at Pinterest.

Now, the catch is that this cannot be proven (so far) other than for the simple case of circles within a circular disk. This was published by Christopher Evens, "(Always) Room for One More," Math Horizons, February 2016. This might be just what you need.

10
On

This proof is incomplete, as noted in the comments and at the end of this answer

Apologies for the length. I tried to break it up in to sections so it's easier to follow and I tried to make all implications really clear. Happy to revise as needed

I'll start with some definitions to keep things clear.

Let

  • The area of a set $S \subset \mathbb{R^2}$ be the 2-Lebesgue measure $\lambda^*_2(S):= A(S)$
  • $p_n$ be the point selected from $X_{n-1}$ such that $P(p_n \in Q) = \frac{A(Q)}{A({X_{n-1}})} \; \forall Q\in \mathcal{B}(X_{n-1})$
  • $C_n(p)$ is the maximal circle drawn around $p \in X_{n-1}$ that fits in $X_{n-1}$: $C_n(p) = \max_r \textrm{Circle}(p,r):\textrm{Circle}(p,r) \subseteq X_{n-1})$
  • $A_n = A(C_n(p_n)) $ be the area of the circle drawn around $p_n$ $($i.e., $X_n = X_{n-1}\setminus C_n(p_n))$

We know that $0 \leq A_n \leq 1$. By your definition of the generating process we can also make a stronger statement:

Also, since you're using a uniform probability measure over (well-behaved) subsets of $X_{n-1}$ as the distribution of $p_n$ we have $P(p_n \in B) := \frac{A(B)}{A(X_{n-1})}\;\;\forall B\in \sigma\left(X_{n-1}\right) \implies P(p_1 \in S) = P(S) \;\;\forall S \in \sigma(X_0)$.

Lemma 1: $P\left(\exists L \in [0,\infty): \lim \limits_{n \to \infty} A(X_{n}) = L\right)=1$

Proof: We'll show this by proving

  1. $P(A_n>0)=1\;\forall n$
  2. $(1) \implies P\left(A(X_{i})\leq A(X_{i-1}) \;\;\forall i \right)=1$
  3. $(2) \implies P\left(\exists L \in [0,\infty): \lim \limits_{n \to \infty} A(X_{n}) = L\right)=1$

$A_n = 0$ can only happen if $p_n$ falls directly on the boundary of $X_n$ (i.e., $p_n \in \partial_{X_{n-1}} \subset \mathbb{R^2})$. However, since each $\partial_{X_{n-1}}$ is the union of a finite number of smooth curves (circular arcs) in $\mathbb{R^2}$ we have ${A}(\partial_{X_{n-1}})=0 \;\forall n \implies P(p_n \in \partial_{X_{n-1}})=0\;\;\forall n \implies P(A_n>0)=1\;\forall n$

If $P(A_n>0)=1\;\forall n$ then since $A(X_i) = A(X_{i-1}) - A_n\;\forall i$ we have that $A(X_{i-1}) - A(X_i) = A_n\;\forall i$

Therefore, $P(A(X_{i-1}) - A(X_i) > 0\;\forall i) = P(A_n>0\;\forall i)=1\implies P\left(A(X_{i})\leq A(X_{i-1}) \;\;\forall i \right)=1$

If $P\left(A(X_{i})\leq A(X_{i-1}) \;\;\forall i \right)=1$ then $(A(X_{i}))_{i\in \mathbb{N}}$ is a monotonic decreasing sequence almost surely.

Since $A(X_i)\geq 0\;\;\forall i\;\;(A(X_{i}))_{i\in \mathbb{N}}$ is bounded from below, the monotone convergence theorem implies $P\left(\exists L \in [0,\infty): \lim \limits_{n \to \infty} A(X_{n}) = L\right)=1\;\;\square$

As you've stated, what we want to show is that eventually we've cut away all the area. There are two senses in which this can be true:

  1. Almost all sequences $\left(A(X_i)\right)_1^{\infty}$ converge to $0$: $P\left(\lim \limits_{n\to\infty}A(X_n) = 0\right) = 1$
  2. $\left(A(X_i)\right)_1^{\infty}$ converges in mean to $0$: $\lim \limits_{n\to \infty} \mathbb{E}[A(X_n)] = 0$

In general, these two senses of convergence do not imply each other. However, with a couple additional conditions we can show almost sure convergence implies convergence in mean. Your question is about (2), and we will get there via proving (1) plus a sufficient condition for $(1)\implies (2)$.

I'll proceed as follows:

  1. Show $A(X_n) \overset{a.s.}{\to} 0$ using Borel-Cantelli Lemma
  2. Use the fact that $0<A(X_n)\leq 1$ to apply the Dominated Convergence Theorem to show $\mathbb{E}[A(X_n)] \to 0$

Step 1: $A(X_n) \overset{a.s.}{\to} 0$

If $\lim_{n\to \infty} A(X_n) = A_R > 0$ then there is some set $R$ with positive area $A(R)=A_R >0$ that is a subset of all $X_n$ (i.e.,$\exists R \subset X_0: A(R)>0\;\textrm{and}\;R \subset X_i\;\;\forall i> 0)$

Let's call a set $S\subset X_0:A(S)>0,\;S \subset X_i\;\;\forall i> 0$ a reserved set $(R)$ since we are "setting it aside". In the rest of this proof, the letter $R$ will refer to a reserved set.

Let's define the set $Y_n = X_n \setminus R$, and the event $T_n:=p_n \in Y_{n-1}$ then

Lemma 2: $P\left(\bigcap_1^n T_i \right) \leq A(Y_0)^n = (1 - A_R)^n\;\;\forall n>0$

Proof: We'll prove this by induction. Note that $P(T_1) = A(Y_0)$ and $P(T_1\cap T_2) = P(T_2|T_1)P(T_1)$. We know that if $T_1$ has happened, then Lemma 1 implies that $A(Y_{1}) < A(Y_0)$. Therefore

$$P(T_2|T_1)<P(T_1)=A(Y_0)\implies P\left(T_1 \bigcap T_2\right)\leq A(Y_0)^2$$

If $P(\bigcap_{i=1}^n T_i) \leq A(Y_0)^n$ then by a similar argument we have

$$P\left(\bigcap_{i=1}^{n+1} T_i\right) = P\left( T_{n+1} \left| \;\bigcap_{i=1}^n T_i\right. \right)P\left(\bigcap_{i=1}^n T_i\right)\leq A(Y_0)A(Y_0)^n = A(Y_0)^{n+1}\;\;\square$$

However, to allow $R$ to persist, we must ensure that not only does $T_n$ occur for all $n>0$ but that each $p_n$ doesn't fall in some neighborhood $\mathcal{N}_n(R)$ around $R$:

$$\mathcal{N}_n(R):= \mathcal{R}_n\setminus R$$ $$\textrm{where}\; \mathcal{R}_n:=\{p \in X_{n-1}: A(C_n(p)\cap R)>0\}\supseteq R$$

Let's define the event $T'_n:=p_n \in X_{n-1}\setminus \mathcal{R}_n$ to capture the above requirement for a particular point $p_n$. We then have the following.

Lemma 3: $A(X_n) \overset{a.s.}{\to} A_R \implies P\left(\bigcap \limits_{i \in \mathbb{N}} T_i'\right)=1$

Proof: Assume $A(X_n) \overset{a.s.}{\to} A_R$. If $P\left(\bigcap \limits_{i \in \mathbb{N}} T_i'\right)<1$ then $P\left(\exists k>0:p_k \in \mathcal{R}_k\right)>0$. By the definition of $ \mathcal{R}_k$, $A(C_k(p_k)\cap R) > 0$ which means that $X_{k}\cap R \subset R \implies A(X_{k}\cap R) < A_R$. By Lemma 1, $(X_i)_{i \in \mathbb{N}}$ is a strictly decreasing sequence of sets so $A(X_{j}\cap R) < A_R \;\;\forall j>i$; therefore, $\exists \epsilon > 0: P\left(A(X_n) \overset{a.s.}{\to} A_R - \epsilon\right)>0$. However, this contradicts our assumption $A(X_n) \overset{a.s.}{\to} A_R$. Therefore, $P\left(\bigcap \limits_{i \in \mathbb{N}} T_i'\right)<1$ is false which implies $P\left(\bigcap \limits_{i \in \mathbb{N}} T_i'\right)=1\;\square$

Corollary 1: $P\left(\bigcap \limits_{i \in \mathbb{N}} T_i'\right)=1$ is a necessary condition for $A(X_n) \overset{a.s.}{\to} A_R$

Proof: This follows immediately from Lemma 3 by the logic of material implication: $X \implies Y \iff \neg Y \implies \neg X$ -- an implication is logically equivalent to its contrapositive.

We can express Corollary 1 as an event $\mathcal{T}$ in a probability space $\left(X_0^{\mathbb{N}},\mathcal{F},\mathbb{P}\right)$ constructed from the sample space of infinite sequences of points $p_n \in X_0$ where:

  • $X_0^{\mathbb{N}}:=\prod_{i\in\mathbb{N}}X_0$ is the set of all sequences of points in the unit disk $X_0 \subset \mathbb{R^2}$

  • $\mathcal{F}$ is the product Borel $\sigma$-algebra generated by the product topology of all open sets in $X_0^{\mathbb{N}}$

  • $\mathbb{P}$ is a probability measure defined on $\mathcal{F}$

With this space defined, we can define our event $\mathcal{T}$ as as the intersection of a non-increasing sequence of cylinder sets in $\mathcal{F}$:

$$\mathcal{T}:=\bigcap_{i=1}^{\infty}\mathcal{T}_i \;\;\;\textrm{where } \mathcal{T}_i:=\bigcap_{j=1}^{i} T'_j = \text{Cyl}_{\mathcal{F}}(T'_1,..,T'_i)$$

Lemma 4: $\mathbb{P}(\mathcal{T}_n) = \mathbb{P}(\bigcap_1^n T'_i)\leq \mathbb{P}\left(\bigcap_1^n T_i\right)\leq (1-A_R)^n$

Proof: $\mathbb{P}(\mathcal{T}_n) = \mathbb{P}(\bigcap_1^n T'_i)$ follows from the definition of $\mathcal{T}_n$. $\mathbb{P}(\bigcap_1^n T'_i)\leq \mathbb{P}\left(\bigcap_1^n T_i\right)$ follows immediately from $R\subseteq \mathcal{R}_n\;\;\forall n\;\square$

Lemma 5: $\mathcal{T} \subseteq \limsup \limits_{n\to \infty} \mathcal{T}_n$

Proof: By definition $\mathcal{T} \subset \mathcal{T}_i \;\forall i>0$. Since $\left(\mathcal{T}_i\right)_{i \in \mathbb{N}}$ is nonincreasing, we have $\limsup \limits_{i\to \infty} \mathcal{T}_i = \limsup \limits_{i\to \infty}\mathcal{T}_i = \lim \limits_{i\to \infty}\mathcal{T}_i = \mathcal{T}\;\;\square$

Lemma 6: $\mathbb{P}\left(\limsup \limits_{i\to \infty} \mathcal{T}_i\right) = 0\;\;\forall A_R \in (0,1]$

Proof: From Lemma 4 $$\sum \limits_{i=1}^{\infty} \mathbb{P}\left(\mathcal{T}_i\right) \leq \sum \limits_{i=1}^{\infty} (1-A_R)^i = \sum \limits_{i=0}^{\infty} \left[(1-A_R) \cdot (1-A_R)^i\right] =$$ $$ \frac{1-A_R}{1-(1-A_R)} = \frac{1-A_R}{A_R}=\frac{1}{A_R}-1 < \infty \;\; \forall A_R \in (0,1]\implies$$ $$ \mathbb{P}\left(\limsup \limits_{i\to \infty} \mathcal{T}_i\right) = 0 \;\; \forall A_R \in (0,1]\textrm{ (Borel-Cantelli) }\;\square$$

Lemma 6 implies that only finitely many $\mathcal{T}_i$ will occur with probability 1. Specifically, for almost every sequence $\omega \in X_0^{\infty}$ there $\exists n_{\omega}<\infty$ such $p_{n_{\omega}} \in \mathcal{R}_{n_{\omega}}$.

We can define this as a stopping time for each sequence $\omega \in X_0^{\infty}$ as follows:

$$\tau(\omega) := \max \limits_{n \in \mathbb{N}} \{n:\omega \in \mathcal{T}_n\}$$

Corollary 2: $\mathbb{P}(\tau < \infty) = 1$

Proof: This follows immediately from Lemma 6 and the definition of $\tau$

Lemma 7: $P(\mathcal{T}) = 0\;\;\forall R:A(R)>0$

Proof: This follows from Lemma 5 and Lemma 6


This is where I'm missing a step For Theorem 1 below to work, Lemma 7 + Corollary 1 are not sufficient.

Just because every subset of positive area $R$ has probability zero of occurring doesn't imply that the probability of the set of all possible subsets of area $R$ has a zero probability. An analogous situation is with continuous random variables -- there are an uncountable number of points, but yet when we draw from it we nonetheless get a point.

What I don't know are the sufficient conditions for the following:

$P(\omega)=0 \;\forall \omega\in \Omega: A(\omega)=R \implies P(\{\omega: A(\omega)=R\})=0$



Theorem 1: $A(X_n) \overset{a.s.}{\to} 0$

Proof: Lemma 7 and Corollary 1 imply $A(X_n)$ does not converge to $A_R$ almost surely, which implies $P(A(X_n) \to A_R) < 1 \;\forall A_R > 0$. Corollary 2 makes the stronger statement that $P(A(X_n) \to A_R)=0\;\forall A_R>0$ (i.e., almost never), since we know that the sequences of centers of each circle $p_n$ viewed as a stochastic process will almost surely hit $R$ (again, since we've defined $R$ such that $A(R)>0)$. $P(A(X_n) \to A_R) = 0 \;\forall A_R>0$ with Lemma 1 implies that $P(A(X_n) \to 0) = 1$. Therefore, $A(X_n) \overset{a.s.}{\to} 0\;\square$


Step 2: $\mathbb{E}[A(X_n)] \to 0$

We will appeal to the Dominated Convergence Theorem to prove this result.

Theorem 2: $\mathbb{E}[A(X_n)] \to 0$

Proof: From Theorem 1 we've shown that $A(X_n) \overset{a.s.}{\to} 0$. Given an almost surely constant random variable $Z\overset{a.s.}{=}c$, we have $c>1 \implies |A(X_n)| < Z\;\forall n$. In addition, $\mathbb{E}[Z]=c<\infty$, hence $Z$ is $\mathbb{P}$-integrable. Therefore, $\mathbb{E}[A(X_n)] \to 0$ by the Dominated Convergence Theorem. $\square$

13
On

Suppose our sample space is the open unit disk $\Bbb{D}$ with the usual uniform (Lebesgue) probability measure. To show that it's practically certain that the non-removed area tends to zero, we begin by constructing a sequence of random variables $R_n$, $n \geq 0$, on $\Bbb{D}$ as follows. (The following argument also works for any open set of finite area.)

Before we start slicing out our random disks, we initialize our random variables $R_0: \Bbb{D} \to [0, 1]$ so that, for any point $x \in \Bbb{D}$, $R_0(x)$ is the distance from $x$ to the edge of $\Bbb{D}$, i.e. $R_0(x)$ is the radius of the largest possible disk contained in $\Bbb{D}$ centered at $x$.

At each stage $n$, $n \geq 1$, we then select a point $x \in \Bbb{D}$ uniformly at random, and remove the disk of radius $R_{n-1}(x)$ centered at $x$. $R_n(y)$ is then defined to be $0$ for points $y$ in the removed disk, and updated to be the infimum distance from $y$ to any previously removed point (or the edge of $\Bbb{D}$) otherwise. If for some $y$, all the newly removed points are more than $R_{n-1}(y)$ away from $y$, then we set $R_n(y) = R_{n-1}(y)$.
Note that because our selections take place uniformly at random on $\Bbb{D}$, we may at some stage $n$ select a point $x$ for which $R_{n-1}(x) = 0$ (i.e. $x$ was already removed). In this case, no new disks are removed, $R_n(x)$ is defined to equal $R_{n-1}(x)$ for all points $x \in \Bbb{D}$, and we pick again.
These steps where "nothing happens" have no effect on the end result that all the area of $\Bbb{D}$ is eventually removed, although they grow increasingly frequent as the removed area takes up a larger and larger percentage of $\Bbb{D}$. The convenience of including these "do-nothing" steps is that we don't have to restrict our sample space at each new stage to only select from the non-removed points where $R_{n-1} > 0$.

We then have, for all $x \in \Bbb{D}$, $$R_0(x) \geq R_1(x) \geq R_2(x) \geq ... \geq 0.$$ The area of the remaining circle centered at $x$ at stage $n$ is $\pi R_n(x)^2$. The removed region at any stage $n$ is connected and closed; the non-removed region at stage $n$ is open. Let $C_0 = \emptyset$ and $C_n$ be the removed region at stage $n$; let $U_0 = \Bbb{D}$ and $U_n = \Bbb{D} \setminus C_n$ be the non-removed region. Since the area of $C_n$ is $\pi \Bbb{P}(C_n)$ and the area of $U_n$ is $\pi \Bbb{P}(U_n)$, it suffices to show that $\Bbb{P}(U_n) \to 0$ with probability $1$. If the points chosen in $\Bbb{D}$ at each stage are $X_1, X_2, X_3, ...$ then we have

\begin{align*} \Bbb{P}(C_n) &= R_1(X_1)^2 + ... + R_n(X_n)^2 \\ \Bbb{P}(U_n) &= 1 - R_1(X_1)^2 + ... + R_n(X_n)^2 \ \end{align*}

We first note that for any point $p \in \Bbb{U_{n-1}}$, $$\Bbb{P}(R_n(p) \leq \frac{1}{2} R_{n-1}(p) | X_n \in B(p, R_{n-1}(p)) = 9/16,$$ because if $X_n$ lands closer to $p$ in $R_{n-1}(p)$ than $\frac{3}{4}R_{n-1}(p)$, the radius $R_{n-1}(X_n) \geq \frac{1}{2}R_{n-1}(p)$ (draw a picture to see this). Since $\Bbb{D}$ and $U_n$ are both separable, we can cover them with a countable union of open balls $B(p_n, R_{n-1}(p_n))$. The infinite monkey theorem then tells us that almost surely, the area of any of the countably many open balls in $U_i$ is at some stage $n \geq i$ reduced by a factor of at least $(1 - (1/4)^2) = 15/16$. Since a countable intersection of almost sure events is almost sure, this implies that this happens for every open ball in $U_i$, and hence the area of $U_i$ is in the long term multiplied by a factor of at most $15/16$ with probability 1. In other words, if we let $U_\infty := \cap_n U_n$, then this means $\Bbb{P}(U_\infty) < 15/16 \Bbb{P}(U_n)$ with probability $1$ for all $n$.

We're close to the end here. Let $\pi L$ be the limiting area of $U_\infty$, so $L$ is the limiting probability of $U_\infty$. This means for any $\epsilon$, there exists $n$ so that the area of $U_n$, which we call $\pi L_n$, is at most $\epsilon$ greater than the area of $U_\infty$: $\pi L_n \leq \pi L + \epsilon$. But we know that with probability $1$, the area of $U_n$ will be reduced by a factor of at least $15/16$ in the limit, which means $\frac{15}{16} \pi L_n \geq \pi L$. Therefore $15/16 \pi L + 15/16 \epsilon > \pi L$ for every epsilon with probability $1$, which (since $L$ is nonnegative) is only possible if $L = 0$. $\blacksquare$