51

Suppose we have intervals of alternating color on $\mathbb{R}$ (say, red / blue / red / blue / …). All intervals have independent length, with all red intervals distributed as $\mathbb{P}_{R}$, all blue intervals distributed as $\mathbb{P}_B$.

Once the intervals are generated, I update them iteratively as follows: if an interval is smaller than its two immediate neighbors, then it get swallowed by its neighbors: all three intervals join to become one big interval of the same color as the two end ones.

For example, if we have: … (red 2) (blue 3) (red 4) (blue 1) (red 3) (blue 4) (red 5) …

Then the next step we will see … (red 2) (blue 3) (red 8) (blue 4) (red 5) …

And the next step we will see … (red 2) (blue 3) (red 17)

And so on.

The question is: does one color eventually wins out, or do the two colors switch back and forth infinitely often? Concretely: for any fixed point $x \in \mathbb{R}$, if $B_n(x)$ denotes the color of $x$ at time step $n$ (= 1 if blue, = 0 if red), then will the sequence $\{B_n(x)\}$ converges, or will it visit 1 and 0 infinitely often (with probability 1)?

Clearly we need to know something about the two distributions. Then, the question really is: under what assumptions on $\mathbb{P}_R$ and $\mathbb{P}_B$ can we answer the above question?

One can check that the process is well defined, in the sense that it doesn't matter which interval is updated first. Also, for the process to get stuck, one needs to have an infinite sequence of increasing numbers, and this happens with probability 0. We further assume that the distributions are such that there is no tie between intervals with positive probability. (E.g.: when the distributions are continuous.)

We know the following:

If $\mathbb{P}_R = \mathbb{P}_B$ (that is, if red and blue have the same distributions), then the answer is infinitely often w.p. 1.

Proof: for any point $x$, the events of {converge to 1} and {converge to 0} lie in the tail sigma-field, hence occur w.p. 1 or 0. But by symmetry they have the same probability, hence it must be 0. Since the process is translation invariant, sequences $\{B_n(x)\}$ have the same tail distribution at all points, and hence there is no convergence in color for the whole line. (In particular, this argument implies that if one point converges to a color with probability 1, then the whole line converges to the same color with probability 1). ▢

THOUGHTS:

Intuitively, suppose that at the start, red has a slight "advantage" over blue. (Say, $\mathbb{P}_R$ stochastically dominates $\mathbb{P}_B$.) Then one would expect this advantage to get amplified in later rounds, and eventually red wins with probability 1. But we can't find a concrete proof of this. The hardest step seems to be finding a characterization of "advantage" (that is, a property of a distribution relative to another) that is invariant under updates, and tells us that it improves the probability of one color winning eventually. Stochastic domination is a fairly strong assumption — we think that there should be a milder condition.

For a toy case, a simple example is when all red intervals have length 1, and all blue intervals have length either $1 - \epsilon_1$ or $1 + \epsilon_2$, where $\epsilon_1, \epsilon_2$ are distinct irrational numbers (so as to stop having ties with positive probability).

For simple examples as above, one can write down the distribution of red and blue after step 1 of the update (one step of the update means: updating all local minima in the previous round) — note that the intervals of the same color are still i.i.d., but red and blue are not independent (since we know that the surviving blues and reds have to have "big" size to stop the neighbor from expanding). However, I think they should be exchangeable, in which case it might be easier to do the following:

  • update the process once. Now red and blue have new distributions $\mathbb{P}_R'$ and $\mathbb{P}_B'$;
  • "peel off" all the reds and blue and "shuffle" them (that is: generate another sequence of interlacing red/blue following the new distribution). Then run another update.

If we have exchangeability, then this does not change the tail distribution. But analytically it is easier to work with. And now one can search for the characterization of an "advantage" that get amplified in further updates.

For instance, one might hope that if red has a "huge advantage" over blue, then $P(B_n(x) = 1)$ decays fast enough so $\sum_{n}P(B_n(x) = 1) < \infty$, then the result follows from Borel–Cantelli. Then the next step would be to show that one can reach this "huge advantage" starting from a "small advantage" in finite time step.

Intuitively it seems that the tightest (minimal definition of advantage) would be $P(B_0(x) = 1) < 1/2$. In this case we can compute and show that $P(B_n(x) = 1)$ is a monotone decreasing function in $n$, but quantifying the rate seems very difficult.

If anybody has fresh idea to approach this problem, please let me know!


Here are some updates on this problem: We can also run this problem in continuous time as follows: at each end point of the intervals there is a balloon. At time $t$ the balloon as radius $t/2$. When two balloons meet they "pop", and the two points are removed from the line.

This put the process in the world of stable matching, hierarchical coalescent process, and some other physics literature. The advantage of this continuous time approach is that one can check that the sequence of points after running for some time $T$ is still a renewal process, and furthermore, one can write down a set of ODEs for the Laplace transform of the distribution of the two lengths. In fact, if the distribution $P_R$ and $P_B$ are identical, then one can show that the normalized length $X_t/t$ converges to a "universal" limit. (Universal in the sense that it almost doesn't depend on the initial distribution).

The relevant paper is: Faggionato, Roberto, and Toninelli - Universality for one-dimensional hierarchical coalescence processes with double and triple merges.

This paper considers a very similar situation, so to apply the result in this paper directly one needs to do a little argument, which unfortunately does not carry over when $P_R$ and $P_B$ have different distributions. I'll skip writing down the argument because I think that we might benefit from a fresh approach.

Edit (Stefan Kohl, 2021-05-01): I undeleted this highly upvoted question which was deleted single-handedly by an (at that time) SE employee in 2014, since I didn't see a reason why it should be deleted. If there is a reason why it should be deleted, feel free to delete again.

Stefan Kohl
  • 19,498
  • 21
  • 73
  • 136
  • It seems there are distributions for which it matters which interval you use to start the update process. So I do not know what you mean by "well-defined" For example, suppose all intervals are length 1, except for two adjoining intervals which are length 2. How do you go about updating this case? Gerhard "Ask Me About System Design" Paseman, 2011.02.19 – Gerhard Paseman Feb 20 '11 at 07:50
  • Ah, numbering the colors leads to part of the confusion. You are looking for oscillatory behaviour versus non-oscillatory behaviour. I will re-examine your claim of "well-defined". Gerhard "Ask Me About System Design" Paseman, 2011.02.19 – Gerhard Paseman Feb 20 '11 at 07:53
  • 3
    This is a wonderful question. You write that $[B_n(x)\to1]$ is a tail event (space $x$ fixed, time $n$ going to infinity). For which sequence of sigma-algebras? Since the update process is deterministic once $B_0=(B_0(x))$ is known, I understand you consider the sigma-algebras $G_h$ generated by ${B_0(x);|x|\ge h}$ and the tail sigma-algebra $T$ of the sequence $(G_h)$. But then, why is $T$ trivial? And, more importantly, why should $[B_n(0)\to1]$ belong to $T$? – Did Feb 20 '11 at 09:06
  • @Didier: I also wondered about that. It is easy to see that the event is translation invariant, which also gives us a proof by the ergodic theorem. – Ori Gurel-Gurevich Feb 20 '11 at 16:57
  • @Ori Even that... To get invariance by translation one should specify the initial distribution. If at time $0$, one imposes that a blue interval begins on the right of site $0$, then the events $[B_0(x)=1]$ do not have the same probability for various sites $x$ hence the probability that $[B_n(x)\to1]$ might depend on $x$. Probably one should start from the stationary distribution of the renewal process of the colors. And I would love to see things written carefully here... – Did Feb 20 '11 at 18:06
  • @Didier: I prefer to think about the as just a bi-infinite sequence of numbers, and forget about the spatial interpretation. Then initially we define all the even indexes as red and the odd as blue, and then change the colors according to the rule. Then the whole process is invariant w.r.t. shift by 2 and we can use the ergodic theorem. – Ori Gurel-Gurevich Feb 20 '11 at 18:23
  • @Ori: Thanks for the indication. So, for you, there are lengthes $(x(i))$ and colors $(c_n(i))$, both indexed by $i\in\mathbb{Z}$, with $c_0(2i)=+1$, $c_0(2i+1)=-1$, $(x(i))$ independent, $x(2i)$ of distribution $\mu_+$, $x(2i+1)$ of distribution $\mu_-$. The lengthes do not change, the color change as follows. If $i$, $k\ge1$ and $\ell\ge1$ are such that $c_n(i-k)=\cdots=c_n(i-1)$, $c_n(i+1)=\cdots=c_n(i+\ell)$, $c_n(i-1)=-c_n(i)=c_n(i+1)$, $x(i-k)+\cdots+x(i-1)>x(i)$ and $x(i+1)+\cdots+x(i+\ell)>x(i)$, then $c_{n+1}(i)=-c_n(i)$, otherwise $c_{n+1}(i)=c_n(i)$. Is this what you have in mind? – Did Feb 20 '11 at 19:35
  • @Didier: Basically, yes, but you have to take into account that the middle segment may also consist of many original segments (In what you arote only $c(i)$ changes). That's what you probably had in mind anyway.

    Of course, if the lengths distribution is reasonable (the means are finite) then one could also construct a stationary point process with this interval distribution. Still, I think it's easier to work with $\mathbb{Z}$.

    – Ori Gurel-Gurevich Feb 21 '11 at 02:09
  • @Ori True, I missed that a whole bunch of sites $i$ can flip together. – Did Feb 21 '11 at 06:24
  • What happens if you take a countable state space and declare in the case of a tie, say (red 4) (blue 4) (red 5), red swallows blue? Then, we can consider a 2-state i.i.d. process $X={r,b}^{\mathbb{Z}}$ with measure $(p,1-p)^{\mathbb{Z}}$, and factor maps where, for example, the block $rrbrr$ maps to $rrrrr$ (blue swallowed by red). Then we ask if the entropy goes to zero after applying our factor map (or maps). – Stephen Shea Feb 21 '11 at 14:38
  • Hi all, Thank you for all the lively discussions.

    @Gerhard: for the case 1 - 1 - 2 - 2 - 1, then the process gets stuck because there are ties. I just assume from the start that the distributions are such that there is no ties between two adjacent blocks with probability 1.

    – Ngoc Mai Tran Feb 21 '11 at 20:49
  • @Didier: yes, given a fixed sequence the process is deterministic. So for each $\omega$, $B(x)(\omega) \in {0,1}^\mathbb{Z}$ is a sequence of 0 and 1, where $B_n(x)(\omega)$ is the $n^{th}$ entry of this sequence. Then by writing $B_n(x)$ I mean the corresponding random variable. (The randomness comes from the initialization of the sequence. So the sigma-field in this case is that generated by the random variables $(B_i, R_j: i, j \in \mathbb{Z})$, where $R_j$ are i.i.d with distribution $P_R$, $B_i$ are i.i.d with distribution $P_B$ (corresponding to lengths of the initial intervals). – Ngoc Mai Tran Feb 21 '11 at 20:49
  • @Ori: the translation invariance set up is nice (that's what I had in mind when I wrote "translation invariance" as well).

    @Stephen: You can ask the same question for the toy model I wrote down (red = constant 1, blue = 1+$\epsilon_1$ or $1-\epsilon_2$). I'm not sure I follow the bit about factor map (how would you go about computing it, for example?)

    Ngoc

    – Ngoc Mai Tran Feb 21 '11 at 20:50
  • Are there any interesting observations or questions for the following combinatorial toy model? For each pair of adjacent intervals $A,B$ write $1$ if $B$ is longer than $A$ and $-1$ otherwise. This gives an infinite sequence of with values in $\pm 1$ (we suppose that there are no ties). Suppress all consecutive pairs $(-1)1$ and iterate (this corresponds to merging except that the lengths are not taken into account). – Roland Bacher Mar 08 '12 at 09:03
  • An interesting example is where the initial blue intervals have length $3n+1$ with probability $\frac18(10/2^n-9/4^n)$, and the initial red intervals have length $3n+2$ with probability $\frac18(8/2^n-6/4^n)$. Then the red and blue distributions both have mean $7$ and variance $24$. –  Feb 21 '22 at 16:52

0 Answers0