## Electronic Journal of Probability

- Electron. J. Probab.
- Volume 15 (2010), paper no. 51, 1574-1593.

### On the One Dimensional "Learning from Neighbours" Model

Antar Bandyopadhyay, Rahul Roy, and Anish Sarkar

#### Abstract

We consider a model of a discrete time "interacting particle system" on the integer line where infinitely many changes are allowed at each instance of time. We describe the model using chameleons of two different colours, *viz.*, red (R) and blue (B). At each instance of time each chameleon performs an independent but identical coin toss experiment with probability α to decide whether to change its colour or not. If the coin lands head then the creature retains its colour (this is to be interpreted as a "success"), otherwise it observes the colours and coin tosses of its two nearest neighbours and changes its colour only if, among its neighbours and including itself, the proportion of successes of the other colour is larger than the proportion of successes of its own colour. This produces a Markov chain with infinite state space. This model was studied by Chatterjee and Xu (2004) in the context of diffusion of technologies in a set-up of myopic, memoryless agents. In their work they assume different success probabilities of coin tosses according to the colour of the chameleon. In this work we consider the symmetric case where the success probability, $\alpha$, is the same irrespective of the colour of the chameleon. We show that starting from any initial translation invariant distribution of colours the Markov chain converges to a limit of a single colour, i.e., even at the symmetric case there is no "coexistence" of the two colours at the limit. As a corollary we also characterize the set of all translation invariant stationary laws of this Markov chain. Moreover we show that starting with an i.i.d. colour distribution with density $p\in[0,1]$ of one colour (say red), the limiting distribution is all red with probability $\Pi(\alpha,p)$ which is continuous in $p$ and for $p$ "small" $\Pi(p)>p$. The last result can be interpreted as the model favours the "underdog".

#### Article information

**Source**

Electron. J. Probab. Volume 15 (2010), paper no. 51, 1574-1593.

**Dates**

Accepted: 15 October 2010

First available in Project Euclid: 1 June 2016

**Permanent link to this document**

https://projecteuclid.org/euclid.ejp/1464819836

**Digital Object Identifier**

doi:10.1214/EJP.v15-809

**Mathematical Reviews number (MathSciNet)**

MR2735375

**Zentralblatt MATH identifier**

1225.60123

**Subjects**

Primary: 60J10: Markov chains (discrete-time Markov processes on discrete state spaces) 60K35: Interacting random processes; statistical mechanics type models; percolation theory [See also 82B43, 82C43]

Secondary: 60C05: Combinatorial probability 62E10: Characterization and structure theory 90B15: Network models, stochastic 91D30: Social networks

**Keywords**

Coexistence Learning from neighbours Markov chain Random walk Stationary distribution

**Rights**

This work is licensed under a Creative Commons Attribution 3.0 License.

#### Citation

Bandyopadhyay, Antar; Roy, Rahul; Sarkar, Anish. On the One Dimensional "Learning from Neighbours" Model. Electron. J. Probab. 15 (2010), paper no. 51, 1574--1593. doi:10.1214/EJP.v15-809. https://projecteuclid.org/euclid.ejp/1464819836.