• Tidak ada hasil yang ditemukan

The evolution of public belief

Dalam dokumen Essays on social learning and social choice (Halaman 34-40)

Chapter II: The speed of sequential asymptotic learning

2.3 The evolution of public belief

Consider a baseline model, in which each agent observes the private signals of all of her predecessors. In this case the public log-likelihood ratio Λœβ„“π‘‘ would equal the sum

β„“Λœπ‘‘ =

𝑑

βˆ‘οΈ

𝜏=1

𝐿𝜏.

Conditioned on the state this is the sum of i.i.d. random variables, and so by the law of large numbers we have that the limit limπ‘‘β„“Λœπ‘‘/𝑑 wouldβ€”conditioned on (say) πœƒ =+1β€”equal the conditional expectation of𝐿𝑑, which is positive.6

Sub-linear public beliefs

Our first main result shows that when agents observe actions rather than signals, the public log-likelihood ratio grows sub-linearly, and so learning from actions is always slower than learning from signals.

Theorem 4. It holds with probability 1 thatlim𝑑ℓ𝑑/𝑑 =0.

Our second main result shows that, depending on the choice of private signal distributions,ℓ𝑑 can grow at a rate that is arbitrarily close to linear: given any sub- linear functionπ‘Ÿπ‘‘, it is possible to find private signal distributions so thatℓ𝑑 grows as fast asπ‘Ÿπ‘‘.

6In fact,E(𝐿𝑑|πœƒ=+1)is equal to the Kullback-Leibler divergence between𝐹+andπΉβˆ’, which is positive as long as the two distributions are different.

Theorem 5. For anyπ‘Ÿ: N β†’ R>0such that limπ‘‘π‘Ÿπ‘‘/𝑑 = 0 there exists a choice of CDFsπΉβˆ’ and𝐹+such that

lim inf

π‘‘β†’βˆž

|ℓ𝑑| π‘Ÿπ‘‘

> 0 with probability 1.

For example, for some choice of private signal distributions,ℓ𝑑grows asymptotically at least as fast as𝑑/log𝑑, which is sub-linear but (perhaps) close to linear.

Long-term behavior of public beliefs

We next turn to estimating more precisely the long-term behavior of the public log-likelihood ratioℓ𝑑. Since signals are unbounded, agents learn the state, so that ℓ𝑑 tends to+∞ifπœƒ =+1, and toβˆ’βˆžifπœƒ =βˆ’1. In particularℓ𝑑 stops changing sign from some𝑑on, with probability 1; all later agents choose the correct action.

We consider without loss of generality the case that πœƒ = +1, so that ℓ𝑑 is positive from some𝑑on. Thus, recalling (2.1), we have that from some𝑑 on,

ℓ𝑑+1=ℓ𝑑+𝐷+(ℓ𝑑).

This is the recurrence relation that we need to solve in order to understand the long term evolution of ℓ𝑑. To this end, we consider the corresponding differential equation:

d𝑓 d𝑑

(𝑑) =𝐷+(𝑓(𝑑)).

Recall that πΊβˆ’ is the CDF of the private log-likelihood ratio 𝐿𝑑, conditioned on πœƒ =βˆ’1. We show (Lemma 8) that 𝐷+(π‘₯) is well approximated byπΊβˆ’(βˆ’π‘₯)for high π‘₯, in the sense that

π‘₯β†’βˆžlim

𝐷+(π‘₯) πΊβˆ’(βˆ’π‘₯) =1.

In some applications (including the Gaussian one, which we consider below), the expression for πΊβˆ’ is simpler than that for 𝐷+, and so one can instead consider the differential equation

d𝑓 d𝑑

(𝑑) =πΊβˆ’(βˆ’π‘“(𝑑)). (2.2)

This equation can be solved analytically in many cases in whichπΊβˆ’ has a simple form. For example, ifπΊβˆ’(βˆ’π‘₯) = eβˆ’π‘₯ then 𝑓(𝑑) =log(𝑑+𝑐), and ifπΊβˆ’(βˆ’π‘₯) = π‘₯βˆ’π‘˜ then 𝑓(𝑑) = ( (π‘˜+1) ·𝑑+𝑐)1/(π‘˜+1).

We show that solutions to this equation have the same long term behavior as ℓ𝑑, given thatπΊβˆ’ satisfies some regularity conditions.

Theorem 6. Suppose thatπΊβˆ’and𝐺+are continuous, and that the left tail ofπΊβˆ’is convex and differentiable. Suppose also that 𝑓: R>0β†’R>0satisfies

d𝑓 d𝑑

(𝑑)=πΊβˆ’(βˆ’π‘“(𝑑)) (2.3)

for all sufficiently large𝑑. Then conditional on πœƒ=+1,

π‘‘β†’βˆžlim ℓ𝑑 𝑓(𝑑) =1 with probability1.

The condition7 on πΊβˆ’ is satisfied when the random variables 𝐿𝑑 (i.e., the log- likelihood ratios associated with the private signals), conditioned onπœƒ=βˆ’1, have a distribution with a probability density function that is monotone decreasing for all π‘₯less than someπ‘₯0. This is the case for the normal distribution, and for practically every non-atomic distribution one may encounter in the standard probability and statistics literatures.

Gaussian signals

In the Gaussian case,𝐹+is Normal with mean+1 and variance𝜎2, andπΉβˆ’is Normal with meanβˆ’1 and the same variance. A simple calculation shows thatπΊβˆ’ is the Gaussian cumulative distribution function, and so we cannot solve the differential equation (2.2) analytically. However, we can bound πΊβˆ’(π‘₯) from above and from below by functions of the form eβˆ’π‘Β·π‘₯2/π‘₯. For these functions the solution to (2.2) is of the form 𝑓(𝑑)=√︁

log𝑑, and so we can use Theorem 6 to deduce the following.

Theorem 7. When private signals are Gaussian, then conditioned onπœƒ =+1, lim

π‘‘β†’βˆž

ℓ𝑑 (2√

2/𝜎) ·√︁

log𝑑

=1 with probability 1.

Recall, that when private signals are observed, the public log-likelihood ratio ℓ𝑑 is asymptoticallylinear. Thus, learning from actions is far slower than learning from signals in the Gaussian case.

7By β€œthe left tail ofπΊβˆ’is convex and differentiable” we mean that there is someπ‘₯0such that, restricted to(βˆ’βˆž, π‘₯0),πΊβˆ’is convex and differentiable.

The expected time to learn

When private signals are unbounded then with probability 1 the agents eventually all choose the correct actionπ‘Žπ‘‘ =πœƒ. A natural question is: how long does it take for that to happen? Formally, we define thetime to learn

𝑇𝐿 =min{𝑑 : π‘Žπœ =πœƒ for all𝜏 β‰₯ 𝑑},

and study its expectation. Note that in the baseline case of observed signals𝑇𝐿 has finite expectation, since the probability of a mistake at time𝑑 decays exponentially with𝑑.

We first study the expectation of𝑇𝐿 in the case of Gaussian signals. To this end we define thetime of first mistakeby

𝑇1=min{𝑑 : π‘Žπ‘‘ β‰ πœƒ}

ifπ‘Žπ‘‘ β‰  πœƒ for some 𝑑, and by𝑇1 =0 otherwise. We calculate a lower bound for the distribution of𝑇1, showing that it decays at most as fast as 1/𝑑.

Theorem 8. When private signals are Gaussian then for everyπœ€ >0there exists a π‘˜ >0such that for all𝑑

P(𝑇1 =𝑑) β‰₯ π‘˜ 𝑑1+πœ€

Β·

Thus𝑇1has a very thick tail, decaying far slower than the exponential decay of the baseline case. In particular,𝑇1 has infinite expectation, and so, since𝑇𝐿 > 𝑇1, the expectation of the time to learn𝑇𝐿 is also infinite.

In contrast, we show that when private signals have thick tailsβ€”that is, when the probability of a strong signal vanishes slowly enoughβ€”then the time to learn has finite expectation. In particular, we show this when the left tail ofπΊβˆ’ and the right tail of𝐺+ are polynomial.8

Theorem 9. Assume that πΊβˆ’(βˆ’π‘₯) =𝑐·π‘₯βˆ’π‘˜ and that 𝐺+(π‘₯) =1βˆ’π‘Β·π‘₯βˆ’π‘˜ for some 𝑐 > 0andπ‘˜ > 0, and for allπ‘₯greater than someπ‘₯0. ThenE(𝑇𝐿) < ∞.

8Recall thatπΊβˆ’is the conditional cumulative distribution function of the private log-likelihood ratios𝐿𝑑.

An example of private signal distributions𝐹+andπΉβˆ’for whichπΊβˆ’and𝐺+have this form is given by the probability density functions

π‘“βˆ’(π‘₯) =







ο£²







ο£³

𝑐·eβˆ’π‘₯π‘₯βˆ’π‘˜βˆ’1 when 1 ≀ π‘₯

0 when βˆ’1< π‘₯ < 1 𝑐· (βˆ’π‘₯)βˆ’π‘˜βˆ’1 whenπ‘₯ ≀ βˆ’1.

and 𝑓+(π‘₯) = π‘“βˆ’(βˆ’π‘₯), for an appropriate choice of normalizing constant 𝑐 > 0. In this caseπΊβˆ’(βˆ’π‘₯) =1βˆ’πΊ+(π‘₯) = 𝑐

π‘˜π‘₯βˆ’π‘˜ for allπ‘₯ >1.9

The proof of Theorem 9 is rather technically involved, and we provide here a rough sketch of the ideas behind it.

We say that there is an upsetat time 𝑑 if π‘Žπ‘‘βˆ’1 β‰  π‘Žπ‘‘. We denote byΞthe random variable which assigns to each outcome the total number of upsets

Ξ =|{𝑑 : π‘Žπ‘‘βˆ’1 β‰ π‘Žπ‘‘}|.

We say that there is arunof lengthπ‘š from time𝑑 ifπ‘Žπ‘‘ =π‘Žπ‘‘+1 =Β· Β· Β· =π‘Žπ‘‘+π‘šβˆ’1. As we will condition onπœƒ =+1 in our analysis, we say that a run from time𝑑 isgoodif π‘Žπ‘‘ =1 andbadotherwise. A trivial but important observation is that the number of maximal finite runs is equal to the number of upsets, and so, ifΞ = 𝑛, and if𝑇𝐿 =𝑑, then there is at least one run of length𝑑/𝑛before time𝑑. Qualitatively, this implies that if the number of upsets is small, and if the time to learn is large, then there is at least one long run before the time to learn.

We show that it is indeed unlikely that Ξ is large: the distribution of Ξ has an exponential tail. Incidentally, this holds foranyprivate signal distribution:

Proposition 8. For every private signal distribution there exist𝑐 >0and0 < 𝛾 <1 such that for all𝑛 > 0

P(Ξβ‰₯ 𝑛) ≀𝑐𝛾𝑛.

Intuitively, this holds because whenever an agent takes the correct action, there is a non-vanishing probability that all subsequent agents will also do so, and no more upsets will occur.

9Theorem 9 can be proved for other thick-tailed private signal distributions: for example, one could take different values of𝑐andπ‘˜forπΊβˆ’and𝐺+, or one could replace their thick polynomial tails by even thicker logarithmic tails. For the sake of readability we choose to focus on this case.

Thus, it is very unlikely that the number of upsetsΞis large. As we observe above, whenΞis small then the time to learn𝑇𝐿 can only be large if at least one of the runs is long. WhenπΊβˆ’has a thin tail then this is possible; indeed, Theorem 8 shows that the first finite run has infinite expected length when private signals are Gaussian.

However, whenπΊβˆ’ has a thick, polynomial left tail of orderπ‘₯βˆ’π‘˜, we show that it is very unlikely for any run to be long: the probability that there is a run of length𝑛 decays at least as fast as exp(βˆ’π‘›π‘˜/(π‘˜+1)), and in particular runs have finite expected length. Intuitively, when strong signals are rare then runs tend to be long, as agents are likely to emulate their predecessor. Conversely, when strong signals are more likely then agents are more likely to break a run, and so runs tend to be shorter.

Putting together these insights, we conclude that it is unlikely that there are many runs, and, in the polynomial signal case, it is unlikely that runs are long. Thus𝑇𝐿 has finite expectation.

Probability of taking the wrong action

Yet another natural metric of the speed of learning is the probability of mistake 𝑝𝑑 =P(π‘Žπ‘‘ β‰  πœƒ).

Calculating the asymptotic behavior of 𝑝𝑑seems harder to tackle.

For the Gaussian case, while we cannot estimate 𝑝𝑑 precisely, Theorem 8 immedi- ately implies a lower bound: 𝑝𝑑is at leastπ‘˜/𝑑1+πœ€, for everyπœ€ >0 andπ‘˜ that depends onπœ€. This is much larger than the exponentially vanishing probability of mistake in the revealed signal baseline case.

More generally, we can use Theorem 4 to show that 𝑝𝑑 vanishes sub-exponentially for any signal distribution, in the sense that

π‘‘β†’βˆžlim 1 𝑑

log𝑝𝑑 =0.

To see this, note that the probability of mistake at time 𝑑 βˆ’1, conditioned on the observed actions, is exactly equal to

min{πœ‡π‘‘,1βˆ’πœ‡π‘‘}; where we recall that

πœ‡π‘‘ =P(πœƒ =+1|π‘Ž1, . . . , π‘Žπ‘‘βˆ’1) = eℓ𝑑 eℓ𝑑 +1

is the public belief. This is due to the fact that if the outside observer, who holds belief πœ‡π‘‘, had to choose an action, she would choose π‘Žπ‘‘βˆ’1, the action of the last player she observed, a player who has strictly more information than her. Thus

𝑝𝑑 =E(min{πœ‡π‘‘,1βˆ’πœ‡π‘‘})=E 1

e|ℓ𝑑|+1

,

and since, by Theorem 4,|ℓ𝑑|is sub-linear, it follows that 𝑝𝑑is sub-exponential.

Dalam dokumen Essays on social learning and social choice (Halaman 34-40)

Dokumen terkait