Next Article in Journal
Validating Game-Theoretic Models of Terrorism: Insights from Machine Learning
Previous Article in Journal
Rock-Paper-Scissors Play: Beyond the Win-Stay/Lose-Change Strategy
Previous Article in Special Issue
Imitation and Local Interactions: Long Run Equilibrium Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Competing Conventions with Costly Information Acquisition

Department of Economics, Ca’ Foscari University of Venice, Dorsoduro 3246, 30123 Venice, Italy
Games 2021, 12(3), 53; https://doi.org/10.3390/g12030053
Submission received: 22 February 2021 / Revised: 22 June 2021 / Accepted: 23 June 2021 / Published: 25 June 2021
(This article belongs to the Special Issue Social Coordination Games)

Abstract

:
We consider an evolutionary model of social coordination in a 2 × 2 game where two groups of players prefer to coordinate on different actions. Players can pay a cost to learn their opponent’s group: if they pay it, they can condition their actions concerning the groups. We assess the stability of outcomes in the long run using stochastic stability analysis. We find that three elements matter for the equilibrium selection: the group size, the strength of preferences, and the information’s cost. If the cost is too high, players never learn the group of their opponents in the long run. If one group is stronger in preferences for its favorite action than the other, or its size is sufficiently large compared to the other group, every player plays that group’s favorite action. If both groups are strong enough in preferences, or if none of the groups’ sizes is large enough, players play their favorite actions and miscoordinate in inter-group interactions. Lower levels of the cost favor coordination. Indeed, when the cost is low, in inside-group interactions, players always coordinate on their favorite action, while in inter-group interactions, they coordinate on the favorite action of the group that is stronger in preferences or large enough.

1. Introduction

Since the seminal contribution of Kandori et al. [1], evolutionary game theorists have used stochastic stability analysis and 2 × 2 coordination games to study the formation of social conventions (Lewis [2] and Bicchieri [3] are classical references on social conventions from philosophy, while for economics, see Schelling [4], Young [5], and Young [6]). Some of these works focus on coordination games such as the battle of sexes: a class that describes situations in which two groups of people prefer to coordinate on different actions. In this framework, the long-run convention may depend on how easily people can learn each other’s preferences.
Think about Bob and Andy, who want to hang out together: they can either go to a football match or to the cinema. Both Andy and Bob prefer football, but they do not know what the other prefers. In certain contexts, learning each other’s preferences may require too much effort. In these cases, if Bob and Andy know that everybody usually goes to the cinema, they go to the cinema without learning each other’s preferences. In other situations, learning each other’s preferences may require a small effort (for instance, watching each other’s Facebook walls). In this case, Bob and Andy learn that they both prefer football, so they go to a football match together.
In this work, we contribute to the literature on coordination games. We show which conventions become established between two groups of people different in preferences if people can learn each other’s preferences by exerting an effort. We do so, formalizing the previous example and studying the evolution of conventions in a dynamic setting. We model the coordination problem as a repeated language game (Neary [7]): we use evolutionary game theory solution concepts and characterize the long-run equilibrium as the stochastically stable state (see Foster and Young [8], Kandori et al. [1] and Young [9]).
We consider a population divided into two groups, which repeatedly play a 2 × 2 coordination game. We assume that one group is larger than the other and that the two groups differ in preferences towards the coordination outcomes. At each period, players can learn the group of their opponent if they pay a cost. Such a cost represents the effort to exert if they want to learn their opponent’s group. If they pay this cost, they can condition the action to the player they meet. If they do not pay it, they can only play the same action with every player. Given this change in the strategic set, we introduce a new possible perturbation. Players can make a mistake in the information choice and a mistake in the coordination choice. We model two situations: one where the cost is equal to zero, and players always learn their opponent’s group, and one where the cost is strictly positive and players can learn their opponent’s group only if they pay that cost. Players decide myopically their best reply based on the current state, which is always observable. We say that a group has a stronger preference for its favorite action than the other if it assigns higher payoffs to its favorite outcome or lower payoffs to the other outcome compared to the other group.
We find that cost level, strength in preferences, and group size are crucial drivers for the long-run stability of outcomes. Two different scenarios can happen, depending on the cost. Firstly, low cost levels favor coordination: players always coordinate on their favorite action with players of their group. If one group has a stronger preference for its favorite action or its size is sufficiently large compared to the other, every player plays the action preferred by that group in inter-group interactions. Interestingly, players from the group that is stronger in preferences never need to buy the information because they play their favorite action with everyone, while players from the other group always need to buy it.
Secondly, when the cost is high, players never learn the group of their opponents, and they play the same action with every player. Some players coordinate on one action that they do not like, even with players of their group. Indeed, we find that when one group is stronger in preferences than the other for its favorite action, or if its size is sufficiently large compared to the other, every player coordinates on that group’s favorite action. Even worse, the two groups may play their favorite action and miscoordinate in inter-group interactions. We find that this outcome occurs when both groups have strong enough preferences for their favorite action or if the two groups are sufficiently close in size.
Neary [7] considers a similar model, where each player decides one single action valid for both groups. Hence, it is as if learning an opponent’s group requires too much effort, and no player ever learns it. Given this scenario, Neary’s results are the same as in our analysis when the cost is high.
It is helpful to highlight our analysis with respect to the one proposed by Neary, from which we started. We firstly enlarge Neary’s analysis to the case when players learn their opponent’s group at zero cost. In this case, only states where all the players in one group buy the information can be stochastically stable: this result was not possible in the analysis of Neary. Overall, controlling for the cost equal to zero may be seen as a robustness exercise; nevertheless, we find that the model is more tractable under this specification than under Neary’s one. Indeed, if the cost is equal to zero, we can consider inter-group dynamics separated from inside-group ones, and hence, we can consider two absorbing states at a time.
The behavioral interpretation is similar for high and low levels of the cost: either the minority adapts to the majority, or the weaker group in preferences adapt to the strongest. Indeed, when the cost is low, the weakest group always needs to buy the information, while the strongest group does not, since it plays its favorite action with everyone. Similarly, when the cost is high, everybody will play the action favored by the strongest group in preferences in the long run. However, comparing the high-cost case with the low-cost case enriches the previous analysis. From this comparison, we can say that reducing the cost of learning the opponent’s group increases the probability of inter-group coordination in the long run. Indeed, inter-group miscoordination does not occur without incomplete information and a high cost. Unlike in Neary, strength in preferences or group size alone does not cause inter-group miscoordination.
The paper is organized as follows: In Section 2, we explain the model’s basic features. In Section 3, we determine the results for the complete information case where the cost is 0. In Section 4, we derive the results for the case with incomplete information and costly acquisition. We distinguish between two cases: low cost and high cost. In Section 5, we discuss results, and in Section 6, we conclude. We give all proofs in the Appendix A and we give the intuition during the text.

2. The Model

We consider N players divided into two groups A and B, N = N A + N B . We assume N A > N B + 1 and N B > 1 . Each period, players are randomly matched in pairs to play the 2 × 2 coordination game represented in Table 1, Table 2 and Table 3. Matching occurs with uniform probability, regardless of the group. Table 1 and Table 2 represent inside-group interactions, while Table 3 represents inter-group interactions (group A row player and group B column player). We assume that Π A > π A , and thus, we name a the favorite action of group A. Equally, we assume Π B > π B , and hence, b is the favorite action of group B. We do not assume any particular order between Π B , and Π A . However, without loss of generality, we assume that Π A + π A = Π B + π B . Consider K { A , B } , and K K { A , B } . We say that group K is stronger in preferences for its favorite action than group K if Π K > Π K or equivalently π K < π K .
Each period, players choose whether to pay a cost to learn their opponent’s group or not before choosing between action a and b. If they do not pay it, they do not learn the group of their opponent, and they play one single action valid for both groups. If they pay it, they can condition the action on the two groups. We call information choice the first and coordination choice the second.
Consider player i K . τ i is the information choice of player i: if τ i = 0 , player i does not learn the group of her/his opponent. If τ i = 1 , player i pays a cost c and learns the group. We assume that c 0 . x 0 i a , b is the coordination choice when τ i = 0 . If τ i = 1 , x 1 i K a , b is the coordination choice when player i meets group K, while x 1 i K a , b is the coordination choice when player i meets group K .
A pure strategy of a player consists of her/his information choice, τ i , and of her/his coordination choices conditioned on the information choice, i.e.,
s i = τ i , x 0 i , x 1 i K , x 1 i K S = 0 , 1 × { a , b } 3 .
Each player has sixteen strategies. However, we can safely neglect some strategies because they are both payoff-equivalent (a player earns the same payoff disregarding which strategy s/he chooses) and behaviorally equivalent (a player earns the same payoff independently from which strategy the other players play against her/him).
We consider a model of noisy best-response learning in discrete time (see Kandori et al. [1], Young [9]).
Each period t = 0 , 1 , 2 , , independently from previous events, there is a positive probability p ( 0 , 1 ) that a player is given the opportunity to revise her/his strategy. When such an event occurs, each player who is given the revision opportunity chooses with positive probability a strategy that maximizes her/his payoff at period t. s i ( t ) is the strategy played by player i at period t. U s i ( s , s i ) is the payoff of player i that chooses strategy s against the strategy profile s i played by all the other players except i. Such a payoff depends on the random matching assumption and the payoffs of the underlying 2 × 2 game. At period t + 1 , player i chooses
s i ( t + 1 ) arg max s S U s i ( s , s i ( t ) ) .
If there is more than one strategy that maximizes the payoff, player i assigns the same probability to each of those strategies. The above dynamics delineates a Markov process that is ergodic thanks to the noisy best response property.
We group the sixteen strategies into six analogous classes that we call behaviors. We name behavior a ( b ) as the set of strategies when player i K chooses τ i = 0 , and x 0 i = a ( b ) . We name behavior a b as the set of strategies when player i chooses τ i = 1 , x 1 i K = a , and x 1 i K = b , and so on. Z is the set of possible behaviors: Z = ( a , b , a b , b a , a a , b b ) . z i ( t ) is the behavior played by player i at period t as implied from s i ( t ) . z i ( t ) is the behavior profile played by all the other players except i at period t as implied from s i ( t ) . Note that behaviors catch all the relevant information as defined when players are myopic best repliers. U z i ( z , z i ( t ) ) is the payoff for player i that chooses behavior z against the behavior profile z i ( t ) . Such a payoff depends on the random matching assumption and the payoffs of the underlying 2 × 2 game. The dynamics of behaviors as implied by strategies coincide with the dynamics of behaviors, assuming that players myopically best reply to a behavior profile. We formalize the result in the following lemma.
Lemma 1.
Given the dynamics of z i ( t + 1 ) as implied by s i ( t + 1 ) , it holds that z i ( t + 1 ) a r g m a x z Z U z i ( z , z i ( t ) ) .
Consider a player i A such that the best thing to do for her/him is to play a with every player s/he meets regardless of the group. In this case, both ( 0 , a , a , b ) and ( 0 , a , b , b ) maximize her/his payoff. In contrast, ( 0 , b , a , b ) does not maximize her/his payoff since in this case, s/he plays b with every player s/he meets. Moreover, the payoff of player i is equal whether s i = ( 0 , a , a , b ) N 1 or s i = ( 0 , a , b , b ) N 1 but different if s i = ( 0 , b , a , b ) N 1 . Therefore, all the strategies that belong to the same behavior are payoff equivalent and behaviorally equivalent.
A further reduction is possible because a a ( b b ) is behaviorally equivalent to a ( b ) for each player. The last observation and the fact that we are interested in the number of players playing a with each group lead us to introduce the following state variable. We denote with n A A ( n B B ) the number of players of group A ( B ) playing action a with group A ( B ) , and n A B ( n B A ) the number of players of group A ( B ) playing action a with group B ( A ) . We define states as vectors of four components: ω = n A A , n A B , n B A , n B B , with Ω being the state space and ω t = n t A A , n t A B , n t B A , n t B B the state at period t. At each t, all the players know all the components of ω t . Consider player i playing behavior z i ( t ) at period t. U z i ( t ) i ( z , ω t ) is the payoff of i if s/he chooses behavior z at period t + 1 against the state ω t . All that matters for a decision-maker is ω t and z i ( t ) . We formalize the result in the following lemma.
Lemma 2.
Given the dynamics of ω t + 1 generated by z i ( t + 1 ) , it holds that U z i ( z , z i ( t ) ) = U z i ( t ) i ( z , ω t ) . Moreover, U a i ( z , ω t ) = U a a i ( z , ω t ) = U a b i ( z , ω t ) , and U b i ( z , ω t ) = U b b i ( z , ω t ) = U b a i ( z , ω t ) .
If players are randomly matched, it is as if each player plays against the entire population. Therefore, each player of group K myopically best responds to the current period by looking at how many players of each group play action a with group K. Moreover, a player that is given the revision opportunity subtracts her/himself from the component of ω t where s/he belongs. If i K is playing behavior a, a a or a b at period t, s/he knows that n t K K 1 players of group K are playing action a with group K at period t.
Define with θ t + 1 the set of players that are given the revision opportunity at period t. Given Lemma 2, it holds that ω t + 1 depends on ω t and on θ t + 1 . That is, we can define a map F ( · ) such that ω t + 1 = F ( ω t , θ t + 1 ) . The set θ t + 1 reveals whether the players who are given the revision opportunity are playing a behavior between a, a a , and a b , or a behavior between b, b b , and b a . In the first case we should look at U a i , while in the second at U b i .
From now on, we will refer to behaviors and states following the simplifications described above.
We illustrate here the general scheme of our presentation. We divide the analysis into two cases: complete information and incomplete information. For each case, we consider unperturbed dynamics (players choose the best reply behavior with probability 1) and perturbed dynamics (players choose a random behavior with a small probability). First, we help the reader understand how each player evaluates her/his best reply behavior and which states are absorbing. Second, we highlight the general structure of the dynamics with perturbation and then determine the stochastically stable states. In the next section, we analyze the case with complete information, hence, when the cost is zero.

3. Complete Information with Free Acquisition

In this section, we assume that each player can freely learn the group of her/his opponent when randomly matched with her/him. Without loss of generality, we assume that players always learn the group of their opponent in this case. We refer to this condition as free information acquisition. Each player has four possible behaviors as defined in the previous section. Z = a a , a b , b a , b b , with a = a a , and b = b b in this case.
Define π a K = Π A if K = A π B if K = B and π b K = π A if K = A Π B if K = B .
Equations (1)–(4) are the payoffs for a player i K playing a a or a b at period t.
U a i ( a a , ω t ) = n t K K 1 N 1 π a K + n t K K N 1 π a K ,
U a i ( a b , ω t ) = n t K K 1 N 1 π a K + N K n t K K N 1 π b K ,
U a i ( b a , ω t ) = N K n t K K N 1 π b K + n t K K N 1 π a K ,
U a i ( b b , ω t ) = N K n t K K N 1 π b K + N K n t K K N 1 π b K .

3.1. Unperturbed Dynamics

We begin the analysis for complete information by studying the dynamics of the system when players play their best reply behavior with probability one.
We can separate the dynamics of the system into three different dynamics. The two regarding inside-group interactions, i.e., n t A A and n t B B , and the one regarding inter-group interaction, i.e., n t A B and n t B A . We call this subset of states n t I = n t A B , n t B A . Both n t A A and n t B B are one-dimensional; n t I instead is two-dimensional.
Lemma 3.
Under free information acquisition, n t + 1 A A = F 1 ( n t A A , θ t + 1 ) , n t + 1 B B = F 4 ( n t B B , θ t + 1 ) and ( n t + 1 A B , n t + 1 B A ) = F 2 , 3 ( n t A B , n t B A , θ t + 1 ) .
The intuition behind the result is as follows. If players always learn their opponent’s group, the inter-group dynamics does not interfere with the inside-group and vice-versa. If player i K is given the revision opportunity, s/he chooses x 1 i K only based on n t K K .
Consider a subset of eight states: ω R = { N A , N A , N B , N B , 0 , N A , N B , N B ,
N A , N A , N B , 0 , N A , 0 , 0 , N B , 0 , N A , N B , 0 , N A , 0 , 0 , 0 , 0 , 0 , 0 , N B and 0 , 0 , 0 , 0 } .
Lemma 4.
Under free information acquisition, the states in ω R are the unique absorbing states of the system.
We call ( N A , N A , N B , N B ) and ( 0 , 0 , 0 , 0 ) Monomorphic States ( M S from now on). Specifically, we refer to the first one as M S a and to the second as M S b . We label the remaining six as Polymorphic States ( P S from now on). We call ( N A , N A , N B , 0 ) P S a and ( N A , 0 , 0 , 0 ) P S b . In M S , every player plays the same action with any other player; in P S , at least one group is conditioning the action. In M S a , every player plays a a ; in M S b , every player plays b b . In P S a , group A plays a a and group B plays b a . In P S b , group A plays a b while group B plays b b . In both P S a and P S b , all players coordinate on their favorite action with their similar.
In the model of Neary, only three absorbing states were possible: the two M S and a Type Monomorphic State where group A plays a a and group B plays b b . The P S were not present in the previous analysis. We observe these absorbing states in our analysis, thanks to the possibility of conditioning the action on the group.
We can break the absorbing states in ω R into the three dynamics in which we are interested. This simplification helps in understanding why only these states are absorbing. For instance, in inter-group interactions, there are just two possible absorbing states, namely ( N A , N B ) and ( 0 , 0 ) . For what concerns inside-group interactions, N A and 0 matter for n t A A , and N B and 0 for n t B B . For each dynamic, the states where every player plays a or where every player plays b with one group are absorbing. In this simplification, we can see the importance of Lemma 3. As a matter of fact, in all the dynamics we are studying, there are just two candidates to be stochastically stable. This result simplifies the stochastic stability analysis.

3.2. Perturbed Dynamics

We now introduce perturbations in the model presented in the previous section; that is, players can make mistakes while choosing their behaviors: there is a small probability that a player does not choose her/his best response behavior when s/he is given the revision opportunity. We use tools and concepts developed by Freidlin andWentzell [10] and refined by Ellison [11].
Given perturbations, ω t + 1 depends on ω t , θ t + 1 and on which players make a mistake among those who are given the revision opportunity. We define with ψ t + 1 the set of players who do not choose their best reply behavior among those who are given the revision opportunity. Formally, ω t + 1 = F ( ω t , θ t + 1 , ψ t + 1 ) .
We use uniform mistakes: the probability of making a mistake is equal for every player and every state. At each period, if a player is given the revision opportunity, s/he makes a mistake with probability ε . In this section, we assume that players make mistakes only in the coordination choice: assuming c = 0 , adding mistakes also in the information choice would not influence the analysis. Note that Lemma 3 is still valid under this specification.
If we consider a sequence of transition matrices P ε ε > 0 , with associated stationary distributions μ ε ε > 0 , by continuity, the accumulation point of μ ε ε > 0 that we call μ , is a stationary distribution of P : = lim ε 0 P ε . Mistakes guarantee the ergodicity of the Markov process and the uniqueness of the invariant distribution. We are interested in states which have positive probability in μ .
Definition 1.
A state ω ¯ is stochastically stable if μ ( ω ¯ ) > 0 and it is uniquely stochastically stable if μ ( ω ¯ ) = 1 .
We define some useful concepts from Ellison [11]. Let ω ¯ be an absorbing state of the unperturbed process. D ( ω ¯ ) is the basin of attraction of ω ¯ : the set of initial states from which the unperturbed Markov process converges to ω ¯ with probability one. The radius of ω ¯ is the number of mistakes needed to leave D ( ω ¯ ) when the system starts in ω ¯ . Define a path from state ω ¯ to state ω as a sequence of distinct states ( ω 1 , ω 2 , , ω T ) , with ω 1 = ω ¯ and ω T = ω . Υ ( ω ¯ , ω ) is the set of all paths from ω ¯ to ω . Define r ( ω 1 , ω 2 , , ω T ) as the resistance of the path ( ω 1 , ω 2 , , ω T ) , namely the number of mistakes that occurs to pass from state ω ¯ to state ω . The radius of ω ¯ is then
R ( ω ¯ ) = min ( ω 1 , ω 2 , , ω T ) Υ ( ω ¯ , Ω D ( ω ¯ ) ) r ( ω 1 , ω 2 , , ω T ) .
Now define the Coradius of ω ¯ as
C R ( ω ¯ ) = max ω D ( ω ¯ ) min ( ω 1 , ω 2 , , ω T ) Υ ( ω , D ( ω ¯ ) ) r ( ω 1 , ω 2 , , ω T )
Thanks to Theorem 1 in Ellison [11], we know that if R ( ω ¯ ) > C R ( ω ¯ ) , then ω ¯ is uniquely stochastically stable.
We are ready to calculate the stochastically stable states under complete information.
Theorem 1.
Under free information acquisition, for N large enough, if π B π A < N B N A , then P S b is uniquely stochastically stable. If π B π A > N B N A , then P S a is uniquely stochastically stable.
When the cost is null, players can freely learn the group of their opponent. Therefore, in the long run, they succeed in coordinating on their favorite action with those who are similar in preference. Hence, n t A A always converges to N A , and n t B B always converges to 0. This result rules out Monomorphic States and the other four Polymorphic States: only P S a and P S b are left. Which of the two is selected depends on strength in preferences and group size. Two effects determine the results in the long run. Firstly, if π A = π B , P S a is uniquely stochastically stable. The majority prevails in inter-group interactions if the two groups are equally strong in preferences.
Secondly, if π A π B , there is a trade-off between strength in preferences and group size. If π B π A > N B N A , either group A is stronger in preferences than group B, or group A is sufficiently larger than group B. In both of the two situations, the number of mistakes necessary to leave P S a is bigger than the one to leave P S b : in a sense, more mistakes are needed to make b best reply for A players than to make a best reply for B players. Therefore, every player will play action a in inter-group interactions. Similar reasoning applies if π B π A < N B N A .
Interestingly, in both cases, only players of one group need to learn their opponent’s group: the players from the group that is weaker in preferences or sufficiently smaller than the other.
Unlike in the analysis of Neary, if learning the opponent’s group is costless, the Monomorphic States are never stochastically stable. This result is a consequence of the possibility to condition the action on the group. Indeed, if players can freely learn the opponent’s group, they will always play their favorite action inside the group.
We provide two numerical examples to explain how the model works in Figure 1 and Figure 2. We represent just n t I , hence, a two-dimensional dynamics. Red states represent the basin of attraction of ( 0 , 0 ) , while green states the one of ( N A , N B ) . From gray states, there are paths of zero resistance both to ( 0 , 0 ) and to ( N A , N B ) . Any path that involves more players playing a within red states has a positive resistance. Every path that involves fewer people playing a within green states has a positive resistance. The radius of ( 0 , 0 ) is equal to the coradius of ( N A , N B ) , and it is the minimum resistance path from ( 0 , 0 ) to gray states. The coradius of ( 0 , 0 ) is equal to the radius of ( N A , N B ) , and it is the minimum resistance path from ( N A , N B ) to gray states.
Firstly, consider the example in Figure 1. N A = 10 , N B = 5 , π A = 8 , Π A = 10 , π B = 3 , Π B = 15 . Clearly, π B π A = 3 8 < 5 10 = N B N A . In this case R ( 10 , 5 ) = C R ( 0 , 0 ) = 1 , while R ( 0 , 0 ) = C R ( 10 , 5 ) = 3 . Hence, ( 0 , 0 ) is the uniquely stochastically stable state. We give here a short intuitive explanation. Starting from ( 0 , 0 ) , the minimum-resistance path to gray states is the one that reaches ( 0 , 3 ) . The minimum resistance path from ( 10 , 5 ) to gray states is the one that reaches ( 9 , 5 ) . Hence, fewer mistakes are needed to exit from the green states than to exit from the red states, and P S b = ( 10 , 0 , 0 , 0 ) is uniquely stochastically stable.
Secondly, consider the example in Figure 2. N A = 10 , N B = 5 , π A = 3 , Π A = 15 , π B = 8 , Π B = 10 . Note that π B π A = 8 3 > 5 10 = N B N A . In this case, R ( 10 , 5 ) = C R ( 0 , 0 ) = 4 , C R ( 10 , 5 ) = R ( 0 , 0 ) = 1 . Hence, P S a = ( 10 , 10 , 5 , 0 ) is uniquely stochastically stable. In this case, the minimum resistance path from ( 10 , 5 ) to gray states is the one that reaches ( 6 , 5 ) or ( 10 , 1 ) . The one from ( 0 , 0 ) to gray states is the one that reaches ( 0 , 1 ) .

4. Incomplete Information with Costly Acquisition

In this section, we assume that each player can not freely learn the group of her/his opponent. Each player can buy this information at cost c > 0 . We refer to this condition as costly information acquisition. It is trivial to notice that Lemma 3 is not valid anymore. Indeed, since players learn the group of their opponent conditional on paying a cost, not every player pays it, and the dynamics are no longer separable.
This time, Z = a , b , a b , b a , a a , b b . It is trivial to show that there are four strictly dominant behaviors; indeed, U z i ( t ) i ( a a , ω t ) = U z i ( t ) i ( a , ω t ) c and U z i ( t ) i ( b b , ω t ) = U z i ( t ) i ( b , ω t ) c . Hence, U z i ( t ) i ( a a , ω t ) < U z i ( t ) i ( a , ω t ) and U z i ( t ) i ( b b , ω t ) < U z i ( t ) i ( b , ω t ) , i N and ω t Ω . We define strictly dominant behaviors as Z o = a , b , a b , b a , with z i o being a strictly dominant behavior of player i.
Equations (5)–(8) are the payoffs at period t, for a player i K currently playing a or a b .
U a i ( a , ω t ) = n t K K + n t K K 1 N 1 π a K ,
U a i ( b , ω t ) = N n t K K n t K K N 1 π b K ,
U a i ( a b , ω t ) = n t K K 1 N 1 π a K + N K n t K K N 1 π b K c ,
U a i ( b a , ω t ) = N K n t K K N 1 π b K + n t K K N 1 π a K c .
Note that if c = 0 , then a a = a and b b = b . We begin the analysis with the unperturbed dynamics.

4.1. Unperturbed Dynamics

So far, there are no more random elements with respect to Section 3. Therefore, ω t + 1 = F ( ω t , θ t + 1 ) . Nine states can be absorbing under this specification.
Lemma 5.
Under costly information acquisition, there are nine possible absorbing states: ω R ( N A , N A , 0 , 0 ) .
We summarize all the relevant information in Table 4. The reader can note two differences with respect to Section 3: firstly, some states are absorbing if and only if some conditions hold, and secondly, there is one more possible absorbing state, that is, ( N A , N A , 0 , 0 ) . Such an absorbing state was also possible in Neary under the same conditions on payoffs and group size.
Where we write “none”, we mean that a state is always absorbing for every value of group size, payoffs, and/or the cost. We name ( N A , N A , 0 , 0 ) the Type Monomorphic State ( T S from now on): each group is playing its favorite action in this state, causing miscoordination in inter-group interactions. In both M S and T S , no player is buying the information, while in P S , at least one group is buying the information.
Monomorphic States are absorbing states for every value of group size, payoffs, and cost. Indeed, when each player is playing one action with any other player, players do not need to learn their opponent’s group (the information cost does not matter): they best reply to these states by playing the same action.
Polymorphic States are absorbing if and only if the cost is low enough: if the cost is too high, buying the information is too expensive, and players best reply to Polymorphic States by playing a or b. The Type Monomorphic State is absorbing if group B is either sufficiently close in size to group A or strong enough in preferences for its favorite action and if the cost is high enough. The intuition is the following. On the one hand, if the cost is high and if group B is weak in preferences or small enough, every player of group B best replies to T S by playing a. On the other hand, if the cost is low enough, every player best replies to this state by buying the information and conditioning the action.

4.2. Perturbed Dynamics

We now introduce perturbed dynamics. In this case, we assume that players can make two types of mistakes: they can make a mistake in the information choice and in the coordination choice. Choosing the wrong behavior, in this case, can mean both. We say that with probability η , a player who is given the revision opportunity at period t chooses to buy the information when it is not optimal. With probability ε , s/he makes a mistake in the coordination choice. We could have chosen to set only one probability of making a mistake with a different behavior or strategy.
The logic behind our assumption is to capture behaviorally relevant mistakes. We assume a double punishment mechanism for players choosing by mistake the information level and the coordination action. Specifically, our mistake counting is not influenced by our definition of behaviors. We could have made the same assumption starting from the standard definition of strategies assuming that players can make different mistakes in choosing the two actions that constitute the strategy. Our assumption is in line with works such as Jackson and Watts [12] and Bhaskar and Vega-Redondo [13], which assume mistakes in the coordination choice and the link choice.
Formally, ω t + 1 = F ( ω t , θ t + 1 , ψ t + 1 c ) , where ψ t + 1 c = { ψ t + 1 ε , ψ t + 1 η } is the set of players who make a mistake at period t among those who are given the revision opportunity, ψ t + 1 ε is the set of players who make a mistake in the coordination choice, and ψ t + 1 η the set of players that make a mistake in the information choice.
Since we assume two types of mistakes, the concept of resistance changes. We then need to consider three types of resistances. We call r ε ( ω t , , ω s ) the path from state ω t to state ω s with ε mistakes (players make a mistake in the coordination choice). We call r η ( ω t , , ω s ) the path with η mistakes (players make a mistake in the information choice). Finally, we call r ε η ( ω t , , ω s ) the path with mistakes both in the coordination choice and the information choice. Since we do not make further assumptions on ε and η (probability of making mistakes uniformly distributed), we can assume η ε .
We count each mistake in the path of both ε and η mistakes as 1; however, r ε η ( ω t , , ω s ) is always double since it implies a double mistake. Indeed, we can see this kind of mistake as the sum of two components, one in η and the other in ε , namely r ε η ( ω t , , ω s ) = r ε η | ε ( ω t , , ω s ) + r ε η | η ( ω t , , ω s ) .
For example, think about ω t = M S a , and that one player from B is given the revision opportunity at period t. Consider the case where s/he makes a mistake both in the information choice and in the coordination choice. For example, s/he learns the group and s/he plays a with A and b with B. This mistake delineates a path from M S a to the state ( N A , N A , N B , N B 1 ) of resistance r ε η ( M S a , , ( N A , N A , N B , N B 1 ) ) = 2 . Next, think about ω t = T S : the transition from T S to ( N A , N A 1 , 0 , 0 ) happens with one η mistake. One player from A should make a mistake in the information choice and optimally choosing a b . In this case, r η ( T S , , ( N A , N A 1 , 0 , 0 ) ) = 1 . With a similar reasoning, r ε ( M S a , , ( N A 1 , N A 1 , N B , N B ) ) = 1 : a player of group A makes a mistake in the coordination choice and chooses b.
Before providing the results, we explain why using behaviors instead of strategies does not influence the stochastic stability analysis. Let us consider all the sixteen strategies as presented in Section 2, and just one kind of mistake in the choice of the strategy. Let us take two strategies s , s z and a third strategy s z . Now consider the state ω ¯ , where s i = s , i N , and the state ω , where s i = s , i { 0 , , N m 1 } and s j = s , j { N m , , N } . Since s and s are both payoff-equivalent and behaviorally equivalent, s and s are the best reply strategies i N in both states ω ¯ and ω . Therefore at each period, every player who is given the revision opportunity in state ω ¯ or ω chooses s and s with equal probability. Now let us consider the state ω ¯ where s i = s , i N . When considering the transition between ω ¯ and ω ¯ , the number of mistakes necessary for this transition is the same whether the path passes through ω or not because the best reply strategy is the same in both ω and ω ¯ . Therefore, when computing the stochastically stable state, we can neglect s and ω .
We divide this part of the analysis into two cases, the first one where the cost is low and the second one when the cost is high.

4.2.1. Low Cost

This section discusses the case when c is as low as possible but greater than 0.
Corollary 1.
Under costly information acquisition, if 0 < c < 1 N 1 min { π A , π B } , M S and P S are absorbing states, while T S is not an absorbing state.
The proof is straightforward from Table 4. In this case, there are eight candidates to be stochastically stable equilibria.
Theorem 2.
Under costly information acquisition, for large enough N, take 0 < c < 1 N 1 min { π A , π B } . If π B π A < N B N A , then P S b is uniquely stochastically stable. If π B π A > N B N A , then P S a is uniquely stochastically stable.
The conditions are the same as in Theorem 1. When the cost is low enough, whenever a player can buy the information, s/he does it. Consequently, the basins of attraction of both Monomorphic States and Polymorphic States have the dimension they had under free information acquisition. Due to these two effects, the results are the same as under free information acquisition. This result is not surprising per se but serves as a robustness check of the results of Section 3.2.

4.2.2. High Cost

In this part of the analysis, we focus on a case when only M S and T S are absorbing states.
Define the following set of values:
Ξ P S = N B π A , N A π B , N B 1 Π B , N A 1 π A , N B Π A , N B 1 π B .
Corollary 2.
Under costly information acquisition, if c > 1 N 1 max { Ξ P S } and π B Π B < N B 1 N A , then only M S and T S are absorbing states. If π B Π B N B 1 N A , then only M S are absorbing states.
The proof is straightforward from Table 4, and therefore, we omit it. We previously gave the intuition behind this result. Let us firstly consider the case in which T S is not an absorbing state, hence, the case when π B Π B N B 1 N A .
Theorem 3.
Under costly information acquisition, for large enough N, take π B Π B N B 1 N A and c > 1 N 1 max { Ξ P S } . If N A > 2 N π A + Π A π A Π A + π A , then M S a is uniquely stochastically stable. If N A < 2 N π A + Π A π A Π A + π A , then M S b is uniquely stochastically stable.
If group A is sufficiently large or strong enough in preferences, the minimum number of mistakes to exit from the basin of attraction of M S a is higher than the minimum number of mistakes to exit from the one of M S b . Therefore, M S a is uniquely stochastically stable: every player plays behavior a in the long run.
Now we analyze the case when also T S is a strict equilibrium.
Theorem 4.
Under costly information acquisition, for large enough N, take π B Π B < N B 1 N A and c > 1 N 1 max { Ξ P S } .
  • If N π B π A > N B Π B N A π B Π B + π B + Π A , then M S a is uniquely stochastically stable.
  • If N π A π B > N A Π A N B π A Π A + Π B + π A , then M S b is uniquely stochastically stable.
  • If min N A Π A N B π A + π A , N B Π B N A π B + π B Π A Π B > N π A + π B , then T S is uniquely stochastically stable.
Moreover, when all the above conditions simultaneously do not hold:
  • If N π B π A > N B ( Π B + π A ) N A ( Π A + π B ) + Π A π A + π B Π B , then M S a is uniquely stochastically stable.
  • If N π A π B > N A ( Π A + π B ) N B ( Π B + π A ) Π A + π A π B + Π B , then M S b is uniquely stochastically stable.
  • If N π A π B = N A ( Π A + π B ) N B ( Π B + π A ) Π A + π A π B + Π B , then both M S a and M S b are stochastically stable.
We divide the statement of the theorem into two parts for technical reasons. However, the reader can understand the results from the first three conditions. The first condition expresses a situation where group A is stronger in preferences than group B or group A is sufficiently larger than group B. In this case, there is an asymmetry in the two costs for exiting the two basins of attraction of M S a and M S b . Exit from the first requires more mistakes than exit from the second. Moreover, reaching M S a from T S requires fewer mistakes than reaching M S b from T S . For this reason, R ( M S a ) > C R ( M S a ) and M S a is uniquely stochastically stable in this case. A similar reasoning applies to the second condition.
The third condition expresses a case where both groups are strong enough in preferences or have sufficiently similar sizes. Many mistakes are required to exit from T S , compared to how many mistakes are required to reach T S from the two M S . Indeed, T S is the state where both groups are playing their favorite action. Since they are both strong in preferences or large enough, in this case, all the players play their favorite action in the long run, but they miscoordinate in inter-group interactions.
The results of Theorems 3 and 4 reach the same conclusions as Neary. However, our analysis allows us to affirm that only with a high cost, the M S or the T S is stochastically stable. This result enriches the previous analysis.
As a further contribution, comparing these results with those in Section 4.2.1, we can give the two conditions for inter-group miscoordination to happen in the long run. First, the cost to pay to learn the opponent’s group should be so high that players never learn their opponent’s group. Second, both groups should be strong enough in preferences or sufficiently close in size. The following lemma states what happens when the cost takes medium values.
Lemma 6.
If 1 N 1 max π A , π B < c < 1 N 1 min { Ξ P S } , then the stochastically stable states must be in the set M = P S a , P S b , M S a , M S b .
When the cost lowers a tiny quantity from the level of Section 4.2.2, T S is not absorbing anymore. Therefore, only P S and M S can be stochastically stable when the cost is in the interval above. However, not all the P S can be stochastically stable, only the two where all the players play their favorite action in inside-group interactions. The intuition of this result is simple: if players condition their action on the groups in the long run, they play their favorite action with those with similar preferences.
We do not study when M S are stochastically stable or when P S are: we leave this question for future analysis. Nevertheless, given the results of Section 4.2.1 and Section 4.2.2, we expect that for higher levels of cost, M S is stochastically stable, and for lower levels, P S is stochastically stable.

5. Discussion

The results of our model involve three fields of the literature. Firstly, we contribute to the literature on social conventions. Secondly, we contribute to the literature on stochastic stability analysis, and lastly, we contribute to the literature on costly information acquisition.
For what concerns social conventions, many works in this field study the existence in the long run of heterogeneous strategy profiles. We started from the original model of Neary [7], which considers players heterogeneous in preferences, but with a smaller strategic set than ours (Heterogeneity has been discussed in previous works such as Smith and Price [14], Friedman [15], Cressman et al. [16], Cressman et al. [17] or Quilter [18]). Neary’s model gives conditions for the stochastic stability of a heterogeneous strategy profile that causes miscoordination in inter-group interactions in a random matching case. Neary and Newton [19] expands the previous idea to investigate the role of different classes of graphs on the long-run result. It finds conditions on graphs such that a heterogeneous strategy profile is stochastically stable. It also considers the choice of a social planner that wants to induce heterogeneous or homogeneous behavior in a population.
Carvalho [20] considers a similar model, where players choose their actions from a set of culturally constrained possibilities and the heterogeneous strategy profile is labeled as miscoordination. It finds that cultural constraints drive miscoordination in the long run. Michaeli and Spiro [21] studies a game between players with heterogeneous preferences and who feel pressure from behaving differently. Such a study characterizes the circumstances under which a biased norm can prevail on a non-biased norm. Tanaka et al. [22] studies how local dialects survive in a society with an official language. Naidu et al. [23] studies the evolution of egalitarian and inegalitarian conventions in a framework with asymmetry similar to the language game. Likewise, Belloc and Bowles [24] examines the evolution and the persistence of inferior cultural conventions.
We introduce the assumption that players can condition the action on the group if they pay a cost. This assumption helps to understand the conditions for the stability of the Type Monomorphic State, where players miscoordinate in inter-group interactions. We show that a low cost favors inter-group coordination: incomplete information, high cost, strength in preferences, and group size are key drivers for inter-group miscoordination. Like many works in this literature, we show the importance of strength in preferences and group size in the equilibrium selection.
Concerning network formation literature, Goyal et al. [25] conducts an experiment on the language game, testing whether players segregate or conform to the majority. van Gerwen and Buskens [26] suggests a variant of the language game similar to our version but in a model with networks to study the influence of partner-specific behavior on coordination. Concerning auctions theory, He [27] studies a framework where each individual of a population divided into two types has to choose between two skills: a “majority” and a “minority” one. It finds that minorities are advantaged in competition contexts rather than in coordination ones. He and Wu [28] tests the role of compromise in the battle of sexes with an experiment.
Like these works, we show that group size and strength in preferences matter for the long-run equilibrium selection. The states where the action preferred by the minority is played in most of the interactions ( M S b or P S b ) are stochastically stable provided that the minority is strong enough in preferences or sufficiently large.
A parallel field is the one of bilingual games such as the one proposed by Goyal and Janssen [29] or Galesloot and Goyal [30]: these models consider situations in which players are homogeneous in preferences towards two coordination outcomes, but they can coordinate on a third action at a given cost.
Concerning the technical literature on stochastic stability, we contribute by applying standard stochastic stability techniques to an atypical context, such as costly information acquisition. Specifically, we show that with low cost levels, Polymorphic States where all players in one group condition their action on the group are stochastically stable. Interestingly, only one group of players needs to learn their opponent’s group. With high cost levels, Monomorphic States where no player conditions her/his action on the group are stochastically stable. Since the seminal works by Bergin and Lipman [31] and Blume [32], many studies have focused on testing the role of different mistake models in the equilibrium selection. We use uniform mistakes, and introducing different models could be an interesting exercise for future studies.
Among the many models that can be used, there are three relevant variants: payoff/cost-dependent mistakes (Sandholm [33], Dokumacı and Sandholm [34], Klaus and Newton [35], Blume [36] and Myatt and Wallace [37]), intentional mistakes (Naidu et al. [38] and Hwang et al. [39]), and condition-dependent mistakes (Bilancini and Boncinelli [40]). Important experimental works in this literature have been done by Lim and Neary [41], Hwang et al. [42], Mäs and Nax [43], and Bilancini et al. [44].
Other works contribute to the literature on stochastic stability from the theoretical perspective (see Newton [45] for an exhaustive review of the field). Recently, Newton [46] has expanded the domain of behavioral rules regarding the results of stochastic stability. Sawa and Wu [47] shows that with loss aversion individuals, the stochastic stability of Risk-Dominant equilibria is no longer guaranteed. Sawa and Wu [48] introduces reference-dependent preferences and analyzes the stochastic stability of best response dynamics. Staudigl [49] examines stochastic stability in an asymmetric binary choice coordination game.
For what concerns the literature on costly information acquisition, many works interpret the information’s cost as costly effort (see the seminal contributions by Simon [50] or Grossman and Stiglitz [51]). Our paper is one of those. Many studies place this framework in a sender-receiver game. This is the case of Dewatripont and Tirole [52], which builds a model of costly communication in a sender-receiver setup.
More recent contributions in this literature are Dewatripont [53], Caillaud and Tirole [54], Tirole [55], and Butler et al. [56]. Bilancini and Boncinelli [57] applies this model to persuasion games with labeling. Both Bilancini and Boncinelli [58] and Bilancini and Boncinelli [59] consider coarse thinker receivers, combining costly information acquisition with the theory of Jehiel [60]. Rational inattention is a recent field where the information cost is endogenous (see Mackowiak et al. [61] for an exhaustive review). We assume that the cost is exogenous and homogeneous for each player.
Güth and Kliemt [62] firstly uses costly information acquisition in evolutionary game theory in a game of trust. It finds conditions such that developing a conscience can be evolutionarily stable. More recently, Berger and De Silva [63] uses a similar concept in a deterrence game where agents can buy costly information on past behaviors of their opponents.
Many works use similar concepts of cost in the evolutionary game theory literature on coordination games. For example, Staudigl and Weidenholzer [64] considers a model where players can pay a cost to form links. The main finding is that when agents are constraint in the possible number of interactions, the payoff-dominant convention emerges in the long run.
The work by Bilancini and Boncinelli [65] extends Staudigl and Weidenholzer [64]. The model introduces the fact that interacting with a different group might be costly for a player. It finds that when this cost is low, the Payoff-Dominant strategy is the stochastically stable one. When the cost is high, the two groups in the population coordinate on two different strategies: one on the risk-dominant and the other on the payoff-dominant. Similarly, Bilancini et al. [66] studies the role of cultural intolerance and assortativity in a coordination context. In that model, there is a population divided into two cultural groups, and each group sustains a cost from interacting with the other group. It finds interesting conditions under which cooperation can emerge even with cultural intolerance.

6. Conclusions

We can summarize our results as follows. When players learn the group of their opponent at a low cost, they always coordinate: they play their favorite action with their similar, while in inter-group interactions, they play the favorite action of the group that is stronger in preferences or with large enough size. If the cost is high, players never learn the group of their opponent. All the players play the same action with every player, or they play their favorite action.
By comparing Section 4.2.1 and Section 4.2.2, we can see the impact of varying the cost levels on the long-run results. Surely a low cost favors inter-group coordination. However, a change in the cost level produces two effects that perhaps need further investigation. The first effect concerns the change in the payoff from the interactions between players. The second concerns the change in the purchase of the information.
Consider a starting situation where the cost is low. Players always coordinate on their favorite action in inside-group interactions. If the cost increases, players stop learning their opponent’s group (hence, they stop paying the cost), and they begin to play the same action as any other player. If this happens, either Monomorphic States are established in the long run, or the Type Monomorphic State emerges. In the first case, a group of players coordinates on its second best option, even in inside-group interactions. For this group, there could be a certain loss in terms of welfare. In the second case, players miscoordinate in inter-group interactions, and hence, all of them could have a certain loss in welfare.
Nevertheless, when the cost is low, there is a “free-riding” behavior that vanishes if the cost increases. In fact, with low cost levels, only one group needs to pay the cost, and the other never needs to pay it. In one case, players of group A play their favorite action both in inside-group and inter-group interactions; hence, they never need to pay the cost, while group B always needs to afford it. In the other case, the opposite happens. Therefore, when the cost increases, one of the two groups will benefit from not paying for the information anymore. Future studies could address the implications of this trade-off between successful coordination and the possibility of not paying the cost.
We conclude with a short comparison of our result with the one of Neary [7]. Indeed, it is worthwhile to mention a contrast that is a consequence of the possibility of conditioning the action on the group of the player. In the model of Neary, a change in the strength of preferences or the group size of one group does not affect the behavior of the other group. We can find this effect even in our model when the cost is high. For example, when M S a is stochastically stable and group B becomes strong enough in preferences or sufficiently large, the new stochastically stable state becomes T S . Therefore, group A does not change its behavior. However, when the cost is sufficiently low, the change in payoffs or group size of one group influences the other group’s behavior in inter-group interactions. For instance, when P S a is stochastically stable, if group B becomes strong enough in preferences or sufficiently large, P S b becomes stochastically stable. Hence, both groups change the way they behave in inter-group interactions.
Nevertheless, we can interpret similarly the passing from M S a to T S and the one from P S a to P S b . In both cases, both groups keep playing their favorite action in inside-group interactions, and what happens in inter-group interactions depends on strength in preferences and group size. Therefore, in this respect, the behavioral interpretation of our results is similar to Neary’s.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

I wish to thank my two supervisors Pietro Dindo and Ennio Bilancini, for their comments and support in building the model and writing the paper. I wish to thank Ulrich Berger (Editor-in-Chief), Jacopo Bizzotto, Leonardo Boncinelli, Fabio Michelucci, Jonathan Newton, and two anonymous referees for their helpful comments and suggestions. Remaining errors are mine.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A. Proofs

Proof of Lemma 1.
We have to formally show that each strategy inside the same behavior is behaviorally and payoff-equivalent for each player. Consider a player i K . Define g i K ( s i ) and g i K ( s i ) as the frequencies of successful coordination for i on action a with group K and K given strategy profile s i .
U s i ( ( 0 , a , a , a ) , s i ) = U s i ( ( 0 , a , b , b ) , s i ) = U s i ( ( 0 , a , b , a ) , s i ) = U s i ( ( 0 , a , a , b ) , s i ) = N K 1 N 1 g i K ( s i ) π a K + N K N 1 g i K ( s i ) π a K .
Therefore, if ( 0 , a , a , a ) is the maximizer, then also ( 0 , a , a , b ) , ( 0 , a , b , a ) , and ( 0 , a , b , b ) are so. Hence, in this case, i maximizes her/his payoff by choosing behavior a. Moreover, consider s i = ( 0 , a , a , b ) N 1 and s i = ( 0 , a , a , a ) N 1 . In this case g i K ( s i ) = g i K ( s i ) , so for g i K . Contrarily, if s i = ( 0 , b , a , a ) N 1 , g i K ( s i ) g i K ( s i ) , so for g i K . Therefore, U s i ( a , s i ) = U s i ( a , s i ) = U z i ( a , a N 1 ) . Thanks to symmetry in payoff matrix, the argument stands for all strategies and behaviors. This passage completes the proof. □
Proof of Lemma 2.
Consider a player i K currently playing behavior a that is given a revision opportunity at period t. g i K ( z i ( t ) ) is the frequency of successful coordinations of player i on action a with group K at period t, given z i ( t ) . In this case, U z i ( a , z i ( t ) ) = N K 1 N 1 g i K ( z i ( t ) ) π a K + N K N 1 g i K ( z i ( t ) ) π a K . Note that g i K ( z i ( t ) ) = g i K ( z i ( t ) , ω t ) and g i K ( z i ( t ) ) = g i K ( z i ( t ) , ω t ) , where g i K ( z i ( t ) , ω t ) is the frequency of successful coordinations of player i on action a with group K at period t, given ω t and that player i is currently playing z i ( t ) . Therefore, U z i ( a , z i ( t ) ) = U z i ( t ) i ( a , ω t ) , with z i ( t ) = a in our case.
Note that g i K ( a , ω t ) = n t K K 1 N K 1 , and g i K ( b , ω t ) = n t K K N K 1 . Moreover, g i K ( a , ω t ) = g i K ( a a , ω t ) = g i K ( a b , ω t ) , and g i K ( b , ω t ) = g i K ( b b , ω t ) = g i K ( b a , ω t ) . Contrarily, g i K ( z i ( t ) , ω t ) = g i K ( ω t ) = n t K K N K , z i ( t ) Z .
Therefore, U a i ( a , ω t ) = U a a i ( a , ω t ) = U a b i ( a , ω t ) . Equally, U b i ( a , ω t ) = U b b i ( a , ω t ) = U b a i ( a , ω t ) . Thanks to symmetry in payoff matrix, the argument stands for all strategies and behaviors. □

Appendix A.1. Proofs of Complete Information with Free Acquisition

Proof of Lemma 3.
Consider a player i K currently playing a a who is given the revision opportunity at period t. On the one hand, n t K K , U a i ( a b , ω t ) = U a i ( a a , ω t ) . On the other hand, n t K K , U a i ( b a , ω t ) = U a i ( a a , ω t ) . Therefore, player i chooses a a or a b depending on n t K K , and b a or a a depending on n t K K .
Moreover, if player i chooses a b instead of a a , n t + 1 K K = n t K K , but n t + 1 K K < n t K K . If player i chooses b a instead of a a , n t + 1 K K < n t K K , but n t + 1 K K = n t K K . This passage completes the proof. □
With abuse of notation, we call best reply (BR) the action optimally taken by a player in one of the three dynamics. For example, if a player of group A earns the highest payoff by playing a against a player of group B, we say that a is her/his BR. We do this in the context of complete information because of the separability of the dynamics.
Proof of Lemma 4.
Thanks to Lemma 3, we can consider the three separated dynamics: n t A A , n t B B , and n t I .
Inside-group interactions.
Firstly, we prove the result for n t A A , and then the argument stands for n t B B thanks to symmetry of payoff matrix. We have to show that all the states in ω R have an absorbing component for n t A A , that is, 0 or N A . When n A A = N A , i A , a is BR against group A at period t. Hence, F 1 ( N A , θ t + 1 ) = N A . Symmetrically, if n A A = 0 , b is always BR, and so F 1 ( 0 , θ t + 1 ) = 0 . Therefore, N A and 0 are fixed points for n t A A .
We need to show that these states are absorbing, that all the other states are transient, and that there are no cycles. Consider player i A , who is given the revision opportunity at period t. We define n ¯ A as the minimum number of A players such that a is BR, and n ̲ A as the maximum number of A players such that b is BR. From Equations (1)–(4), we know that n ¯ A = N A π A + Π A Π A + π A , and that n ̲ A = N A π A π A Π A + π A . Assume n t A A n ¯ A . There is always a positive probability that a player not playing a is given the revision opportunity. Hence, F 1 ( n t A A , θ t + 1 ) n t A A . Symmetrically, we can say that if n t A A < n ̲ A , F 1 ( n t A A , θ t + 1 ) n t A A .
We now prove that if n t A A n ̲ A 0 ,
P r lim t F 1 ( n t A A , θ t + 1 ) = n t A A = 0 .
Equally, if n t A A n ¯ A N A ,
P r lim t F 1 ( n t A A , θ t + 1 ) = n t A A = 0 .
We prove the first case, and the result stands for the second, thanks to symmetry in payoff matrices. Consider a period s in a state n s A A < n ̲ A 0 . For every player, b is BR. Define P r n s + 1 A A = n s A A = p . Such a probability represents the event that only players playing b are given the revision opportunity. P r n s + 2 A A = n s A A = p 2 , P r n s + k A A = n s A A = p k . If k , P r n s + k A A = n s A A = 0 . Therefore,
If n 0 A A n ̲ A , P r lim t F 1 ( n t A A , θ t + 1 ) = 0 = 1 ,
If n 0 A A n ¯ A , P r lim t F 1 ( n t A A , θ t + 1 ) = N A = 1 .
Next, consider n ̲ A < n 0 A A < n ¯ A . For every i playing a, b is BR, while, for every i playing b, a is BR. There are no absorbing states between these states. If only players playing a are given the revision opportunity, they all choose b, and if enough of them are given the revision opportunity, n 1 A A < n ̲ A . The opposite happens if only players playing b are given the revision opportunity.
Inter-group interactions.
We now pass to the analysis of n t I . We define four important values for n A B and n B A : T A = m i n n B A | n B A > π A N B Π A + π A , T B = m i n n A B | n A B > Π B N A Π B + π B ,
D A = m a x n B A | n B A < π A N B Π A + π A , and D B = m a x n A B | n A B < Π B N A Π B + π B .
Given these values, we also define two sets of states, Ω I b and Ω I a :
Ω I a = n I | n B A T A and n A B T B and Ω I b = n I | n B A D A and n A B D B .
With similar computation as for n t A A , we can say that ( 0 , 0 ) and ( N A , N B ) are two fixed points for n t I . Are they absorbing states?
Consider the choice of a player i A against player j B and vice-versa. There can be four possible combinations of states: states in which a is BR for every player, states in which b is BR for every player, states, in which i A , a is the best reply and b is the best reply j B , and states for which the opposite is true. Let us call the third situation as Ω I a b and the fourth as Ω I b a .
Firstly, we prove that Ω I a and Ω I b are the regions where a and b are BR for every player. Secondly, we prove that there is no other absorbing state in Ω I a than ( N A , N B ) , and no other absorbing state in Ω I b than ( 0 , 0 ) .
Assume that player i A is given a revision opportunity at period t. From Equations (1)–(4), a is the BR against group B if n t B A > π A N B Π A + π A . Since T A is defined as the minimum value s.t., the latter holds, n t B A T A , i A , a is BR against B groups. Now, assume that j B is given the revision opportunity, a is the BR against group A if n t A B > Π B N A Π B + π B . Since T B is defined as the minimum value s.t., this relation is true, n t A B T B , a is the best reply j B . Therefore, if n 0 I Ω I a , n s I Ω I a , s 0 . Similarly, if n 0 I Ω I b , n s I Ω I b , s 0 .
Consider being in a generic state ( T B + d , T A + d ) Ω I a at period t, with d [ 0 , N A T B ) and d [ 0 , N B T A ) . In such a state, there is always a probability p that a player not playing a is given the revision opportunity.
Therefore, if n t I Ω I a \ ( N A , N B ) , P r F 2 , 3 n t I , θ t + 1 n t I > p (meaning that n t I > n t I if either n t A B > n t A B and n t B A = n t B A or n t B A > n t B A and n t A B = n t A B or both n t B A > n t B A and n t A B > n t A B ). Similar to what we proved before,
if n t I Ω I a \ ( N A , N B ) , P r lim t F 2 , 3 ( n t I , θ t + 1 ) = n t I = 0 ,
if n t I Ω I b \ ( 0 , 0 ) , P r lim t F 2 , 3 ( n t I , θ t + 1 ) = n t I = 0 .
Consequently,
If n 0 I Ω I a P r lim t F 2 , 3 ( n t I , θ t + 1 ) = ( N A , N B ) = 1 ,
if n 0 I Ω I b , P r lim t F 2 , 3 ( n t I , θ t + 1 ) = ( 0 , 0 ) = 1 .
We now consider Ω I a b and Ω I b a . Take an n 0 I Ω I a b : at each period, there is a positive probability that only players of group A are given the revision opportunity, since for them a is the best reply, in the next period, there will be more or equal players in A playing a. Hence, if enough players of A that are currently playing b are given the revision opportunity, n 1 I Ω I a . By the same reasoning, there is also a positive probability that only players from B are given the revision opportunity; hence, that n 1 I Ω I b . The same can be said for every state in Ω I b a . Hence, starting from every state in Ω I a b Ω I b a , there is always a positive probability to end up in Ω I a or Ω I b . □
Lemma A1.
Under complete information,
P r lim t n t I = ( N A , N B ) = 1 P r lim t n t I = ( 0 , 0 ) .
P r lim t n t A A = N A = 1 P r lim t n t A A = 0 .
P r lim t n t B B = N B = 1 P r lim t n t B B = 0 .
Proof. 
We prove the result for n t I , and the argument stands for the two other dynamics thanks to symmetry in the payoff matrix. Firstly, note that whenever the process starts in Ω I a Ω I b , the lemma is always true thanks to the proof of Lemma 4. We need to show that this is the case, as well as when the process starts inside Ω I a b Ω I b a . We prove the result for Ω I a b using the same logic, and the result stands for Ω I b a for symmetry of payoff matrix.
Take n 0 I Ω I a b , define as p a the probability of extracting m players from A that are currently playing b, and who would change action a if given a revision opportunity. Define as p b the probability of picking m players from B currently choosing a who would change action to b if given a revision opportunity. The probability 1 p a p b defines all the other possibilities.
Let us take k steps forward in time:
P r n k I Ω I a ( p a ) k
P r n k I Ω I b ( p b ) k
P r n k I Ω I a b Ω I b a ( 1 p a p b ) k .
Consider period k + d :
P r n k + d I Ω I a ( p a ) k
P r n k + d I Ω I b ( p b ) k
P r n k + d I Ω I a b Ω I b a ( 1 p a p b ) k + d .
Clearly, the probability of being in Ω I a ( Ω I b ) is now greater than or equal to ( p a ) k ( ( p b ) k ) : we know that once in Ω I a ( Ω I b ) , the system stays there. The probability of being in Ω I a b Ω I b a consequently, is lower than ( 1 p a p b ) k + d .
Taking the limit for d that goes to infinity
lim d P r n k + d I Ω I a b Ω I b a = 0 .
This means that if we start in a state in Ω I a b there is no way of ending up in Ω I a b Ω I b a in the long run; hence, the system ends up either in Ω I a or in Ω I b , but given this, we know that it ends up either in ( 0 , 0 ) or in ( N A , N B ) . □
Corollary A1.
Under complete information,
P r lim t n t I = ( N A , N B ) = 1 IFF n 0 I Ω I a .
P r lim t n t I = ( 0 , 0 ) = 1 IFF n 0 I Ω I b .
P r lim t n t A A = N A = 1 IFF n 0 A A n ¯ A , N A , and
P r lim t n t A A = 0 = 1 IFF n 0 A A 0 , n ̲ A .
P r lim t n t B B = N B = 1 IFF n 0 B B n ¯ B , N B , and
P r lim t n t B B = 0 = 1 IFF n 0 B B 0 , n ̲ B .
This result is a consequence of the previous lemmas, and therefore, the proof is omitted. Since the only two absorbing states in the dynamics of n t I are ( 0 , 0 ) and ( N A , N B ) , they are the only two candidates to be stochastically stable states. From now on we call ( 0 , 0 ) as I n b and ( N A , N B ) as I n a . We define as 0 A the state where all players of group A play b with group A and 0 B the state where all players of group B play b with group B.
Let us call E A and E B the two values for which players in A and in B are indifferent in playing a or b in inter-group interactions. E A = N B π A Π A + π A and E B = N A Π B Π B + π B . From now on, we often use values of N large enough to compare the arguments inside ceiling functions.
Lemma A2.
Under free information acquisition, for large enough N, R ( I n b ) = C R ( I n a ) = E A for all values of payoffs and sizes of groups, while
R ( I n a ) = C R ( I n b ) = N A E B i f π B Π A < N B N A N B E A i f π B Π A > N B N A
Proof. 
Firstly, we know from Ellison [11] that if there are just two absorbing states, the radius of one is the coradius of the other and vice-versa. Hence, R ( I n b ) = C R ( I n a ) , and R ( I n a ) = C R ( I n b ) . Moreover, from the proof of Lemma 4, we know that D ( I n a ) = Ω I a and D ( I n b ) = Ω I b .
We prove that the minimum resistance path to exit the basin of attraction of I n b is the one that reaches ( E B , 0 ) or ( 0 , E A ) , and that the one to exit the basin of attraction of I n a is the one that reaches either ( E B , N B ) or ( N A , E A ) . To prove this statement for I n b , firstly, note that once inside Ω I b , every step that involves a passage to a state with more people playing a requires a mistake. Secondly, note that in a state that is out of Ω I b , at least one of the two groups is indifferent in playing b or a, in other words, in a state where either n A B = E B or n B A = E A or both. Hence, the minimum resistance path to exit from I n b is the one either to ( E B , 0 ) or to ( 0 , E A ) . It is straightforward to show that all the other paths have greater resistance than the two above. Since we use uniform mistakes, every mistake counts the same value, and without loss of generality, we can count each of them as 1. Since every resistance counts as 1, then R ( I n b ) = min { E B ; E A } = E A . Similarly, R ( I n a ) = min { N A E B ; N B E A } , and
N A E B < N B E A π B Π A < N B N A .
 □
Lemma A3.
Under free information acquisition, for large enough N, R ( 0 A ) = N A π A π A Π A + π A , R ( N A ) = N A Π A + Π A Π A + π A , R ( 0 B ) = N B Π B + Π B Π B + π B and R ( N B ) = N B π B π B Π B + π B .
Proof. 
The proof is straightforward; indeed, the minimum resistance path in terms of mistakes required to reach one absorbing state starting from the other one is the cost of exit from the basin of attraction of the first. As a matter of fact, let us consider R ( 0 A ) ; we know from the proof of Lemma 4 that we are out of the basin of attraction of 0 A when we reach the state n ̲ A . Hence, R ( 0 A ) = N A π A π A Π A + π A . The same applies to the other states. □
Proof of Theorem 1.
We divide the proof for the three dynamics described so far: for what concerns n t A A , N A is uniquely stochastically stable, and for what concerns n t B B , 0 B is uniquely stochastically stable; this proof follows directly from Lemma A3 and therefore is omitted. Let us pass to n t I . We know from Lemma A2 that R ( I n b ) = E A and that the value of R ( I n a ) depends on payoffs and group size. Let us firstly consider the case when π B Π A < N B N A and R ( I n a ) = N A E B . It is sufficient that E A > N A E B for I n b to be uniquely stochastically stable. Indeed, if this happens, R ( I n b ) > C R ( I n b ) . This is the case IFF
π A N B Π A + π A > π B N A Π B + π B π B π A < N B N A .
To complete the proof, we show that whenever π B π A > N B N A , then I n a is the uniquely stochastically stable state. Firstly, note that π B Π A < π B π A ; hence, for π B π A > N B N A > π B Π A , R ( I n a ) = N A E B and E A = R ( I n b ) . However, Equation (A1) is reversed, so, I n a is uniquely stochastically stable. For π B π A > π B Π A > N B N A , R ( I n a ) = N B E A and still R ( I n b ) = E A . In this case, I n a is the uniquely stochastically stable if E A < N B E A , hence, IFF
π A N B Π A + π A < Π A N B Π A + π A .
This happens for every value of the payoffs (given that Π A > π A ) and of the group size. We conclude that whenever π B π A < N B N A , P S b is uniquely stochastically stable and when π B π A > N B N A , P S a is uniquely stochastically stable. □

Appendix A.2. Proofs of Incomplete Information with Costly Acquisition

For convenience, we call behavior τ 1 the optimal behavior when a player decides to acquire the information: τ 1 = max ( a b , b a , a a , b b ) .
We will use in some proofs the concept of Modified Coradius from Ellison [11]. We write here the formal definition. Suppose ω ¯ is an absorbing state and ( ω 1 , ω 2 , , ω T ) is a path from state ω to ω ¯ . Let L 1 , L 2 , , L r = ω ¯ be the sequence of limit sets through which the path passes consecutively. The modified resistance is the original resistance minus the radius of the intermediate limit sets through which the path passes,
r * ( ω 1 , ω 2 , , ω T ) = r ( ω 1 , ω 2 , , ω T ) i = 2 r 1 R ( L i ) .
Define
r * ( ω , ω ¯ ) = min ( ω 1 , ω 2 , , ω T ) Υ ( ω , ω ¯ ) r * ( ω 1 , ω 2 , , ω T ) ,
the Modified Coradius is defined as follows
C R * ( ω ¯ ) = max ω ω ¯ r * ( ω , ω ¯ ) .
Note that C R * ( ω ¯ ) C R ( ω ¯ ) . Thanks to Theorem 2 in Ellison [11], we know that when R ( ω ¯ ) > C R * ( ω ¯ ) , ω ¯ is uniquely stochastically stable.
Proof of Lemma 5.
We first show that the nine states are effectively strict equilibria, that there is no other possible equilibrium, and that a state is absorbing if and only if it is a strict equilibrium.
Monomorphic States.
It is easy to show that ( N A , N A , N B , N B ) and ( 0 , 0 , 0 , 0 ) are two strict equilibria. We take the first case, and the argument stands also for the second, thanks to the symmetry of the payoff matrix. Consider player i K who is given the revision opportunity at period t:
U a i ( a , ω t ) = N K + N K 1 N 1 π a K = π a K ,
U a i ( b , ω t ) = N N K N K N 1 π b K = 0 ,
U a i ( τ 1 , ω t ) = N K + N K 1 N 1 π a K c = π a K c .
( N A , N A , N B , N B ) is a strict equilibrium since π a K > 0 and c > 0 .
Polymorphic States.
Firstly let us consider the case of ( N A , 0 , 0 , N B ) . Since in this case, every player is playing a b , the state is a strict equilibrium IFF max z i o = a b , i N . If player i K is given the revision opportunity at period t,
U a i ( a , ω t ) = N K 1 N 1 π a K ,
U a i ( b , ω t ) = N K N 1 π b K ,
U a i ( τ 1 , ω t ) = N K 1 N 1 π a K + N K N 1 π b K c .
For group A players, U a i ( a , ω t ) > U a i ( b , ω t ) since N B N 1 π A < N A 1 N 1 Π A . For group B players, U a i ( b , ω t ) > U a i ( a , ω t ) as N B 1 N 1 π B < N A N 1 Π B . U a i ( τ 1 , ω t ) is the highest of the three i N IFF c < 1 N 1 min N B π A , N B 1 π B .
Consider the case of ( 0 , N A , N B , 0 ) ; since every player is playing b a , it must be that max z i o = b a . i K faces the following payoffs
U b i ( a , ω t ) = N K N 1 π a K ,
U b i ( b , ω t ) = N K 1 N 1 π b K ,
U b i ( τ 1 , ω t ) = N K 1 N 1 π b K + N K N 1 π a K c .
Note that U b i ( a , ω t ) > U b i ( b , ω t ) IFF π b K π a K < N K N K 1 , and therefore b a is the best reply behavior in this case if c < N K 1 N 1 π b K . When the opposite happens, and so π b K π a K > N K N K 1 , b a is the best reply behavior if c < N K N 1 π a K . These conditions take the form of the ones in Table 4.
Consider the remaining four PS; they are characterized by the following fact: B R ( n K K ) = B R ( n K K ) but B R ( n K K ) B R ( n K K ) . In this case, it must be that τ i = 0 is optimal for i K while τ j = 1 is optimal for j K . Thanks to the symmetry in payoff matrices, we can say that the argument to prove the results for these four states is similar to the one for ( N A , 0 , 0 , N B ) and ( 0 , N A , N B , 0 ) . All the conditions are listed in Table 4.
Type Monomorphic State.
( N A , N A , 0 , 0 ) is a strict equilibrium if a is the BR i A and b, j B . Consider a player i A , who is given the revision opportunity at period t:
U a i ( a , ω t ) = N A 1 N 1 Π A ,
U a i ( b , ω t ) = N B N 1 π A ,
U a i ( τ 1 , ω t ) = N A 1 N 1 Π A + N B N 1 π A c .
Given that U a i ( a , ω t ) > U a i ( b , ω t ) , a is the best reply behavior IFF c > N B N 1 π A . Consider player j B :
U b j ( a , ω t ) = N A N 1 π B ,
U b j ( b , ω t ) = N B 1 N 1 Π B ,
U b j ( τ 1 , ω t ) = N A N 1 π B + N B 1 N 1 Π B c .
In this case, when π B Π B > N B 1 N A , b is never the best reply and a is the best reply; hence, the state can not be a strict equilibrium. When π B Π B < N B 1 N A , U b j ( b , ω t ) > U b j ( a , ω t ) , and U b j ( b , ω t ) > U b j ( τ 1 , ω t ) IFF c > N A N 1 π B .
No other state is a strict equilibrium.
For what concerns states where not all players of a group are playing the same action with the same group, this is easy to prove. Indeed, by definition, in these states, either not all players are playing their best reply action, or players are indifferent between two or more behaviors. In the first case, the state is not a strict equilibrium by definition; in the second case, there is no strictness of the equilibrium since there is not one best reply, but more behaviors can be best reply simultaneously. Hence, such states can not be strict equilibria. We are left with the seven states where every player of one group is doing the same thing against the same group. Such states are: 0 , 0 , N B , N B , 0 , N A , 0 , N B , N A , 0 , N B , 0 , 0 , 0 , N B , 0 , N A , N A , 0 , N B , 0 , N A , 0 , 0 , and N A , 0 , N B , N B . It is easy to prove that these states enter in the category of states where not every player is playing her/his best reply. Therefore, they can not be strict equilibria.
Strict equilibria are always absorbing states.
We first prove the sufficient and necessary conditions to be a fixed point, and second that every fixed point is an absorbing state. To prove the sufficient part, we rely on the definition of strict equilibrium. In a strict equilibrium, every player is playing her/his BR, and no one has the incentive to deviate. Whoever is given the revision opportunity does not change her/his behavior. Therefore, F ( ω t , θ t + 1 ) = ω t . To prove the necessary condition, think about being in a state that is not a strict equilibrium; in this case, by definition, we know that not all the players are playing their BR. Among them, there are states in which there are no indifferent players. In this case, with positive probability, one or more players who are not playing their BR are given the revision opportunity and they change action; therefore, F ( ω t , θ t + 1 ) ω t for some realization of θ t + 1 . In states where some players are indifferent between two or more behaviors, thanks to the tie rule, there is always a positive probability that the indifferent player changes her/his action since s/he is randomizing her/his choice. Moreover, there is also a positive probability to select a player indifferent between two or more behaviors. In this case, s/he changes the one that is currently playing with a positive probability too. Knowing this, we are sure that no state outside strict equilibria can be a fixed point. In our case, a fixed point is also an absorbing state by definition. Indeed, every fixed point absorbs at least one state: the one where all players except one are playing the same behavior. In this case, if that player is given the revision opportunity, s/he changes for sure her/his behavior into the one played by every player. □
Here, we prove the results of the stochastic stability analysis of Section 4.
Proof of Theorem 2.
We split the absorbing states into two sets and then apply Theorem 1 by Ellison [11]. Define the following two sets of states: M 1 = P S a , P S b and M 2 = ( P S \ M 1 ) M S . Similarly, define M 1 = P S b and M 2 = M S ( P S \ M 1 ) .
Analysis with M 1 and M 2 .
R ( M 1 ) is the minimum number of mistakes to escape the basins of attraction of both P S a and P S b . The dimension of these basins of attraction is determined by the value of c. In a state inside D ( P S a ) , b a is BR for B, and a is BR for A. Similarly, a b is optimal for A inside D ( P S b ) and b is optimal for B. The minimum resistance paths that start in P S a , and P S b and exit from their basins of attraction involve ε mistakes.
We calculate the dimension of these basins of attraction for 0 < c < 1 N 1 min { π A , π B } . We start from P S a and the argument stands for the other states in P S for symmetry of payoffs matrix.
Firstly, we consider the minimum number of mistakes that makes a BR for B players. Consider the choice of a B player inside a category of states where n B B 0 , N B Π B Π B Π B + π B and n A B N A Π B Π B + π B , N A . Referring to Equations (5)–(8), the optimal level of c s.t. 1 is the best reply, and for B players, it is
c < min N B Π B n B B ( Π B + π B ) Π B N 1 , n A B ( Π B + π B ) N A Π B N 1 .
If 0 < c < 1 N 1 min { π A , π B } , whenever n B B 0 , N B Π B Π B Π B + π B and n A B N A Π B Π B + π B , N A , 1 is the BR for B. Therefore, a path towards a state where n B B N B Π B + π B Π B + π B is a transition out of the basin of attraction of P S a . Starting from n B B = 0 , the cost of this transition is N B Π B + π B Π B + π B . This cost is determined by ε mistakes, since once in P S a , it is sufficient that a number of B plays by mistake b. Another possible path is to make b a BR for A. The cost of this transition is N A Π A + π A Π A + π A . With similar arguments, it is possible to show that the cost of exit from M 1 starting from P S b is the same. For this reason, R ( M 1 ) = min N B Π B + π B Π B + π B , N A Π A + π A Π A + π A .
We can show that the minimum resistance path to exit from the basin of attraction of M 2 reaches either P S a from M S a or P S b from M S b . Therefore, R ( M 2 ) = min N A π A + Π A Π A + π A , N B π B + Π B Π B + π B . R ( M 1 ) > R ( M 2 ) for every value of payoffs and group size: the stochastically stable state must be in M 1 .
Analysis with M 1 and M 2
Let us consider the path that goes from M 1 to P S a . Starting in P S b , it is sufficient that N B π A Π A + π A players from A play a for a transition from P S b to D ( P S a ) to happen.
Since N B π A Π A + π A < min N B Π B + π B Π B + π B , N A Π A + π A Π A + π A , we can say that R ( M 1 ) = N B π A Π A + π A . With a similar argument, it can be shown that R ( M 2 ) = N A π B Π B + π B . When R ( M 2 ) > R ( M 1 ) , P S a is uniquely stochastically stable. When R ( M 1 ) > R ( M 2 ) , P S b is uniquely stochastically stable.
R ( M 2 ) R ( M 1 ) when N B N A π B π A . □
Proof of Theorem 3.
In this case, R ( M S a ) = C R ( M S b ) and R ( M S b ) = C R ( M S a ) . Therefore, we just need to calculate the two Radius.
Radius of each state.
Let us consider R ( M S a ) . Since the basin of attraction of M S a is a region where a is the best reply behavior for both groups, many players should make a mistake such that b becomes BR for one of the two groups. For b to be BR for B players, it must be that n A B + n B B N Π B Π B Π B + π B . This state can be reached with ε mistakes at cost N π B + Π B Π B + π B . In a state where n A A + n B A N π A π A Π A + π A , b is BR for A, this path happens at cost N Π A + π A Π A + π A . In principle, N Π A + π A Π A + π A > N π B + Π B Π B + π B , hence, R ( M S a ) should be N π B + Π B Π B + π B . However, it may not be sufficient to reach such a state.
Consider reaching a state s.t. n A B + n B B = N Π B Π B Π B + π B , since N Π B Π B Π B + π B > N π A π A Π A + π A , it must be that a is still the best reply i A , and therefore, there is a path of zero resistance to M S a . Nevertheless, once in that state, it can happen that only B players are given the revision opportunity, and that they all choose behavior b. This creates a path of zero resistance to a state ( n ¯ A A , n ¯ A B , 0 , 0 ) . Once in this state, if n ¯ A A < N π A π A Π A + π A , the state is in the basin of attraction of M S b . This happens only if N π A π A Π A + π A + N B = N Π B Π B Π B + π B . More generally, considering k 0 , this happens if N π A π A Π A + π A + N B = N Π B Π B Π B + π B k . Fixing payoffs and groups size, k = N Π B Π B Π B + π B N π A π A Π A + π A N B ; hence, the cost of this path would be
N π B + Π B Π B + π B + N Π B Π B Π B + π B N π A π A Π A + π A N B = N A N π A π A Π A + π A .
With a similar reasoning, R ( M S b ) = N π A + Π A Π A + π A .
We prove that all the other paths with η mistakes are costlier than ones with ε . We know that a is the BR for every state inside the basin of attraction of M S a , nobody in the basin of attraction of M S a optimally buys the information, and every player who once bought the information (by mistake) plays behavior a a . Every path with an η mistake also involves an ε mistake, and hence is double that of the one described above.
Conditions for stochastically stable states.
M S a is stochastically stable IFF N A N π A π A Π A + π A > N π A + Π A Π A + π A , this is verified when N A > 2 N π A + Π A π A Π A + π A . Therefore, we conclude that M S a is stochastically stable in the above scenario, while if the opposite happens, M S b is stochastically stable. □
Proof of Theorem 4.
We first calculate radius, coradius, and modified coradius for the three states we are interested in, and then we compare them to draw inference about stochastic stability.
Radius of each state.
The Radius of M S a is the minimum number of mistakes that makes b BR for B players. This number is N π B + Π B Π B + π B . The alternative is to make b BR for A: hence, a path to state ( 0 , 0 , N B , N B ) , and then to ( 0 , 0 , 0 , 0 ) . The number of ε mistakes for this path is N Π A + π A Π A + π A . Therefore, R ( M S a ) = N π B + Π B Π B + π B . With a similar reasoning, we can conclude that R ( M S b ) = N π A + Π A Π A + π A .
Consider T S : the minimum-resistance path to exit from its basin of attraction reaches either M S a or M S b , depending on payoffs. In other words, the minimum number of mistakes to exit from D ( T S ) is the one that makes either a or b as BR. Consider the path from T S to M S a : in this case, some mistakes are needed to make a BR for B. The state in which a is BR for B depends on payoffs and group size. In a state ( N A , N A , k , k ) , a is BR for every player in B if ( N A + k 1 ) π B > ( N N A k ) Π B . This inequality is obtained by declining Equations (5)–(8), comparing B playing a / a b or b / b a . Fixing payoffs, we can calculate the exact value of k that is N B Π B N A π B + π B Π B + π B ; this would be the cost of the minimum mistake transition from T S to M S a . With a similar argument, the cost of the minimum mistake transition from T S to M S b is N A Π A N B π A + π A Π A + π A .
There are no paths involving η mistakes that are lower than the two proposed above. The intuition is the following. Consider a situation in which m players of A are given the revision opportunity at one period, and they all choose to buy the information. In this case, they all optimally choose behavior a b . This means that at the cost of n, there is a path to a state in which N A m players are playing b against B, in this state, b is still the BR for group B, while a is still the BR for A. Hence, from that state, there is a path of zero resistance to T S . The same happens when B players choose by mistake to buy the information. Therefore, R ( T S ) = min N B Π B N A π B + π B Π B + π B , N A Π A N B π A + π A Π A + π A .
Coradius of each state.
We start from T S : in this case, we have to consider the two minimum-resistance paths to reach it from M S a and M S b . Therefore, N π A + Π A Π A + π A and N π B + Π B Π B + π B . Firstly, the argument to prove that these two are the minimum resistance paths to reach T S from M S a and M S b are given by the previous part of the proof. Secondly, we have to prove that this path is the maximum among the minimum resistance paths starting from any other state and ending in T S . There are three regions from which we can start and end up in T S : the basin of attraction of M S b , the one of M S a , and all the other states that are not in the basins of attraction of the three states considered. We can say that from this region, there is always a positive probability of ending up in M S a , M S b , or T S . Hence, we can consider as 0 the cost to reach T S from this region. The other two regions are the one considered above, and since we are taking the maximum path to reach T S from any other state, we have to take the sum of this two. Hence, C R ( T S ) = N π A + Π A Π A + π A + N π B + Π B Π B + π B .
Let us think about M S . Similarly to the two previous proofs, we can focus only on ε paths. Note that in this case, T S is always placed between the two M S . Let us start from M S b : in this case we can consider three different paths starting from any state and arriving to M S b . The first one starts in T S , the second starts in every state outside the basin of attractions of the three absorbing states, and the last starts in M S a . In the second case, there is at least one transition of zero resistance to M S b . Next, assume starting in T S : the minimum number of mistakes to reach M S b from T S is the one that makes b BR for A players. Therefore, N A Π A N B π A + π A Π A + π A .
Now, we need to consider the case of starting in M S a . Firstly, consider the minimum number of mistakes to make b BR for A players. This number is N Π A + π A Π A + π A . Secondly, consider the minimum number of mistakes to make b BR for B players, and then once T S is reached, the minimum number of mistakes that makes b BR for A players.
min r ( M S a , M S b ) = min N Π A + π A Π A + π A , N π B + Π B Π B + π B + N A Π A N B π A + π A Π A + π A .
Since the two numbers in the expression are all greater than N A Π A N B π A + π A Π A + π A , we can say that C R ( M S b ) = min r ( M S a , M S b ) .
Reaching a state where b is BR for group A from T S is for sure less costly than reaching it from M S a , since in T S there are more people playing b. Therefore, N Π A + π A Π A + π A N π B + Π B Π B + π B + N A Π A N B π A + π A Π A + π A , hence, C R ( M S b ) = N π B + Π B Π B + π B + N A Π A N B π A + π A Π A + π A . With a similar reasoning, C R ( M S a ) = N π A + Π A Π A + π A + N B Π B N A π B + π B Π B + π B .
Modified Coradius of each state.
Firstly, note that C R ( T S ) = C R * ( T S ) , since between M S and T S , there are no intermediate states. Formally,
C R * ( T S ) = min r * ( M S a , , T S ) + min r * ( M S b , , T S ) = N π A + Π A Π A + π A + N π B + Π B Π B + π B .
The maximum path of minimum resistance from each M S to the other M S passes through T S . Hence, for each M S , we need to subtract from the coradius the cost of passing from T S to the other M S . Let us consider C R * ( M S a ) ; we need to subtract from the coradius the cost of passing from T S to M S b : this follows from the definition of modified coradius. Hence,
C R * ( M S a ) = N π A + Π A Π A + π A + N B Π B N A π B + π B Π B + π B N A Π A N B π A + π A Π A + π A .
Similarly,
C R * ( M S b ) = N π B + Π B Π B + π B + N A Π A N B π A + π A Π A + π A N B Π B N A π B + π B Π B + π B .
Note that C R * ( M S a ) < C R ( M S a ) and C R * ( M S b ) < C R ( M S b ) .
Conditions for stochastically stable states.
By comparing all the possibilities, it is possible to verify that if R ( M S a ) > C R ( M S a ) , both R ( M S b ) < C R ( M S b ) and R ( T S ) < C R ( T S ) . Similar for R ( M S b ) > C R ( M S b ) or R ( T S ) > C R ( T S ) . When R ( M S a ) C R ( M S a ) , R ( M S b ) C R ( M S b ) , and R ( T S ) C R ( T S ) , we need to use Modified Coradius. Given that C R ( T S ) = C R * ( T S ) it will never be that R ( T S ) > C R * ( T S ) . We can show that when R ( M S a ) > C R * ( M S a ) , then R ( M S b ) < C R * ( M S b ) and vice-versa.
When R ( M S a ) = C R * ( M S a ) , it is also possible that R ( M S b ) = C R * ( M S b ) . Thanks to Theorem 3 in Ellison [11], we know that either both states are stochastically stable, or neither of the two is. Note that for the ergodicity of our process, the second case is impossible; hence, it must be that when both R ( M S a ) = C R * ( M S a ) and R ( M S b ) = C R * ( M S b ) , both M S b and M S a are stochastically stable. □
Proof of Lemma 6.
Recall from Section 3 that ω R = P S , M S . Firstly, if 1 N 1 max π A , π B < c < 1 N 1 min Ξ P S , T S is not an absorbing state (see Corollary 2), all P S are absorbing states (see Corollary 1), and M S are absorbing states (see Table 4). Secondly, consider the set M = P S a , P S b , M S a , M S b and the set ω R \ M containing all the P S not in M. If R ( M ) > R ( ω R \ M ) , then the stochastically stable state must be in M. Since the level of the cost is not fixed, the radius of these two sets depend on the cost level. Following the same logic as in Theorem 2 but computing the result as a function of c, we can calculate the two Radii:
R ( M ) = min N A π B + c ( N 1 ) Π A + π A , N B π A + c ( N 1 ) Π A + π A , N Π A + π A Π A + π A , N Π B + π B Π B + π B
R ( ω R \ M ) = min N B π B c ( N 1 ) + Π B Π B + π B , N B π A c ( N 1 ) Π A + π A , N A π B c ( N 1 ) Π B + π B .
By comparing all the twelve possibilities case by case, it is possible to show that for every value of payoffs, group size, and cost, R ( M ) > R ( ω R \ M ) . Therefore the stochastically stable state must be in the set M. □

References

  1. Kandori, M.; Mailath, G.J.; Rob, R. Learning, mutation, and long run equilibria in games. Econom. J. Econom. Soc. 1993, 61, 29–56. [Google Scholar] [CrossRef]
  2. Lewis, D. Convention: A Philosophical Study; Harvard University Press: Cambridge, MA, USA, 1969. [Google Scholar]
  3. Bicchieri, C. The Grammar of Society: The Nature and Dynamics of Social Norms; Cambridge University Press: Cambridge, UK, 2006. [Google Scholar]
  4. Schelling, T.C. The Strategy of Conflict; Harvard University Press: Cambridge, MA, USA, 1960. [Google Scholar]
  5. Young, H.P. The economics of convention. J. Econ. Perspect. 1996, 10, 105–122. [Google Scholar] [CrossRef] [Green Version]
  6. Young, H.P. Individual Strategy and Social Structure: An Evolutionary Theory of Institutions; Princeton University Press: Princeton, NJ, USA, 1998. [Google Scholar]
  7. Neary, P.R. Competing conventions. Games Econ. Behav. 2012, 76, 301–328. [Google Scholar] [CrossRef]
  8. Foster, D.; Young, H.P. Stochastic evolutionary game dynamics *. Theor. Popul. Biol. 1990, 38, 219–232. [Google Scholar] [CrossRef] [Green Version]
  9. Young, H.P. The evolution of conventions. Econom. J. Econom. Soc. 1993, 61, 57–84. [Google Scholar] [CrossRef]
  10. Freidlin, M.I.; Wentzell, A.D. Random Perturbations of Dynamical Systems, 2nd ed.; Springer: Berlin, Germany, 1984. [Google Scholar]
  11. Ellison, G. Basins of attraction, long-run stochastic stability, and the speed of step-by-step evolution. Rev. Econ. Stud. 2000, 67, 17–45. [Google Scholar] [CrossRef] [Green Version]
  12. Jackson, M.O.; Watts, A. On the formation of interaction networks in social coordination games. Games Econ. Behav. 2002, 41, 265–291. [Google Scholar] [CrossRef]
  13. Bhaskar, V.; Vega-Redondo, F. Migration and the evolution of conventions. J. Econ. Behav. Organ. 2004, 55, 397–418. [Google Scholar] [CrossRef] [Green Version]
  14. Smith, J.M.; Price, G.R. The logic of animal conflict. Nature 1973, 246, 15–18. [Google Scholar] [CrossRef]
  15. Friedman, D. On economic applications of evolutionary game theory. J. Evol. Econ. 1998, 8, 15–43. [Google Scholar] [CrossRef] [Green Version]
  16. Cressman, R.; Garay, J.; Hofbauer, J. Evolutionary Stability Concepts for N–species Frequency-dependent Interactions. J. Theor. Biol. 2001, 211, 1–10. [Google Scholar] [CrossRef]
  17. Cressman, R.; Ansell, C.; Binmore, K. Evolutionary Dynamics and Extensive Form Games; MIT Press: Cambridge, MA, USA, 2003; Volume 5. [Google Scholar]
  18. Quilter, T. Noise matters in heterogeneous populations. In ESE Discussion Papers 169; University of Edinburgh, Edinburgh School of Economics: Edinburgh, UK, 2007. [Google Scholar]
  19. Neary, P.R.; Newton, J. Heterogeneity in Preferences and Behavior in Threshold Models. J. Mech. Inst. Des. 2017, 2, 1. [Google Scholar]
  20. Carvalho, J.P. Coordination and culture. Econ. Theory 2017, 64, 449–475. [Google Scholar] [CrossRef] [Green Version]
  21. Michaeli, M.; Spiro, D. From peer pressure to biased norms. Am. Econ. J. Microecon. 2017, 9, 152–216. [Google Scholar] [CrossRef]
  22. Tanaka, C.M.; Lee, J.H.; Iwasa, Y. The persistence of a local dialect when a national standard language is present: An evolutionary dynamics model of cultural diversity. Bull. Math. Biol. 2018, 80, 2761–2786. [Google Scholar] [CrossRef]
  23. Naidu, S.; Hwang, S.H.; Bowles, S. The Evolution of Egalitarian Sociolinguistic Conventions. Am. Econ. Rev. 2017, 107, 572–577. [Google Scholar] [CrossRef]
  24. Belloc, M.; Bowles, S. The persistence of inferior cultural-institutional conventions. Am. Econ. Rev. 2013, 103, 93–98. [Google Scholar] [CrossRef] [Green Version]
  25. Goyal, S.; Hernández, P.; Martínez-Cánovasz, G.; Moisan, F.; Munoz-Herrera, M.; Sánchez, A. Integration and segregation. Exp. Econ. 2020, forthcoming. [Google Scholar]
  26. Van Gerwen, N.; Buskens, V. Partner-specific behavior in social networks: Coordination among actors with heterogeneous preferences. Soc. Netw. 2018, 52, 68–80. [Google Scholar] [CrossRef]
  27. He, S. Minority advantage and disadvantage in competition and coordination. J. Econ. Behav. Organ. 2019, 163, 464–482. [Google Scholar] [CrossRef]
  28. He, S.; Wu, J. Compromise and coordination: An experimental study. Games Econ. Behav. 2020, 119, 216–233. [Google Scholar] [CrossRef]
  29. Goyal, S.; Janssen, M.C. Non-exclusive conventions and social coordination. J. Econ. Theory 1997, 77, 34–57. [Google Scholar] [CrossRef] [Green Version]
  30. Galesloot, B.M.; Goyal, S. Costs of flexibility and equilibrium selection. J. Math. Econ. 1997, 28, 249–264. [Google Scholar] [CrossRef]
  31. Bergin, J.; Lipman, B.L. Evolution with state-dependent mutations. Econom. J. Econom. Soc. 1996, 64, 943–956. [Google Scholar] [CrossRef] [Green Version]
  32. Blume, L.E. How noise matters. Games Econ. Behav. 2003, 44, 251–271. [Google Scholar] [CrossRef]
  33. Sandholm, W.H. Population Games and Evolutionary Dynamics; MIT Press: Cambridge, MA, USA, 2010. [Google Scholar]
  34. Dokumacı, E.; Sandholm, W.H. Large deviations and multinomial probit choice. J. Econ. Theory 2011, 146, 2151–2158. [Google Scholar] [CrossRef]
  35. Klaus, B.; Newton, J. Stochastic stability in assignment problems. J. Math. Econ. 2016, 62, 62–74. [Google Scholar] [CrossRef] [Green Version]
  36. Blume, L.E. The statistical mechanics of strategic interaction. Games Econ. Behav. 1993, 5, 387–424. [Google Scholar] [CrossRef] [Green Version]
  37. Myatt, D.P.; Wallace, C. A multinomial probit model of stochastic evolution. J. Econ. Theory 2003, 113, 286–301. [Google Scholar] [CrossRef] [Green Version]
  38. Naidu, S.; Hwang, S.H.; Bowles, S. Evolutionary bargaining with intentional idiosyncratic play. Econ. Lett. 2010, 109, 31–33. [Google Scholar] [CrossRef] [Green Version]
  39. Hwang, S.H.; Naidu, S.; Bowles, S. Social conflict and the evolution of unequal conventions. In Santa Fe Institute Working Paper; Santa Fe Institute: Santa Fe, NM, USA, 2018. [Google Scholar]
  40. Bilancini, E.; Boncinelli, L. The evolution of conventions under condition-dependent mistakes. Econ. Theory 2020, 69, 497–521. [Google Scholar] [CrossRef] [Green Version]
  41. Lim, W.; Neary, P.R. An experimental investigation of stochastic adjustment dynamics. Games Econ. Behav. 2016, 100, 208–219. [Google Scholar] [CrossRef]
  42. Hwang, S.H.; Lim, W.; Neary, P.; Newton, J. Conventional contracts, intentional behavior and logit choice: Equality without symmetry. Games Econ. Behav. 2018, 110, 273–294. [Google Scholar] [CrossRef] [Green Version]
  43. Mäs, M.; Nax, H.H. A behavioral study of “noise” in coordination games. J. Econ. Theory 2016, 162, 195–208. [Google Scholar] [CrossRef] [Green Version]
  44. Bilancini, E.; Boncinelli, L.; Nax, H.H. What noise matters? Experimental evidence for stochastic deviations in social norms. J. Behav. Exp. Econ. 2020, 90, 101626. [Google Scholar] [CrossRef]
  45. Newton, J. Evolutionary game theory: A renaissance. Games 2018, 9, 31. [Google Scholar] [CrossRef] [Green Version]
  46. Newton, J. Conventions under Heterogeneous Behavioural Rules. Rev. Econ. Stud. 2020. [Google Scholar] [CrossRef]
  47. Sawa, R.; Wu, J. Prospect dynamics and loss dominance. Games Econ. Behav. 2018, 112, 98–124. [Google Scholar] [CrossRef]
  48. Sawa, R.; Wu, J. Reference-dependent preferences, super-dominance and stochastic stability. J. Math. Econ. 2018, 78, 96–104. [Google Scholar] [CrossRef]
  49. Staudigl, M. Stochastic stability in asymmetric binary choice coordination games. Games Econ. Behav. 2012, 75, 372–401. [Google Scholar] [CrossRef]
  50. Simon, H.A. A behavioral model of rational choice. Q. J. Econ. 1955, 69, 99–118. [Google Scholar] [CrossRef]
  51. Grossman, S.J.; Stiglitz, J.E. On the impossibility of informationally efficient markets. Am. Econ. Rev. 1980, 70, 393–408. [Google Scholar]
  52. Dewatripont, M.; Tirole, J. Modes of communication. J. Political Econ. 2005, 113, 1217–1238. [Google Scholar] [CrossRef] [Green Version]
  53. Dewatripont, M. Costly communication and incentives. J. Eur. Econ. Assoc. 2006, 4, 253–268. [Google Scholar] [CrossRef] [Green Version]
  54. Caillaud, B.; Tirole, J. Consensus building: How to persuade a group. Am. Econ. Rev. 2007, 97, 1877–1900. [Google Scholar] [CrossRef] [Green Version]
  55. Tirole, J. Cognition and incomplete contracts. Am. Econ. Rev. 2009, 99, 265–294. [Google Scholar] [CrossRef] [Green Version]
  56. Butler, J.V.; Guiso, L.; Jappelli, T. Manipulating reliance on intuition reduces risk and ambiguity aversion. In CEPR Discussion Paper No. DP9461; CEPR Centre for Economic Policy Research: London, UK, 2013. [Google Scholar]
  57. Bilancini, E.; Boncinelli, L. When market unraveling fails and mandatory disclosure backfires: Persuasion games with labeling and costly information acquisition. J. Econ. Manag. Strategy 2021. [Google Scholar] [CrossRef]
  58. Bilancini, E.; Boncinelli, L. Rational attitude change by reference cues when information elaboration requires effort. J. Econ. Psychol. 2018, 65, 90–107. [Google Scholar] [CrossRef] [Green Version]
  59. Bilancini, E.; Boncinelli, L. Signaling to analogical reasoners who can acquire costly information. Games Econ. Behav. 2018, 110, 50–57. [Google Scholar] [CrossRef] [Green Version]
  60. Jehiel, P. Analogy-based expectation equilibrium. J. Econ. Theory 2005, 123, 81–104. [Google Scholar] [CrossRef]
  61. Mackowiak, B.; Matejka, F.; Wiederholt, M. Rational Inattention: A Review. In CEPR Discussion Papers; CEPR Centre for Economic Policy Research: London, UK, 2020. [Google Scholar]
  62. Güth, W.; Kliemt, H. Competition or co-operation: On the evolutionary economics of trust, exploitation and moral attitudes. Metroeconomica 1994, 45, 155–187. [Google Scholar] [CrossRef]
  63. Berger, U.; De Silva, H. Evolution of deterrence with costly reputation information. PLoS ONE 2021, 16, e0253344. [Google Scholar] [CrossRef] [PubMed]
  64. Staudigl, M.; Weidenholzer, S. Constrained interactions and social coordination. J. Econ. Theory 2014, 152, 41–63. [Google Scholar] [CrossRef] [Green Version]
  65. Bilancini, E.; Boncinelli, L. Social coordination with locally observable types. Econ. Theory 2018, 65, 975–1009. [Google Scholar] [CrossRef] [Green Version]
  66. Bilancini, E.; Boncinelli, L.; Wu, J. The interplay of cultural intolerance and action-assortativity for the emergence of cooperation and homophily. Eur. Econ. Rev. 2018, 102, 1–18. [Google Scholar] [CrossRef] [Green Version]
Figure 1. P S b = ( 0 , 0 ) is uniquely stochastically stable: π B π A < N B N A .
Figure 1. P S b = ( 0 , 0 ) is uniquely stochastically stable: π B π A < N B N A .
Games 12 00053 g001
Figure 2. P S a = ( 10 , 5 ) is uniquely stochastically stable: π B π A > N B N A .
Figure 2. P S a = ( 10 , 5 ) is uniquely stochastically stable: π B π A > N B N A .
Games 12 00053 g002
Table 1. Payoff matrix of interactions inside group A.
Table 1. Payoff matrix of interactions inside group A.
ab
a Π A , Π A 0 , 0
b 0 , 0 π A , π A
Table 2. Payoff matrix of interactions inside group B.
Table 2. Payoff matrix of interactions inside group B.
ab
a π B , π B 0 , 0
b 0 , 0 Π B , Π B
Table 3. Payoff matrix of inter-group interactions.
Table 3. Payoff matrix of inter-group interactions.
ab
a Π A , π B 0 , 0
b 0 , 0 π A , Π B
Table 4. Necessary and sufficient conditions for absorbing states.
Table 4. Necessary and sufficient conditions for absorbing states.
StateCondition on Group Size and PayoffsConditions on c
M S a nonenone
M S b nonenone
T S π B Π B < N B 1 N A c > max N B N 1 π A , N A N 1 π B
P S b none c < N B N 1 π A
P S a (1) π B Π B > N B 1 N A
(2) π B Π B < N B 1 N A
(1) c < N B 1 N 1 Π B
(2) c < N A N 1 π B
( 0 , N A , N B , N B ) (1) π A Π A < N B N A 1
(2) π A Π A > N B N A 1
(1) c < N A 1 N 1 π A
(2) c < N B N 1 Π A
( N A , 0 , 0 , N B ) none c < min N B N 1 π A , N B 1 N 1 π B
( 0 , N A , N B , 0 ) (1) π A Π A < N B N A 1 and π B Π B > N B 1 N A
(2) π A Π A > N B N A 1 and π B Π B > N B 1 N A
(3) π A Π A < N B N A 1 and π B Π B < N B 1 N A
(4) π A Π A > N B N A 1 and π B Π B < N B 1 N A
(1) c < min N A 1 N 1 π A , N B 1 N 1 Π B
(2) c < min N B N 1 Π A , N B 1 N 1 Π B
(3) c < min N A 1 N 1 π A , N A N 1 π B
(4) c < min N B N 1 Π A , N A N 1 π B
( 0 , 0 , 0 , N B ) none c < N B 1 N 1 π B
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rozzi, R. Competing Conventions with Costly Information Acquisition. Games 2021, 12, 53. https://doi.org/10.3390/g12030053

AMA Style

Rozzi R. Competing Conventions with Costly Information Acquisition. Games. 2021; 12(3):53. https://doi.org/10.3390/g12030053

Chicago/Turabian Style

Rozzi, Roberto. 2021. "Competing Conventions with Costly Information Acquisition" Games 12, no. 3: 53. https://doi.org/10.3390/g12030053

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop