Academia.eduAcademia.edu

Continuum time limit and stationary states of the minority game

2001, Physical Review E

We discuss in detail the derivation of stochastic differential equations for the continuum time limit of the Minority Game. We show that all properties of the Minority Game can be understood by a careful theoretical analysis of such equations. In particular, i) we confirm that the stationary state properties are given by the ground state configurations of a disordered (soft) spin system; ii) we derive the full stationary state distribution; iii) we characterize the dependence on initial conditions in the symmetric phase and iv) we clarify the behavior of the system as a function of the learning rate. This leaves us with a complete and coherent picture of the collective behavior of the Minority Game. Strikingly we find that the temperature like parameter which is introduced in the choice behavior of individual agents turns out to play the role, at the collective level, of the inverse of a thermodynamic temperature.

Continuum time limit and stationary states of the Minority Game Matteo Marsili arXiv:cond-mat/0102257v3 [cond-mat.stat-mech] 3 Jul 2001 Istituto Nazionale per la Fisica della Materia (INFM), Trieste-SISSA Unit, Via Beirut 2-4, Trieste 34014, Italy Damien Challet Theoretical Physics, Oxford University, 1 Keble Road, Oxford OX1 3NP, United Kingdom (February 1, 2008) We discuss in detail the derivation of stochastic differential equations for the continuum time limit of the Minority Game. We show that all properties of the Minority Game can be understood by a careful theoretical analysis of such equations. In particular, i) we confirm that the stationary state properties are given by the ground state configurations of a disordered (soft) spin system; ii) we derive the full stationary state distribution; iii) we characterize the dependence on initial conditions in the symmetric phase and iv) we clarify the behavior of the system as a function of the learning rate. This leaves us with a complete and coherent picture of the collective behavior of the Minority Game. Strikingly we find that the temperature like parameter which is introduced in the choice behavior of individual agents turns out to play the role, at the collective level, of the inverse of a thermodynamic temperature. I. INTRODUCTION acterize the average behavior of agents by computing the frequency with which they play their strategies. This step can be translated in the study of the ground state properties of a soft spin disordered Hamiltonian. Secondly we characterize the fluctuations around the average behavior. To do this, we explicitly solve the Fokker-Planck equation associated to the stochastic dynamics. The new results which we derive are: Even under the most demanding definition, the Minority Game (MG) [1,2] definitely qualifies as a complex system. The MG can be regarded as an Ising model for systems of heterogeneous adaptive agents, who interact via a global mechanism that entails competition for limited resource, as found for instance in biology and financial markets. In spite of more than three years of intense research, its rich dynamical behavior is still the subject of investigations, many variations of the basic MG being proposed, each uncovering new surprising regions of phase space. Most importantly, Refs. [3–7] have shown that much theoretical insight can be gained on the behavior of this class of models, using non-equilibrium statistical physics and statistical mechanics of disordered systems. The approach of Refs. [3–6] rests on the assumption that, in a continuum time limit (CTL), the dynamics of the MG can be described by a set of deterministic equations. From these, one derives a function H which is minimized along all trajectories; hence, the stationary state of the system corresponds to the ground state of H, which can be computed exactly by statistical mechanics techniques. This approach has been challenged in Refs. [8,9], which have proposed a stochastic dynamics for the MG, thus leading to some debate in the literature [10,11]. In this paper, our aim is to analyze in detail the derivation of the CTL in order to clarify this issue. We show that a proper derivation of the CTL indeed reconciles the two approaches: the resulting dynamical equations – Eqs. (15-17) below, which are our central result – are indeed stochastic, as suggested in Ref. [8,9], but still the stationary state of the dynamics is described by the minima of the function H, as suggested in Refs. [3,4]. We then confirm the analytic results derived previously. In few words, our analysis follows two main steps: first we char- 1. we derive the full probability distribution in the stationary state. Remarkably we find that the parameter which is introduced as a temperature in the individual choice model, turns out to play the role of the inverse of a global temperature; 2. for α > αc the distribution factorizes over the agents whereas in the symmetric phase (α < αc ) agents play in a correlated way. In the latter case, the correlations contribute to the stochastic force acting on agents. We show how the dependence of global efficiency on individual temperature found in Ref. [8] arises as a consequence of these correlations; 3. we extend the analytic approach of Refs. [3,4] to the α < αc phase and asymmetric initial conditions. The dependence on the initial conditions in this phase, first noticed and discussed in Refs. [3,4], has been more recently studied quantitatively in Refs. [9,7]. We clarify the origin of this behavior and derive analytic solutions in the limit Γ → 0. 4. we show that the stronger is the initial asymmetry in agents evaluation of their strategies, the larger is the efficiency and the more stable is the system against crowd effects [12]. 5. we derive the Hamiltonian of MGs with non-linear payoffs. 1 degrees of freedom. As a consequence also Us,i (t) and hence the probability with which agents chose si (t) are subject to stochastic fluctuations. Our analysis will indeed focus on the characterization of the low-frequency fluctuations of Us,i , by integrating our the high frequency fluctuations of µ(t) and si (t). This will become clearer in the next section. For the time being let it suffice to say that there are two level of fluctuations, that of “fast” variables µ(t) and si (t) and that of “slow” degrees of freedom Us,i (t). The key parameter is the ratio α = P/N [18] and two relevant quantities are This leaves us with a coherent picture of the collective behavior of the Minority Game which is an important reference framework for the study of complex systems of heterogeneous adaptive agents. II. THE MODEL The dynamics of the MG is defined in terms of dynamical variables Us,i (t) in discrete time t = 0, 1, . . . . These are scores, propensities or “attractions” [14] which each agent i = 1, . . . , N attaches to each of his possible choices s = 1, . . . , S. Each agent takes a decision si (t) with eΓi Us,i (t) Prob{si (t) = s} = P Γ U ′ (t) i s ,i s′ e σ 2 = hA2 i, (1) N X µ(t) i=1 Our approach, which follows that of Refs. [3,4], is based on two key observations: each agent updates his scores according to µ(t) A(t) Us,i (t + 1) = Us,i (t) − as,i P . (4) III. THE CONTINUUM TIME LIMIT (2) asi (t),i P 1 X hA|µi2 P µ=1 which measure, respectively, global efficiency and predictability† . Generalizations of the model, where agents account for their market impact [3,4], where deterministic agents – so-called producers – are present [5], or where agents are allowed not to play [20–24], have been proposed. Rather than dealing with the most generic model which would depend on too many parameters, we shall limit our discussion to the plain MG. Furthermore we shall specialize, in the second part of the paper to the case S = 2 which lends itself to a simpler analytic treatment. The analysis carries through in obvious ways to the more general cases discussed in Refs. [4,5,3,24]. where Γi > 0 appears as an “individual inverse temperature”. The original MG corresponds to Γi = ∞ [1], and was generalized later to Γi ≡ Γ < ∞ [8]. The public information variable µ(t) is given to all agents; it belongs to the set of integers (1, . . . , P ), and can either be the binary encoding of last M winning choices [1], or drawn at random from a uniform distribution [15]; we stick to the latter case for sake of simplicity∗ . The acµ(t) tion asi (t),i of each agent depends the on its choice si (t) and on µ(t). The coefficients aµs,i , called strategies, plays the role of quenched disorder: they are randomly drawn signs (Prob{aµs,i = ±1} = 1/2), independently for each i, s and µ. On the basis of the outcome A(t) = H= 2 1. the scaling σ√ ∼ N , at fixed α, suggests that typically A(t) ∼ N . Hence time increments of U√ s,i (t), √ in Eq. (3) are small (i.e. of order N /P ∼ 1/ N ); (3) The idea of this equation is that agents reward [Us,i (t + 1) > Us,i (t)] those strategy which would have predicted the minority sign −A(t)/|A(t)|. The MG was initially proposed with a nonlinear dependence on A(t), i.e. with µ(t) a dynamics Us,i (t + 1) = Us,i (t) − as,i sign A(t). This leads to qualitatively similar results. The extension of our theory to non-linear cases is dealt with in the appendix. We shall not discuss any longer the interpretation of the model, which is discussed at length elsewhere [4,17–19]. The source of randomness are in the choices of µ(t) by Nature and of si (t) by agents. These are fast fluctuating 2. characteristic times of the dynamics are proportional to P . Naively this is because agents need to “test” their strategies against all P values of µ, which requires of order P time steps. More precisely, one can reach this conclusion by measuring relaxation or correlation times and verifying that they indeed grow linearly with P (see Ref. [10]). The second observation implies that one needs to study the dynamics in the rescaled time τ = t/P . This makes † Averages h. . .i stand for time averages in the stationary state of the process. Then h. . . |µi stands for time averages conditional on µ(t) = µ. ∗ Both prescriptions lead to qualitatively similar results for the quantities we study here. See [16] for more details. 2 Now the first term is of order dτ as required for a deterministic term. In addition it remains finite as N → ∞ [25]. The second term is a sum of P dτ random variables Xs,i (t) with zero average. We take dτ fixed and N very large, so that P dτ ≫ 1 and we can use limit theorems. The variables Xs,i (t) are independent from time to time, because both µ(t) and sj (t) are drawn independently at each time. Hence Xs,i (t) for P τ ≤ t < P (τ + dτ ) are independent and identically‡ distributed. For P dτ ≫ 1 we may approximate the second term dWs,i of Eq. (10) by a Gaussian variable with zero average and variance our approach differ from that of Refs. [8,9], where the time is not rescaled. In order to study the dynamics for P, N ≫ 1 at fixed α, we shall focus on a fixed small increment dτ such that P dτ = αN dτ ≫ 1. This means that we take the continuum time limit dτ → 0 only after the thermodynamic limit N → ∞. We focus only on the leading order in N . Furthermore we shall also consider Γi finite and Γi dτ ≪ 1 (5) which means that the limit Γi → ∞ should be taken after the limit dτ → 0. The orders in which these limits are taken, given the agreement with numerical simulation results, does not really matters: as we shall see differences only enter in the finite size corrections. We shall come back later to these issues. Iteration of the dynamics for P dτ time steps, from t = P τ to t = P (τ + dτ ) gives hdWs,i (τ )dWr,j (τ ′ )i = δ(τ − τ ′ ) P2 P (τ +dτ ) X t=P τ = δ(τ − τ ′ )dτ hXs,i (t)Xr,j (t)iπ hXs,i (t)Xr,j (t)iπ P (6) where the δ(τ − τ ′ ) comes from independence of Xs,i (t) and Xr,j (t′ ) for t 6= t′ and the fact that Xs,i (t) are identically distributed in time leads to the expression in the second line. Now: where we have introduced the functions us,i (τ ) = Us,i (P τ ). Let us separate a deterministic (dus,i ) from a stochastic (dWs,i ) term in this equation by replacing hXs,i (t)Xr,j (t)iπ as,i ar,j hA2 iπ as,i hAiπ ar,j hAiπ = − P P P (11) us,i (τ + dτ ) − us,i (τ ) = − 1 P P (τ +dτ )−1 X µ(t) as,i A(t). t=P τ µ(t) as,i A(t) = as,i hAiπ + Xs,i (t). The second term always vanishes for N → ∞ because as,i hAiπ is of order N 0 [25]. In the first term, instead, (7) Here and henceforth, we denote averages over µ by an over-line hA2 |µiπ = N + P 1 X µ R= R , P µ=1 1 P dτ P (τ +dτ )−1 X t=P τ eΓi Us,i (t) P Γ U (t) , i r,i re (8) which is the frequency with which agent i plays strategy s in the time interval P τ ≤ t < P (τ + dτ ). Notice that πs,i (τ ) will themselves be stochastic variables, hence we also define the average on the stationary state as Z 1 τ0 +T dτ h. . .iπ(τ ) (9) h. . .i = lim τ0 ,T →∞ T τ 0 hA2 |µiπ ≈ hA2 i ≡ σ 2 . P (τ +dτ )−1 X (13) ‡ Strictly speaking, si (t) is drawn from the distribution in Eq. (1) and not from πs,i of Eq. (8). However these two distributions differ by a negligible amount as long as the condition (5) holds (see later). us,i (τ + dτ ) − us,i (τ ) = dus,i (τ ) + dWs,i (τ ) 1 P (12) This approximation can be justified naively by observing that the dependence on πs,i of the correlations only involves the global quantity hA2 |µiπ /P for which one expects some sort of self-averaging properties. Numerical results suggest that terms which are ignored by Eq. (13) are negligible for N ≫ 1, but we were unable to prove where the average inside the integral is performed with the probabilities πs,i (τ ). Hence: = −as,i hAiπ dτ + k6=l=1 aµs′ ,k aµr′ ,l πs′ ,k πr′ ,l s′ ,r ′ is of order N which then gives a positive contribution in Eq. (11) for N → ∞. Eq. (12) leads to a stochastic dynamics where the noise covariance depends on the stochastic variables us,i (τ ) themselves. The complications which result from this fact can be avoided if one takes the approximation while h. . .iπ stands for an average over the distributions πs,i (τ ) = N X X Xs,i (t) (10) t=P τ 3 confirm Eq. (17) is valid even for α < αc . Fig. 2 compares the results of numerical simulations of the MG, as a function of Γ, with those of a semi-analytic solution of Eqs. (15-17), to be discussed later. The agreement of the two approaches shows that Eqs. (15-17) are valid even in the symmetric phase (α ≤ αc ) for all values of Γ. The instantaneous probability distribution of s, in the continuum time limit, reads this in general [26] (Eq. (13) holds trivially in the limit Γ → 0). Within this approximation, the correlation, for N ≫ 1, become hdWs,i (τ )dWr,j (τ ′ )i ∼ = σ2 as,i ar,j δ(τ − τ ′ )dτ αN (14) Note that, for r 6= s or j 6= √i, correlations hdWs,i (τ )dWr,j (τ ′ )i ∝ as,i ar,j ∼ 1/ N vanishes as N → ∞. However it is important to keep the offdiagonal terms because they keeps the dynamics of the phase space point |U (t)i = {Us,i (t)}s=1,...,S, i=1,...,N constrained to the linear space spanned by the vectors |aµ i = {aµs,i }s=1,...,S, i=1,...,N which contains the initial condition |U (0)i. The original dynamics of Us,i (t) indeed posses this property. It is important to remark that the approximation Eq. (13) makes our approach a self-consistent theory for σ 2 . We introduce σ 2 as a constant in Eq. (13) which then has to be computed self-consistently from the dynamic equations. eΓi us,i (τ ) πs,i (τ ) = P Γ u (τ ) . i r,i re This and Ito calculus then lead to a dynamic equation for πs,i (τ ). We prefer to exhibit this for Γi = Γ and using the rescaled time t = Γτ : i h dπs,i = −πs,i as,i hAi − ~πi · ~ai hAi dt  √ σ2 Γ + πs,i πs,i − ~πi2 + Γπs,i (ηs,i − ~πi ~ηi ) . (19) αN The first term in the r.h.s. comes from the deterministic part of Eq. (15), the second from the Ito term √ (where we neglected terms proportional to ai,s ai,s′ ∼ 1/ N for s 6= s′ ) and the third from the stochastic part. It is clear that, in the limit Γ → 0 the last two term vanish and the dynamics becomes deterministic. We see then that Γ tunes the strength of stochastic fluctuations in much the same way as temperature does for thermal fluctuations in statistical mechanics. The “individual inverse temperature” Γ should indeed more correctly be interpreted as a learning rate§ . Furthermore it plays the role of a “global temperature”. We shall pursue this discussion in detail below for the case S = 2. At this point, let us comment on the limit Γ → ∞, which is of particular importance, since it corresponds to the original MG. It is clear that in the limit Γ → ∞ the dynamical equations (19) become problematic. The origin of the problem lies in the order in which the limits N → ∞ and Γ → ∞ is performed. Indeed in Eq. (8) Us,i (t) ≃ us,i (τ ) + O(dτ ) for P τ ≤ t < P (τ + dτ ). Therefore, as long as Γdτ ≪ 1 the difference between Eq. (8) and πs,i in Eq. (18) is negligible. In practice, in order to satisfy both Γdτ ≪ 1 and P dτ ≫ 1, one needs Γ ≪ P . When this condition is not satisfied the instantaneous probability Eq. (1) fluctuates very rapidly at each time-step. Eq. (8) averages out these high frequency fluctuations so that, even for Γ = ∞, the distribution πs,i (τ ) of Eq. (8) is not a discontinuous step function of 10 fluctuations X αX 2 σ /N 1 0.1 0.1 1 α 10 FIG. 1. Noise strength averaged over all agents (stars); when it is multiplied by α (diamonds), one recovers σ 2 /N (circles) (P = 32, S = 2, 300P , iterations, Γ = ∞, average over 50 samples). Dashed lines are for eye guidance only Summarizing, the dynamics of us,i is described by a continuum time Langevin equation: dus,i (τ ) = −as,i hAi + ηs,i (τ ) dτ hηs,i (τ )i = 0 (18) (15) (16) 2 σ hηs,i (τ )ηr,j (τ ′ )i ∼ as,i ar,j δ(τ − τ ′ ). = αN (17) § This is an a posteriori learning rate. Indeed 1/Γ is the time the dynamics of the scores needs in order to learn a payoff difference. From a different viewpoint, Γ tunes the randomness of the response of agents. The larger the randomness, the longer it takes to average fluctuations out. Equation (15), given its derivation, has to be interpreted in the Ito sense. The expression for the noise strength by figure 1 where the measure of P is confirmed 2 i/(P N S) in a MG is reported; note that X = i,s h(Xi,s these numerical simulations were done for Γ = ∞ and 4 (22) are one and the same problem∗∗ . In other words fs,i can be computed from the constrained minimization of H as proposed in Refs. [3,4]. Hence the statistical mechanics approach based on the study of the ground state of H is correct. This approach gives the frequency fs,i with which agents play their strategies. We remark once more that H is a function of the stationary state probabilities fs,i . Also note that us,i (τ ), as suggested by Eq. (18). High frequency fluctuations contribute to the functional form of πs,i on us,i which will differ from Eq. (18). Summarizing, when we let Γ → ∞ only after the limit N → ∞ has been taken no problem arises. There is no reason to believe that results change if the order of the limits is interchanged. This expectations, as we shall see, is confirmed by numerical simulations (see Fig. 2): direct numerical simulations of the MG deviate from the prediction of Eqs. (15-17) only for finite size effects which vanish as N → ∞. Eqs. (15-17) are our central result. We shall devote the rest of the paper to discuss their content and to show that all of the observed behavior of the MG can be derived from these equations. H̃{πs,i } = Let us take the average, denoted by h. . .i, of Eq. (15) on the stationary state (SS). Let fs,i = hπs,i i be the frequency with which agent i plays strategy s in the SS. Then we have   X dUs,i = −as,i hAi, hA|µi = fs′ ,j aµs′ ,j vs,i ≡ dτ ′ j,s A. Independence on Γ for α > αc Given the relation between πs,i and Us,i and considering that the long time dynamics of Us,i in the SS is Us,i (τ ) = const+vs,i τ , we have that i) each strategy which is played in the SS by agent i must have the same “velocity” vs,i = vi∗ , and ii) strategies which are not played (i.e. with fs,i = 0) must have vs,i < vi∗ . In other words −as,i hAi ≤ vi∗ , When α > αc the solution to Eq. (22) is unique and H > 0. Hence fs,i does not depends on Γ, neither does H. In addition we shall see that hπs,i πs′ ,j i = hπs,i ihπs′ ,j i = fs,i fs′ ,j min {fs,i ≥0} for i 6= j (23) implying that ∀i, s such that fs,i > 0 ∀i, s such that fs,i = 0. (20) σ2 ≡ N + (21) Consider now the problem of constrained minimization of H in Eq. (4), subject to fs,i ≥ 0 for all s, i and the normalization conditions. Introducing Lagrange multipliers P λi to enforce s fs,i = 1 for all i, this problem reads ( as,i as′ ,j πs,i πs′ ,j i,j=1 s,s′ =1 as a function of the instantaneous probabilities πs,i , is not a Lyapunov function of the dynamics. The dynamical variables πs,i √ (t) are subject to stochastic fluctuations of the order of Γi around their average values fs,i . Only in the limit Γi → 0, when the dynamics becomes deterministic and πs,i → fs,i , the quantity H̃{πs,i } becomes a Lyapunov function. The solution to the minimization of H reveals two qualitatively distinct phases [3,4] which are separated by a phase transition occurring as α → αc . We discuss qualitatively the behavior of the solution for a generic S and leave for the next section a more detailed discussion in the simpler case S = 2. IV. STATIONARY STATE − as,i hAi = vi∗ , N S X X hAi2 − N X i=1 λi 1− S X s=1 fs,i !) . XX i6=j s,r as,i ar,j hπs,i πr,j i does not depend on Γ either. Hence the solution {fs,i } uniquely determines all quantities in the SS, as well as ∗∗ Indeed both problems can be put in the form of a Linear Complementarity problem [27]: (22) Taking derivatives, we find that if fs,i > 0 then as,i hAi + λi = 0 whereas if fs,i = 0 then as,i hAi + λi ≥ 0. These are exactly Eqs. (20,21) where vi∗ = λi . We then conclude that the two problems Eqs. (20,21) and Eq. X j,s′ fs,i " X j,s′ fs,i ≥ 0 as,i as′ ,j fs′ ,j + vi∗ ≥ 0 as,i as′ ,j fs′ ,j + vi∗ # =0 This problem has a solution for all values of vi∗ because of non-negativity of the matrix as,i as′ ,j , see Ref. [27]. 5 the parameters which enter into the dynamics (notice the dependence on σ 2 in Eq. 17). In particular σ 2 does not depend on Γ. V. THE CASE S = 2 We work in this section with the simpler case of S = 2 strategies, labelled by s = ±. We also set Γi = Γ for all i. Following Refs. [3,13] we introduce the variables B. Dependence on Γ and on initial conditions for α < αc ξiµ = For α < αc the solution to the minimization problem is not unique: there is a connected set of points {fs,i } such that H = 0. Let us first discuss the behavior of the system in the limit Γ → 0, where the dynamics becomes deterministic. The dynamics reaches a stationary state {fs,i } which depends on the initial conditions. In order to see this, let us introduce the vector notation |vi = {vs,i , s = 1, . . . , S, i = 1, . . . , N }. Then for all times |u(τ )i is of the form |u(τ )i = |u(0)i + P X µ=1 µ aµ+,i − aµ−,i , 2 Ωµ = N X aµ+,i + aµ−,i i=1 2 . Let us rescale time t = Γτ and introduce the variables yi (t) = Γ U+,i (τ ) − U−,i (τ ) 2 Then, using Eq. (8), the dynamical equations (15-17) become N X dyi = −ξi Ω − ξi ξj tanh(yj ) + ζi dt j=1 µ |a iC (τ ) hζi (t)ζj (t′ )i = where C µ (τ ) are P functions of time. If there are vectors hv| such that hv|aµ i = 0 for all µ, then hv|u(τ )i = hv|u(0)i, i.e. the components of the scores will not change at all along these vectors. As a result the SS will depend on initial conditions |u(0)i. These vectors hv| exist exactly for α < αc [4], because the “dimensionality” of the vectors |u(τ )i is larger than P †† . The picture is made even more complex by the fact that, for α < αc , when Γ is finite Eq. (23) does not hold. Hence σ 2 has a contribution which depends on the stochastic fluctuations around fs,i . The strength of these fluctuations, by Eqs. (17,19), depends on Γ and σ 2 itself. We face, in this case, a self-consistent problem: σ 2 enters as a parameter of the dynamics but should be computed in the stationary state of the dynamics itself. Therefore the solution to this problem and hence σ 2 depends on Γ. The solution {fs,i } to the minimization of H should also be computed self-consistently. As a result, the SS properties acquire a dependence on Γ. The condition (23), which is similar to the clustering property in spin glasses [28], plays then a crucial role. We show below how the condition (23), the dependence on initial conditions and on Γ enter into the detailed solution for S = 2. By similar arguments our conclusion can be generalized to all S > 2. Γσ 2 ξi ξj δ(t − t′ ). αN (24) (25) The Fokker-Planck (FP) equation for the probability distribution P ({yi }, t) under this dynamics reads  N N X ∂P ({yi }, t) X ∂  ξi Ω + ξi ξj tanh(yj )+ = ∂t ∂yi  j=1 i=1  N 1X ∂  ξi ξj P ({yi }, t) (26) β ∂yj  j=1 where we have introduced the parameter β= 2αN . Γσ 2 (27) Multiply Eq. (26) by yi and integrate over all variables. Using integration by parts, assuming that P → 0 fast as yj → ∞, one gets N X ∂ ξi ξj htanh(yj )i . hyi i = −ξi Ω − ∂t j=1 Let us look for solutions with hyi i ∼ vi t and define mi = htanh(yi )i. Hence for t → ∞ we have vi = −ξi Ω − †† In order to compute the dimensionality of the vectors |ui we have to take into account the N normalization conditions and the fact that strategies which are not played (fs,i = 0) should not be counted. So if there are N> variables fs,i > 0, the relevant dimension of the space of |ui is N> − N . Hence vectors hv| orthogonal to all |aµ i exist for N> − N > P , i.e. for α < αc = N> (αc )/N − 1. N X ξi ξj mj . (28) j=1 Now, either vi = 0 and hyi i is finite, or vi 6= 0, which means that yi → ±∞ and mi = sign vi . In the latter case (vi 6= 0) we say that agent i is frozen [13], we call F the set of frozen agents and φ = |F |/N the fraction of frozen agents. 6 As in the general case, the parameters vi for i ∈ F and mi ≡ htanh(yi )i for i 6∈ F are obtained by solving the constrained minimization of #2 " N P X 1 X ξiµ mi . Ωµ + H= P µ=1 i=1 A. α > αc For α > αc the solution of min H is unique, and hence mi depends only on the realization of the disorder. In addition, the number N − |F | ≡ N (1 − φ) of unfrozen agents is less than P and the constraint is ineffective, i.e. Py(0) ≡ 1. The scores |yi of unfrozen agents span a linear space which is embedded in the one spanned by the vectors |ξ µ i. Hence the dependence on initial conditions yi (0) drops out. Therefore the probability distribution of yi factorizes, as in Eq. (29). Then the third term of Eq. (32), which is the only one which depends on β, vanishes identically. We conclude that, for α > αc , σ 2 only depends on mi and is hence independent of Γ as confirmed by numerical simulations. Summarizing, for α > αc one derives a complete solution of the MG by finding first the minimum {mi } of H and then by computing σ 2 , β and the full distribution of yi from Eq. (29). When the solution of min H is unique, i.e. for α > αc , the parameters mi depend only on the realization of disorder {ξiµ , Ωµ }, and their distribution can be computed as in Ref [3]. When the solution is not unique, i.e. for α < αc , we are left with the problem of finding which solution the dynamics selects. Let us suppose that we have solved this problem (we shall come back later to this issue), so that all mi are known. Using the stationary condition Eq. (28), we can write the FP equation for the probability distribution Pu (yi , i 6∈ F) of unfrozen agents. For times so large that all agents in F are indeed frozen (i.e. si (t) = sign vi ) this reads:   X ∂ X ∂Pu 1 ∂ Pu . = ξi ξj tanh(yj ) − mj + ∂t ∂yi β ∂yj B. α ≤ αc : dependence on Γ and crowd effects j6∈F i6∈F This has a solution     X Pu ∝ exp −β [log cosh yj − mj yj ] .   When α < 1 − φ, on the other hand, the solution of min H is not unique. Furthermore the constraint cannot be integrated out and the stationary distribution depends on the initial conditions. Numerical simulations [8] show that σ 2 increases with Γ for α < αc (see Fig. 2). This effect has been related to crowd effects in financial markets [12]. Ref. [6] has shown that crowd effects can be fully understood in the limit α → 0: as Γ exceeds a critical learning rate Γc , the time independent SS becomes unstable and a bifurcation to a period two orbit occurs. Neglecting the stochastic term ζi , Ref. [6] shows that this picture can be extended to α > 0‡‡ . This approach suggests a crossover to a (29) j6∈F Finally we have to impose the constraint that |y(t)i = {yi (t)}N i=1 must lie on the linear space spanned by the vectors |ξ µ i which contains the initial condition |y(0)i. This means that     X Pu ∝ Py(0) exp −β [log cosh yj − mj yj ] (30)   j6∈F where the projector Py(0) is given by # " P P Z ∞ N X Y Y µ µ µ δ yi − yi (0) − c ξi . (31) Py(0) ≡ dc −∞ µ=1 i=1 µ=1 ‡‡ The idea of Ref. [6] is the following: imagine that our system is close to a SS point yi∗ at time t = tk , when µ(t) = 1. Will the system be close to yi∗ when the pattern µ = 1 occurs again the next time t′ = tk+1 ? To see this, let us integrate Eq. (24) from t = tk to tk+1 . In doing this we i) neglect the noise term (i.e. ζi = 0) and ii) assume that tanh yi (t) ≈ tanh yi (tk ) stays constant in the integration time interval. This latter assumption is similar to the recently introduced [7] batch version of the MG, where agents update their strategies every P time-steps. This leads to study a discrete time dynamical system We find it remarkable that Γ, which is introduced as the inverse of an individual “temperature” in the definition of the model, actually turns out to be proportional to β −1 (see Eq. 27) which plays collectively a role quite similar to that of temperature. Using the distribution Eq. (30), we can compute σ2 = H + N X i=1 X i6=j ξi2 (1 − m2i ) + ξi ξj h(tanh yi − mi )(tanh yj − mj )i. (32) " yi (tk+1 ) = yi (tk ) − Γ Ωξi + This depends on β, i.e. on σ 2 itself by virtue of Eq. (27). The stationary state is then the solution of a selfconsistent problem. Let us analyze in detail the solution of this self-consistent problem. N X j=1 # ξi ξj tanh[yj (tk )] (33) where the factor Γ comes because tk+1 −tk is on average equal to Γ. The linear stability of fixed point solutions is analyzed setting yi (tk ) = yi∗ + δyi (k) and computing the eigenvalues of 7 N N 1 X 1 X hsi (t)si (0)i = mi t→∞ N N i=1 i=1 “turbulent” dynamics for Γ > Γc , where 4 √ 2 Γc (α) = (1 + α) (1 − Q) M ≡ lim (34) which measures the overlap of the SS configuration with the initial condition. Symmetric initial conditions are related to M = 0 SS. These are the states we focus on. The solution is derived in two steps: and N 1 X 2 Q= m . N i=1 i 1. find the minimum {mi } of H, with M = 0; Both Q and Γc can be computed exactly in the limit N → ∞ within the statistical mechanics approach [3,6]. This approach however i) does not properly takes into account the stochastic term, ii) does not explain what happens for Γ > Γc and iii) does not explains why such effects occur only for α < αc . 2. compute self-consistently σ 2 . The numerical procedure for solving the problem is the following: given the realization of disorder {ξiµ , Ωµ }, step (1) — finding the minimum {mi } of H — is straightforward. For step (2) we sample the distribution Eq. (30) with the Montecarlo method§§ at inverse temperature β and measure the β dependent contribution of σ 2 in Eq. (32): X Σ(β) = ξi ξj h(tanh yi − mi )(tanh yj − mj )iβ . 4 N=160 MG N=320 MG N=640 MG N=160 Montecarlo N=320 Montecarlo N=640 Montecarlo Theory Γ −> 0 2 i6=j Here h. . .iβ stands for an average over the distribution Eq. (30) with parameter β. Finally we solve the equation   2αN 2 2 σ (Γ) = σ (0) + Σ . (36) Γσ 2 (Γ) 0.8 2 σ (Γ)/Ν 3 Γc 0.6 1 0.4 0 −2 10 10 −1 (35) 10 0 Γ 0 0.1 0.2 10 1 0.3 10 This procedure was carried out for different system sizes and several values of Γ. The results, shown in Fig. 2, agree perfectly well with direct numerical simulations of the MG. Actually Fig. 2 shows that, for Γ ≫ 1, the solution of the self-consistent equation (36) suffers much less of finite size effects than the direct numerical simulations of the MG. Fig. 2 also shows that, even if only approximate, Eq. (34) provides an useful estimate of the point where the crossover occurs. It is possible to compute σ 2 (Γ) to leading order in Γ ≪ 1. The calculation is carried out in the appendix in detail. The result is   1 − Q + α(1 − 3Q) σ2 ∼ 1 − Q 2 1+ Γ + O(Γ ) . (37) = N 2 4α 2 FIG. 2. Global efficiency σ 2 /N versus Γ for α = 0.1 < αc and different system sizes. Lines refer to direct simulations of the MG with N = 160, 320 and 640. Finite size effect for Γ ≫ 1 are evident. Symbols refer instead to the solution of the self-consistent equation (36) for the same system sizes. For both methods and all values of N , σ 2 is averaged over 100 realizations of the disorder. The arrow marks the location of Γc predicted by Eq. (34). In the inset, the theoretical prediction Eq. (37) on the leading behavior of σ 2 /N for Γ ≪ 1 (solid line) is tested against numerical simulations of the MG (points) for the same values of N . The stochastic dynamics derived previously gives detailed answers to all these issues. We first restrict attention to symmetric initial conditions yi (0) = 0 and then discuss the dependence on initial conditions. The choice of yi (0) = 0 ∀i is convenient because it allows one to use this same symmetry to identify the solution {mi } to the minimization of H, independently of Γ. To be more precise, one can introduce a “magnetization” §§ The Montecarlo procedure follows the usual basic steps: i) A move yi → yi + ǫξiµ is proposed, with µ and ǫ drawn at random, ii) the “energy” E{yi } = N X i=1 [log cosh yi − mi yi ] of the new configuration is computed and iii) The move is accepted with a probability equal to min(1, e−β∆E ) where ∆E is the “energy” difference. P the linearized map δyi (k+1) ≃ j Ti,j δyj (k). There is a critical value of Γ above which the solution yi∗ become unstable, which is given by Eq. (34). 8 and χ = β(Q − q)/α is a “spin susceptibility”. There are two possible solutions: one with χ < ∞ finite as η → 0 which describes the α > αc phase. The other has χ ∼ 1/η which diverges as η → 0. This solution describes the α < αc phase. We focus on this second solution, which can be conveniently parameterized by two parameters z0 and ǫ0 . We find  if z ≤ −z0 − ǫ0  −1 z+ǫ0 if −z s0 (z) = 0 − ǫ0 < z < z0 − ǫ0  z0 1 if z ≥ z0 − ǫ0 The inset of Fig. 2 shows that this expression indeed reproduces quite accurately the small Γ behavior of σ 2 . Note finally that Eq. (36) has a finite solution σ 2 (∞) = σ 2 (0) + Σ(0) in the limit Γ → ∞. Furthermore it is easy to understand the origin of the behavior σ 2 /N ∼ 1/α for Eq. (36). Because of the constraint, when ξi ξj is positive (negative) the fluctuations of tanh(yi )− mi are positively (negatively) correlated with tanh(yj ) − mj . If we assume that h[tanh(yi ) − mi ][tanh(yj ) − mj ]i ≃ cξi ξj for some P 2 constant c, we find Σ ≃ c i6=j ξi ξj . This leads easily to Σ/N ≃ c/(4α), which explains the divergence of σ 2 /N as α → 0 for Γ ≫ 1. Indeed Eq. (38) p gives Q(z0 , ǫ0 )R and Eq. (39) which for χ → ∞ reads α(1 + Q) = Dzzs0 (z), then gives α(z0 , ǫ0 ). With ǫ0 6= 0 one finds solutions with a non-zero “magnetization” M = hsi i. This quantity is particularly meaningful, in this context, because it measures the overlap of the behavior of agents in the SS with their a priori preferred strategies C. Selection of different initial conditions in the Replica calculation As discussed above, the stationary state properties of the MG in the symmetric phase depend on the initial conditions. Can the statistical mechanics approach to the MG [3,4] be extended to characterize this dependence for Γ ≪ 1? If this is possible, how do we expect the resulting picture to change when Γ increases? We shall first focus on the first question (i.e. Γ ≪ 1) and then discuss the second. Of course one can introduce the constraint on the distribution of yi in the replica approach in a straightforward manner. This leads however to tedious calculations. We prefer to follow a different approach. In the α < αc phase the minimum of H is degenerate, i.e. H = 0 occurs on a connected set of points. Each of these points corresponds to a different set of initial conditions, as discussed above. In order to select P a particular point with H = 0 we can add a potential η i (si − s∗i )2 /2 to the Hamiltonian H, which will favor the solutions closer to s∗i , and then let the strength η of the potential go to zero. This procedure lifts the degeneracy and gives us the statistical features of the equilibrium close to s∗i . The nature of the stationary state changes as the asymmetry in the initial conditions changes. If we take s∗i = s∗ , the state at s∗ = 0 describes symmetric initial conditions and increasing s∗ > 0 gives asymmetric states. The saddle point equations of the statistical mechanics approach of Ref. [3] can be reduced to two equations: Z ∞ Q= Dzs20 (z) (38) −∞ Z ∞ 1+χ Dzzs0 (z) (39) χ= p α(1 + Q) −∞ N 1 X hsi (t)si (0)i. Dz s0 (z) = lim M≡ t→∞ N −∞ i=1 Z ∞ (41) Note indeed that one can always perform a “gauge” transformation in order to redefine s = +1 as the initially preferred strategy. This amounts to taking yi (0) ≥ 0 for all i. Which SS is reached from a particular initial condition is, of course, a quite complex issue which requires the integration of the dynamics. However, the relation between Q and M derived analytically from Eqs. (38,41) can easily be checked by numerical simulations of the MG. Figure 3 shows that the self-overlap Q and the magnetization M computed in numerical simulations with initial conditions yi (0) = y0 for all i, perfectly match the analytic results. The inset of this figure shows how the final magnetization M and the self-overlap Q in the SS depend on the asymmetry y0 of initial conditions. 1 1 M Q 0.8 0.5 Q 0.6 0 −1 0 10 10 y 0 0.4 1 2 10 10 2 where Dz = √dz e−z /2 , s0 (z) ∈ [−1, 1] is the value of s 2π which minimizes r 1 1+Q 1 2 s − zs + η(1 + χ)(s − s∗ )2 (40) Vz (s) = 2 α 2 0.2 0 0 0.2 0.4 M 9 0.6 0.8 FIG. 3. Relation between Q and M , for α = 0.1, derived from analytic calculation (full line) and from numerical simulations of the MG with different initial conditions y0 (⋄, P = 32, N = 320, Γ = 0.1). The inset shows the dependence of Q and M from the initial condition y0 . D. The maximally magnetized stationary state The maximally magnetized SS (MMSS), obtained in the limit y0 → ∞, is also the one with the largest value of Q, and hence with the smallest value of σ 2 = N (1 − Q)/2. σ 2 /N is plotted against α both for symmetric y0 = 0 initial conditions and for maximally asymmetric ones y0 → ∞ in fig. 5. The inset shows the behavior of Q and M in the MMSS. Let us finally discuss the dependence on Γ for asymmetric initial conditions. Eq. (34) provides a characteristic value of Γ as a function of α and Q. This theoretical prediction is tested against numerical simulations of the MG in Fig. 4: when plotted against Γ/Γc , the curves of σ 2 /N obtained from numerical simulations, approximately collapse one onto the other in the large Γ region. Fig. 4 suggests that Γc in Eq. (34) provides a close lower bound for the onset of saturation to a constant σ 2 for large Γ. We find it remarkable that a formula such as Eq. (34) which is computed in the limit Γ → 0, is able to predict the large Γ behavior. 1.00 1 0.8 0.6 0.4 2 σ /Ν 0 2 Γc= 0.260 Γc= 0.447 Γc= 0.757 Γc=1.382 Γc=2.099 0.00 −2 10 0.1 0.2 α 0.3 −1 10 0 α 10 1 10 FIG. 5. σ 2 /N for the MG with initial conditions y0 = 0 (full line) and y0 → ∞ (dashed line). The inset reports the behavior of M and Q in the y0 → ∞ SS. Note that Q is linear in α. 1 0 −1 10 0 0.50 y0=0 M=Mmax 2 σ /Ν 3 M(α) Q(α) 0.2 0 10 1 Γ/Γc 10 Remarkably we find that σ 2 /N vanishes linearly with α in the MMSS∗∗∗ . This means that, at fixed P , as N increases the fluctuation σ 2 remains constant. This contrast with what happens in the y0 = 0 state, for Γ ≪ Γc , where σ 2 increases linearly with N , and with the case Γ ≫ Γc where σ 2 ∝ N 2 [18,6]. Note also that the lowest curve of Fig. 5 also gives an upper bound to the σ 2 of Nash equilibria (see Refs. [3,4,29]). The MMSS is also the most stable state against crowd effects: if we put Q(α, y0 = ∞) ∼ = 1 − cα, as appropriate for the MMSS we find that Γc ∼ 1/α diverges with α. 2 10 FIG. 4. Global efficiency σ 2 /N versus Γ/Γc for α = 0.1 < αc , N = 160 agents and different initial conditions y0 . The value Γc is computed from Eq. (34) and is shown in the legend. With respect to the dependence on initial conditions, we observe that Γc is an increasing function of Q and hence it increases with the asymmetry y0 of initial conditions. Hence, for a fixed Γ, the fluctuation dependent part Σ of σ 2 decreases with y0 because it is an increasing function of Γ/Γc . This effects adds up to the decrease of the Γ independent part of σ 2 discussed previously. Fig. 4 also shows that the Γ ≫ Γc state is independent of initial conditions y0 . This can naively be understood observing that stochastic fluctuations induce fluctuations δyi which increase with Γ. For Γ ≫ 1 the asymmetry y0 of initial conditions is small compared to stochastic fluctuations δyi and hence the system behaves as if y0 ≈ 0. VI. CONCLUSIONS We have clarified the correct derivation of continuous time dynamics for the MG. This on the one hand reconciles the two current approaches [3,9]. On the other it leads to a complete understanding of the collective behavior of the MG. We confirm that stationary states are characterized by the minimum of a Hamiltonian which ∗∗∗ 10 This results was also found analytically in [7]   P Z √ 1 X ∞ dx −x2 /2 X µ √ e Hg = fs,i as,i + Dx G P µ=1 −∞ 2π i,s measures the predictability of the game. For α > αc we find a complete analytic solution, whereas for α < αc the statistical mechanics approach of Ref. [3] is valid for Γ → 0. It is in principle possible to introduce the new elements discussed here in the approach of Ref. [3] and to derive a full analytic solution. We have indeed derived the first term of the series expansion for Γ ≪ 1, which agrees perfectly with numerical data. The extension of the approach of Ref. [3] involves lengthy calculations and it shall be pursued elsewhere. Finally we note that the results derived in this paper generalize to more complex models. It is worth to remark that the solution to the FP equation is no more factorizable, in general, when agents account for their market impact as in Refs. [3,6,4]. Hence, as long as there are unfrozen agents, we expect that the stationary state depends on Γ. However, when the agents take fully into account their market impact, all of them are frozen and the conclusion that agents converge to Nash equilibria remains valid. We acknowledge constructive discussions with A.C.C. Coolen, A. Engel, J.P. Garrahan, J.A.F. Heimel and D. Sherrington. with g(x) = dG(x) dx and D = σ 2 − H which must be determined selfconsistently. Indeed taking the derivative of Hg w.r.t. fs,i and imposing the constraint fs,i ≥ 0 and normalization, we arrive at exactly the same equations which describe the stationary state of the process. The Hamiltonian for the original MG is derived setting g(x) = sign x, which leads to Hsign   P  1 X 1 −hA|µi2 /D hA|µi hA|µi √ √ √ = . e + erf P µ=1 π D D vi = −as,i hg(A)i, if fs,i > 0 The analysis of stochastic fluctuations can be extended to non-linear cases in a straightforward manner. Again the key point is that the dynamics is constrained to the linear space spanned by the vectors |aµ i. For α > αc we have no dependence on initial conditions. However it is not easy to show in general that the distribution of scores factorizes across agents. This means that there may be a contribution of fluctuations to σ 2 – i.e. Σ > 0 – so we cannot rule out a dependence of σ 2 on Γ. Numerical simulations for g(x) = sign x show that such a dependence, if it exists, is very weak. Anyway even though σ 2 only depend on fs,i , the minimization problem depends on D = σ 2 − H which must then be determined self-consistently. For α < αc the dependence on initial conditions induces a correlation of scores across agents. As a result σ 2 depends on Γ just as in the linear case discussed above. vi > −as,i hg(A)i, if fs,i = 0. APPENDIX B: SMALL Γ EXPANSION APPENDIX A: NON-LINEAR MINORITY GAMES Take a generic dynamics µ(t) Us,i (t + 1) = Us,i (t) − as,i g [A(t)] with g(x) some function. When we carry out the limit to continuous time we find a deterministic term which is proportional to −as,i hg(A)i. The stationary state conditions then read and For any fixed µ, A(t) is well approximated by a Gaussian variable with mean X fs,i aµs,i hA|µi = For Γ ≪ 1 it is appropriate to consider β ≫ 1 and to take zi yi = arc tanh mi + √ β i,s and variance D = σ 2 − H. Here we neglect dependences on µ. Also we treat D as a parameter and neglect its dependence on the stationary state probabilities fs,i . Hence we can write Z ∞  √  2 dx √ e−x /2 g hA|µi + Dx hg(A)|µi = 2π −∞ so that β[log cosh yi − mi yi ] ≃ 21 (1 − m2i )zi2 + O(β −1/2 ). Hence we have to sample a distribution P 2 2 1 P {zi } ∝ e− 2 i (1−mi )zi where zi has the form The stationary state conditions of the dynamics above can again be written as a minimization problem of the functional zi = P X µ=1 11 cµ ξiµ . It is convenient to express everything in terms of the coefficients cµ . Their pdf is derived from that of zi and it reads: P {cµ } ∝ e − 21 P µ,ν cµ T µ,ν cν T µ,ν = , N X i=1 [5] Challet D., M. Marsili and Y.-C. Zhang, Physica A 276, 284 (2000) e-print cond-mat/9909265 [6] M. Marsili and D. Challet, Adv Compl. Sys. 3-I (2001), e-print cond-mat/0004376 e-print cond-mat/0004196 [7] J. A. F. Heimel, A. C. C. Coolen, Phys. Rev. E 63, 056121 (2001), e-print cond-mat/0012045. [8] A. Cavagna, J.P. Garrahan, I. Giardina, D. Sherrington, Phys. Rev. Lett. 83, 4429 (1999) e-print condmat/9903415 (1999). [9] J. P. Garrahan, E. Moro, D. Sherrington, Phys. Rev. E 62, R9 (2000) [10] D. Challet, M. Marsili and R. Zecchina, Phys. Rev. Lett. 85, 5008 (2000), e-print cond-mat/0004308 [11] A. Cavagna, J. P. Garrahan, I. Giardina, D. Sherrington, Phys. Rev. Lett. 85, 5009 (2000), e-print condmat/0005134 [12] M. Hart, P. Jefferies, P. Hui, N. F. Johnson, e-print condmat/0008385 [13] D. Challet and M. Marsili, Phys. Rev. E 60, 6271 (1999), e-print cond-mat/9904071 [14] The learning process belongs to the class of models described in C. Camerer and T.-H. Ho, Econometrica 67, 827 (1999) with φ = κ = δ = 1. [15] A. Cavagna, Phys. Rev. E 59, R3783 (1999) [16] D. Challet and M. Marsili, Phys Rev. E 62, 1862 (2000), [17] Y.-C. Zhang, Evolving models of financial markets, Europhys. News 29, 51 (1998). [18] Savit R., Manuca R., and Riolo R., Phys. Rev. Lett. 82(10), 2203 (1999), e-print adap-org/9712006. [19] D. Challet, A. Chessa, M. Marsili, Y.-C. Zhang, Quant. Fin. I-1 (2001), e-print cond-mat/0011042 [20] F. Slanina and Y.-C. Zhang, Physica A 272, 257 (1999), e-print cond-mat/9906249 [21] P. Jefferies, M.L. Hart, P.M. Hui, N.F. Johnson, Int. J. Th and Appl. Fin. 3-3 (2000), e-print cond-mat/9910072. [22] T. S. Lo, P. M. Hui, N. F. Johnson, e-print condmat/0008387 [23] J.-P. Bouchaud, I. Giardina, M. Mezard, e-print condmat/0012156 [24] D. Challet, M. Marsili and Y.-C. Zhang, Physica A 294, 514 (2001), e-print cond-mat/0101326 √ [25] Indeed aµs,i hA|µiπ is of order N but its sign fluctuates. When we average over P ∼ N different µ, we get a quantity of order N 0 . We are implicitly assuming that hAi ≃ 0 which definitely holds for large times. [26] We thank J. A. F. Heimel for pointing out this issue. [27] Katta G. Murty, Linear Complementarity, Linear and Nonlinear Programming, Heldermann Verlag, (1988) [28] M. Mezard, G. Parisi, M. A. Virasoro, Spin glass theory and beyond World Scientific (1987). [29] A. De Martino and M. Marsili (2000), e-print condmat/0007397 (1 − m2i )ξiµ ξiν . From this we find hcµ cν i = [T −1 ]µ,ν . Now we split the term Σ(β) in two contributions, " Σ(β) = h N X i=1 N X i=1 #2 ξi (tanh yi − mi ) i+   ξi2 m2i − (tanh yi )2 and work them out separately. For the first we use N X 1 X µ,ν ν ξiµ (tanh yi − mi ) = √ T c β ν i=1 so that " h N X i=1 #2 ξi (tanh yi − mi ) i= = 1 X µ,ν µ,γ ν γ T T hc c i βP µ,ν,γ TrT ∼ 1 − Q N. = βP 2β Within the approximation (1 − 3m2i ) ≈ (1 − 3Q) we are able to derive a closed expression also for the second term: N X N   1X 2 ξi2 m2i −(tanh yi )2 = ξi (1−m2i )(1−3m2i )hzi2 i β i=1 i=1 ≈ α 1 − 3Q N. β 2 Hence we find   1 − Q 1 − 3Q 1 ∼ Σ(β) = + α + O(β −2 ). 2 2 β This and Eq. (27) lead to Eq. (37). [1] Challet D. and Zhang Y.-C., Physica A 246, 407 (1997) e-print adap-org/9708006 [2] See the Minority Game’s web page on http://www.unifr.ch/econophysics. [3] D. Challet, M. Marsili and R. Zecchina, Phys. Rev. Lett. 84, 1824 (2000), e-print cond-mat/9904392 [4] M. Marsili, D. Challet and R. Zecchina, Physica A 280, 522 (2000), e-print cond-mat/9908480 12