15
American Political Science Review Vol. 98, No. 1 February 2004 “Machiavellian” Intelligence as a Basis for the Evolution of Cooperative Dispositions JOHN ORBELL University of Oregon TOMONORI MORIKAWA Waseda University JASON HARTWIG University of Oregon JAMES HANLEY University of Illinois, Springfield NICHOLAS ALLEN University of Melbourne H ow to promote cooperative behavior is classically solved by incentives that lead self-interested individuals in socially desirable directions, but by now well-established laboratory results show that people often do act cooperatively, even at significant cost to themselves. These results suggest that cooperative dispositions might be an evolved part of human nature. Yet such dispositions appear inconsistent with the “Machiavellian intelligence” paradigm, which develops the idea that our brains have evolved, in substantial part, for capturing adaptive advantage from within-group competition. We use simulation to address the evolutionary relationship between basic Machiavellian capacities and cooperative dispositions. Results show that selection on such capacities can (1) permit the spread of cooperative dispositions even in cooperation-unfriendly worlds and (2) support transitions to populations with high mean cooperative dispositions. We distinguish between “rationality in action” and “rationality in design”—the adaptive fit between a design attribute of an animal and its environment. The combination of well-developed Machiavellian intelligence, modest mistrust, and high cooperative dispositions appears to be a rational design for the brains of highly political animals such as ourselves. H ow to promote group-benefiting action when individuals have strong incentives against such action is perhaps the fundamental problem of political theory (Axelrod 1981; Ostrom 1998). Clas- sically spelled out by Hobbes ([1651] 1947) but now usually framed in terms of the Prisoner’s Dilemma, the issue is how to persuade individuals to act in the inter- ests of the collectivity (to “cooperate”) when there is a clear incentive for them to do otherwise (to “defect”). The incentive structure means that there is only one so- lution for rational and self-interested individuals: The incentives they confront must, somehow, be changed so that cooperation rather than defection offers the greater return. We now understand that this does not necessarily imply a centralized Hobbesian Leviathan, John Orbell is Professor, Department of Political Science, 936 PLC Hall, 1415 Kincaid Street, University of Oregon, Eugene, OR 97403- 1284 ( [email protected]). Tomonori Morikawa is Professor, International College, Waseda University, Tokyo, Japan ([email protected]). Jason Hartwig is Instructor, Political Science Department, University of Oregon, Eugene, OR 97403. ([email protected]). James Hanley is Assistant Professor, Political Studies Program, Uni- versity of Illinois, Springfield, IL 62703 ([email protected]). Nicholas Allen is Principal Research Fellow, ORYGEN Research Centre, and Associate Professor, Department of Psychology, Univer- sity of Melbourne, Melbourne, Australia ([email protected]. edu.au). Oleg Smirnov gave us valuable critical and statistical help, and Ron Mitchell provided constructive readings of early drafts. Cian Mont- gomery helped develop an earlier simulation out of which the present one grew. The College of Arts and Sciences at the University of Ore- gon gave critical financial assistance in the early stage of the project, which was subsequently funded by the National Science Foundation Grant 9808041. Morikawa is partly funded by a 2003 Grant-in-Aid for Scientific Research of the Japan Society for the promotion of Science, under the title The Evolution of Political Intelligence and Decision-Making. and decentralized mechanisms might be sufficient—for example, by the existence of “altruistic punishment” (Boyd et al. 2003). Nevertheless, without additional or “selective” (Olson 1965) incentives, rational and self- interested individuals will defect in Prisoner’s Dilemma situations. But that does not mean that real people con- fronting real PD-like situations will always defect. In fact, as Field (2001) has recently emphasized, there is now strong laboratory evidence documenting humans’ frequent willingness to cooperate even in one-shot Prisoner’s Dilemmas where the incentive to defect is substantial and unambiguous (e.g., Caporael et al. 1989; Orbell and Dawes 1993; and Ostrom, Walker, and Gardner 1992). The incidence of such cooperation varies between studies and experimental conditions, but people often do cooperate in prisoner’s dilemmas, sometimes in large numbers. The extensive literature on ultimatum and dictator games, although not ad- dressing cooperation per se, supports the same general conclusion: People are frequently prepared to carry sig- nificant private costs to benefit of others (Camerer and Thaler 1995). Taking such findings seriously has implications for how we address classic problems in political science, notably normative issues surrounding the design of in- stitutions. Certainly, it is foolish to design institutions in such a way that socially desirable outcomes depend on people consistently acting against their private interests (G. Hardin 1977). But “incentive compatibility” and co- operative dispositions might interact. On the positive side, perhaps awareness that “compatible” incentives are in place will increase players’ expectations that oth- ers will cooperate, increasing thereby the likelihood of their own cooperative dispositions being expressed in their own behavior. This might happen, for example, 1

Machiavellian Intelligence

Embed Size (px)

DESCRIPTION

Machiavellian Intelligence

Citation preview

  • American Political Science Review Vol. 98, No. 1 February 2004

    Machiavellian Intelligence as a Basis for the Evolution ofCooperative DispositionsJOHN ORBELL University of OregonTOMONORI MORIKAWA Waseda UniversityJASON HARTWIG University of OregonJAMES HANLEY University of Illinois, SpringfieldNICHOLAS ALLEN University of Melbourne

    How to promote cooperative behavior is classically solved by incentives that lead self-interestedindividuals in socially desirable directions, but by now well-established laboratory results showthat people often do act cooperatively, even at significant cost to themselves. These results suggestthat cooperative dispositions might be an evolved part of human nature. Yet such dispositions appearinconsistent with the Machiavellian intelligence paradigm, which develops the idea that our brainshave evolved, in substantial part, for capturing adaptive advantage from within-group competition.We use simulation to address the evolutionary relationship between basic Machiavellian capacities andcooperative dispositions. Results show that selection on such capacities can (1) permit the spread ofcooperative dispositions even in cooperation-unfriendly worlds and (2) support transitions to populationswith high mean cooperative dispositions. We distinguish between rationality in action and rationalityin designthe adaptive fit between a design attribute of an animal and its environment. The combinationof well-developed Machiavellian intelligence, modest mistrust, and high cooperative dispositions appearsto be a rational design for the brains of highly political animals such as ourselves.

    How to promote group-benefiting action whenindividuals have strong incentives against suchaction is perhaps the fundamental problem ofpolitical theory (Axelrod 1981; Ostrom 1998). Clas-sically spelled out by Hobbes ([1651] 1947) but nowusually framed in terms of the Prisoners Dilemma, theissue is how to persuade individuals to act in the inter-ests of the collectivity (to cooperate) when there is aclear incentive for them to do otherwise (to defect).The incentive structure means that there is only one so-lution for rational and self-interested individuals: Theincentives they confront must, somehow, be changedso that cooperation rather than defection offers thegreater return. We now understand that this does notnecessarily imply a centralized Hobbesian Leviathan,

    John Orbell is Professor, Department of Political Science, 936 PLCHall, 1415 Kincaid Street, University of Oregon, Eugene, OR 97403-1284 ([email protected]).Tomonori Morikawa is Professor, International College, WasedaUniversity, Tokyo, Japan ([email protected]).Jason Hartwig is Instructor, Political Science Department, Universityof Oregon, Eugene, OR 97403. ([email protected]).James Hanley is Assistant Professor, Political Studies Program, Uni-versity of Illinois, Springfield, IL 62703 ([email protected]).Nicholas Allen is Principal Research Fellow, ORYGEN ResearchCentre, and Associate Professor, Department of Psychology, Univer-sity of Melbourne, Melbourne, Australia ([email protected]).

    Oleg Smirnov gave us valuable critical and statistical help, and RonMitchell provided constructive readings of early drafts. Cian Mont-gomery helped develop an earlier simulation out of which the presentone grew. The College of Arts and Sciences at the University of Ore-gon gave critical financial assistance in the early stage of the project,which was subsequently funded by the National Science FoundationGrant 9808041. Morikawa is partly funded by a 2003 Grant-in-Aidfor Scientific Research of the Japan Society for the promotion ofScience, under the title The Evolution of Political Intelligence andDecision-Making.

    and decentralized mechanisms might be sufficientforexample, by the existence of altruistic punishment(Boyd et al. 2003). Nevertheless, without additional orselective (Olson 1965) incentives, rational and self-interested individuals will defect in Prisoners Dilemmasituations.

    But that does not mean that real people con-fronting real PD-like situations will always defect. Infact, as Field (2001) has recently emphasized, there isnow strong laboratory evidence documenting humansfrequent willingness to cooperate even in one-shotPrisoners Dilemmas where the incentive to defectis substantial and unambiguous (e.g., Caporael et al.1989; Orbell and Dawes 1993; and Ostrom, Walker,and Gardner 1992). The incidence of such cooperationvaries between studies and experimental conditions,but people often do cooperate in prisoners dilemmas,sometimes in large numbers. The extensive literatureon ultimatum and dictator games, although not ad-dressing cooperation per se, supports the same generalconclusion: People are frequently prepared to carry sig-nificant private costs to benefit of others (Camerer andThaler 1995).

    Taking such findings seriously has implications forhow we address classic problems in political science,notably normative issues surrounding the design of in-stitutions. Certainly, it is foolish to design institutions insuch a way that socially desirable outcomes depend onpeople consistently acting against their private interests(G. Hardin 1977). But incentive compatibility and co-operative dispositions might interact. On the positiveside, perhaps awareness that compatible incentivesare in place will increase players expectations that oth-ers will cooperate, increasing thereby the likelihood oftheir own cooperative dispositions being expressed intheir own behavior. This might happen, for example,

    1

  • Machiavellian Intelligence February 2004

    when responses are governed by fear of being suckeredrather than by greed for the free riders payoff (Orbellet al. 1986). On the negative side, Frohlich and Oppen-heimer (1996) have shown that incentive compatibilitycan, sometimes at least, undermine socially productivebehavior that would otherwise result from innate co-operative dispositions. Perhaps the social capital rep-resented by such dispositions would be substantiallylost if we were to organize social life exclusively as aresponse to private incentives. In general, the now well-documented existence of innate cooperative disposi-tions means that it is almost certainly an error to baserecommendations about institutional design exclu-sively on incentive compatibilityor on the assump-tion that cooperative dispositions are universal and in-fallible in their production of cooperative behavior.

    We do take those findings seriously and believe, withField (2001), that it is appropriate to address the prob-lem in biologicalthat is to say, evolutionaryterms.Doing so, of course, goes against the standard socialscience model (Tooby and Cosmides 1992) that seeksexplanations of human behavior exclusively in culturalor environmental terms, but drawing a strict dichotomybetween culture and biology is now widely rec-ognized as erroneous (e.g., Boyd and Richerson 1985and Dunbar, Knight, and Power 1999). Humans arecertainly cultural animals, and culture certainly has alot to do with human behavior, including cooperation.But accepting these facts only poses further evolution-ary questions, for example, How and why have hu-mans evolved the capacity for culture? What are theconstraints that evolved human nature places on thesubstance of culture? and What are the constraintsthat evolving culture places on human nature? Grantedthat the issue is human nature, biology is the ultimatesource of understandingand, as Dobzhansky (1973)famously pointed out, Nothing in biology makes senseexcept in the light of evolution.

    Several evolutionary arguments about cooperativedispositions are well known in the life sciences. Thetheory of kin selection (Hamilton 1964) shows how adisposition to assist others with whom one is relatedcan rebound to ones own genetic advantage even ifadaptive costs are involved in the helping act, sug-gesting that our cooperative responses toward nonkinmight be founded, at least in part, on an evolved dis-position to cooperate toward kin. The theory of reci-procity (Axelrod 1984; Trivers 1971) provides an Illscratch your back if you scratch mine explanation forcooperative dispositions that does not require geneticrelatedness, but only that individuals (even membersof different species) encounter each other through aniterated sequence of Prisoners Dilemma-like games.A further model, the one favored by Field (2001), isgroup selection, now staging a comeback (Sober andWilson 1998) after many years of having been widelydiscounted (Dawkins 1976; Williams 1966). Althoughstill hotly disputed (Reeve 2000), group selections ca-pacity to promote cooperative dispositions requiresthat a groups survival prospects be increased by mem-bers cooperative choices, with the cooperators fitnessgains from the groups success being greater than the

    fitness costs that individuals incur as a result of theircooperative choices.1

    Notably absent from this list is any model of howcognitive capacities designed by natural selection foraddressing prior problems might have provided a suf-ficient, even necessary, basis from which cooperativedispositions could subsequently evolve. As is widelyrecognized, natural selection must build from existingstructuresin Dennetts (1995) terms, using cranesrather than skyhooksand any significant coopera-tive dispositions must have evolved in the context ofprior adaptations. In this spirit, the broad question weaddress is, Might cognitive capacities originally designedto address other adaptive problems have made possiblethe evolution of cooperative dispositions such as thosenow strongly suggested by the empirical data?

    In Tinbergens (1963) famous distinction, our in-terest is in ultimate rather than proximate pro-cessesthat is, we address adaptive pressures acrossmany generations as opposed to particular mecha-nisms that might have evolved to capture adaptivegains in the here-and-now. This does not detract, ofcourse, from the interest attaching to proximate mecha-nisms. At the conclusion of his original paper, for ex-ample, Trivers (1971) laid out a formidable researchagenda for psychology by proposing that moralistic ag-gression, gratitude, sympathy, guilt, friendship, gos-sip, cheating, and the ability to detect cheating mighthave evolved as proximate mechanisms for capturingthe gains available from reciprocity. But in evolution-ary theory, ultimate selective pressures and proximatemechanisms are distinct issues, and we do not addressthe latter here.

    Our particular interest is in cognitive mecha-nisms fundamental to the Machiavellian intelligence(Byrne and Whiten 1988; Whiten and Byrne 1997)hypothesis.2 In its broadest terms, this proposes thatgroup living selects strongly for whatever cognitive ca-pacities facilitate an individuals successful negotiationof the competitive and highly complex social environ-ment of the group. A seminal paper by Humphrey(1976; see also Jolly 1966) pointed to the complexity,fluidity, and recursiveness of social relations within pri-mate groups as posing a particularly difficult adaptiveproblem for members of those groupsfar more dif-ficult, Humphrey argued, than the standard adaptiveproblems of gathering resources and avoiding preda-tors:

    Once a society has reached a certain level of complexity,then new internal pressures must arise which act to increaseits complexity still further. For . . . an animals adversariesare members of his own breeding community. If intellectual

    1 We use fitness in the standard (although sometimes debated)evolutionary sense of an individuals relative success in populat-ing subsequent generations with its descendants. Adaptationscomplex structures produced by natural selection in response tochallenges in a species ancestral environment of evolutionary adap-tation(Bowlby 1969)necessarily involve fitness costs as well asfitness benefits, with the presumption being that an adaptation wouldnot have been selected for if the net were not positive.2 Known also as the social (Brothers 1997) and political (Boehm1997) intelligence hypothesis.

    2

  • American Political Science Review Vol. 98, No. 1

    prowess is correlated with social success, and if social suc-cess means high biological fitness, then any heritable traitwhich increases the ability of an individual to outwit his fel-lows will soon spread through the gene pool. And in thesecircumstances there can be no going back: an evolutionaryratchet has been set up, acting like a self-winding watchto increase the general intellectual standing of the species.(311)

    Although there are substantial adaptive advantagesto group livingmost obviously, better protectionfrom predators, improved success as predators, andmore ready access to matesthere is also geneticcompetition between group members, meaning thatthose attributes best equipped to win the never-endingMachiavellian games of social chess or plot andcounter plot will be positively selected.

    Thus social primates are required by the very nature of thesystem they create and maintain to be calculating beings;they must be able to calculate the consequences of theirown behavior, to calculate the likely behavior of others, tocalculate the balance of advantage and lossand this all ina context where the evidence on which their calculationsare based is ephemeral, ambiguous and liable to change,not least as a consequence of their own actions. (309)

    By this hypothesis, evolutionary competition be-tween group members selects for capacities that allowindividuals to deceive and exploit other group mem-bers, while simultaneously avoiding being deceived andexploited by others. Employing the terms introducedby Dawkins and Krebs (1978), the two fundamentalMachiavellian capacities are (1) Senders capacity topersuade another group member to accept as true whatit is in Senders interest to have it believe is trueviz,manipulation3and (2) Receivers capacity to pene-trate to the truth underlying messages from potentiallymanipulative othersviz, mindreading. More specifi-cally, therefore, our research question is, Could evolvedMachiavellian capacities for manipulation and min-dreading be a basis for the evolution of cooperative dis-positions among social animals such as ourselves?

    This has the air of paradox. On the one hand, theMachiavellian intelligence hypothesis seems to dove-tail quite nicely with the harder models of rationalaction, implying a brain designed to facilitate the pur-suit of private welfare by whatever means necessary.On the other, cooperative behavior involves, by def-inition, rejecting a dominant incentive, an alternativethat is superior for the acting individual regardless ofwhat another individual might choose. If our brains aredesigned for Machiavellian purposes, should we notexpect evolved dispositions to defectat least when wecan get away with itnot to cooperate?

    We use evolutionary simulation to show how, un-der specified parameters, it is not only possible thatMachiavellian capacities provide an evolutionary foun-dation for cooperative dispositions, but in fact quite

    3 As Trivers (1985) has commented, One of the most importantthings to realize about systems of animal communication is that theyare not systems for the dissemination of the truth. An animal se-lected to signal to another animal may be selected to convey correctinformation, misinformation, or both (395).

    likely. If we are, by nature, Machiavellian animals ben-efiting from group living but also exploiting other groupmembers when that is possible, our findings show howMachiavellian capacities could be a critical underpin-ning for evolutionary selection on cooperative dispo-sitions capable of motivating us, perhaps quite often,to act cooperativelythus helping to resolve collec-tive action problems absent appropriate selective in-centives.

    DESIGN OF THE SIMULATIONWe use simulation for two reasons. First and most ob-viously, the processes that interest us occurred over aperiod of perhaps hundreds of thousands of years and,because they involve social relationships, could leaveonly little by way of a directly observable fossil record(Wynn 2002). Although careful inference from the ex-isting fossil record can help us understand when partic-ular cognitive structures evolved (e.g., Mithen 1996),such work is necessarily informed by plausible modelsof process, and simulation is one way of constructingsuch models. Second, simulation can be a tool for dis-covery, letting the researcher explore the consequencesof diverse parameter settings and, perhaps, identifytheoretically interesting processes that might not havebeen considered using standard analytic techniques.This has been our approachmeaning that a first orderof business here is to specify the design decisions madein constructing the simulation.

    A simulation for studying the relationship betweenMachiavellian capacities and cooperative dispositionsrequires modeling decisions at several levels of analysis:

    Individual attributes: We incorporate two disposi-tions and two cognitive (information processing) capac-ities. The dispositions are, first, to act in a cooperativemanner within joined Prisoners Dilemma games4 and,second, to mistrust others willingness to do so (thusits converse, to trust them). The cognitive capacitiesare, first, the capacity to manipulate the persuasivenessof what the individual communicatestrue or falseto others and, second, the capacity to mindread thetruth underlying others efforts at such manipulation(Dawkins and Krebs 1978). At the ultimate or func-tional level at which we operate, of course, we cansidestep the always fascinating proximate questionsabout particular mechanisms facilitating such manip-ulation and mindreading.

    Interindividual communication: Because we are in-terested in the evolution of cooperation, when two

    4 Although much recent theoretical and empirical work has focusedon cooperation within multiple-person Prisoners Dilemma games,the problem of predicting others behavior in such games is com-putationally formidable, perhaps requiring simplifying heuristics ofone kind or another (Orbell and Dawes 1993), whereas the problemof choice and incentives remains basically the same in two-personand n-person PDs (although see Dawes 1975). Accordingly, we focusexclusively on the two-person PD heredefined, as usual, by a gamewith (1) a binary choice between cooperation and defection;(2) the standard payoff notation being t = lone defection or freeriding, c = mutual cooperation, d = mutual defection, and s = lonecooperation or being suckered; and (3) t > c > d > s and 2c > (t + s)> 2d.

    3

  • Machiavellian Intelligence February 2004

    individuals encounter each other they must choose be-tween entering and not entering a potentially cooper-ative game. In this context, Machiavellian intelligencerequires that each sends messages to the other abouthis or her intentions within any such game (subject tomanipulation) and that each evaluates the truth-valueof such messages received from the other (a task per-formed by mindreading).

    Individual decision-making: Individuals must makesocial decisionswhether to enter a particular po-tentially cooperative game and, if they do, whether tocooperate or defectand those decisions must be in-formed, at least in part, by their estimates of what theother individual is likely to do. Importantly, the optionof not entering such a game implies the availability ofan alternative course of action that has some value, thusthat the individuals play vs. not play decision will bebased on its comparison between what it expects to getfrom playing a cooperative game and what it expectsfrom such an alternative. There are, of course, many dif-ferent games that social animals playRapoport et al.(1976) identified 78 logically distinct two-by-two games,and many of those have multiple-person variantsbutmodeling a social ecology with all that complexity, ifindeed possible, would detract from our primary goalof understanding how the evolution of Machiavelliancognitive capacities and cooperative dispositions mightbe related, although, particularly interesting alterna-tives are (1) making a hawk challenge, indicating awillingness to fight over some resources otherwise incontrol of another individual (Orbell, Morikawa, andAllen 2002), and (2) acting in a solitary, go it alonemanner, extracting resources from the environment ab-sent any interaction with ones own speciesan en-tirely nonsocial course of action.

    Natural selection: In an evolutionary model theremust be a process by which individuals actions duringtheir lifetimes are translated into relative reproductivesuccess, thus that allows for selective retention in thepopulation of whatever attributes produced that suc-cess (Dennett 1995). As is well recognized, populationattributes can change by drift as well as by relativereproductive success; in fact, we will show that driftmight have played an important role in the evolutionof cooperative dispositions. In either case, however,evolution requires a source of variation that is ran-dom with respect to the individuals success but thatmakes it possible for some individuals to be reproduc-tively more successful than others. We recognize thatvariation in the natural world can be produced by, forexample, recombination as well as by mutation, but forsimplicity in what follows we use the term mutationto include all sources of variation. Critically for ourpurposes, mutations must happen on the dispositionaland cognitive attributes that bear on an individualssuccess in complex social environments.

    Dispositional and Cognitive AttributesWe model an individuals disposition to cooperate as aprobability between 0.0 and 1.0 (its probability of coop-erating, or PC). Each decision it makes is determined

    by a random draw from its PC; it will cooperate if thedraw falls at or below its PC and defect if it falls abovethat value. Notice that such decisions are not appropri-ately thought of as rational. Granted, an individualwith a low PC might be classified as more rationalthan one with a higher such value; cooperation is, afterall, dominated. But that has no particular bearing onour evolutionary model in which such an individual issimply undisposed to cooperate. If such dispositionsare adaptive, then they will be positively selected andwill spread throughout the population. If they are not,then they will be selected against and, probably, willvanishrational or not.

    We assume that every agent, on encountering an-other, makes the claim I will always cooperate as itmakes no strategic sensecertainly in Machiavelliantermsto claim anything less than perfect coopera-tiveness. The total communication is modeled as 100messages, some of which (PC 100) are true (reflectingthe true probability of cooperating) and the remainderof which ({1 PC} 100) are false (reflecting the trueprobability of noncooperation). But the sender of anysuch communication confronts the discrete manipula-tive problems of being convincing in both its true and itsfalse messages. Each individual is therefore equippedwith separate manipulative capacities for its true mes-sages (Manipulate True) and for its false ones (Manip-ulate False), in each case varying between 0.0 and 1.0.Agents in our simulation might, therefore, be goodtruth-tellers and good liars, good truth-tellers but poorliarsand so on. Particular truths and particular lies arenot all equally persuasive but are normally distributedaround the agents mean manipulative capacity in eachcase.

    Correspondingly, individuals are modeled as havinga mindreading capacity that varies between 0.0 and 1.0.At the low extreme of mindreading, an individual hasno capacity to penetrate the truth of a potential part-ners I will always cooperate claim. While at the highextreme it has a perfect capacity to do so, being (poten-tially) able to recognize all Senders lies as lies and all itstruths as truths. Notice that we do not model mindread-ing capacities specialized separately for responding totruths and lies. That would imply a cognitive apparatusequipped with a priori knowledge of which is which,thus that the problem of mindreading has already beensolved.5

    Trust is a concept that has attracted much inter-est in recent years (e.g., Fukuyama 1995 and Nesse2001), and it plays an important role in the simulation.As specified above, mindreading lets agents addresswhether or not they should trust a particular other toact on its I will always cooperate claim (cf. R. Hardin

    5 Of course, an individual could be better at (for example) detectingtrue statements than at detecting lies, and as our results will show,populations can evolve such differential abilities. Similarly, individ-uals might or might not be aware that they have such differentialabilities, if they do. But our present modeling concern is with thestructure of cognitive architecture, and a mechanism designed formindreadingfor detecting whether or not another individual istelling the truthcannot be constructed on the assumption that itknows the answer to that question in the particular case.

    4

  • American Political Science Review Vol. 98, No. 1

    1991 and Orbell, Dawes, and Schwartz-Shea 1994). Buta distinct issuerelated to the concept of social cap-ital (Putnam 2000)concerns trust as a generalizeddisposition toward others, a willingness to accept oth-ers I will always cooperate messages independentof expectations that mindreading might have producedabout particular individuals, and we have incorporatedthis into the simulation. We model generalized mis-trust as a value between 0.0 and 1.0, with individualsat the low end of that distribution being disposed toaccept most or all of another individuals messageswhether true or false, and operationally independentof what their mindreading tells them they should ac-cept from this particular individualand individu-als at the high end being disposed to reject most orall such messages. Only messages whose persuasive-ness, modified by Senders manipulation and Receiversmindreading, places them above Receivers mistrustthreshold are believed by Receiver and incorporatedinto its decision-making.

    Individual Decision-MakingAn encounter between two individuals obliges each todecide between entering a PD game with the othervs. taking the alternative, non-PD course of action(which we call ALT). By assumption, individualsknow their own cooperate vs. defect choice, and theyalso know the payoffs available from a joined PD gameand from ALT.6 The only additional information theyneed before making a decision, therefore, is the prob-ability with which the other individual will cooperatein a joined PD game. A decision-makers estimate ofthis will be a function of (1) the other individualsactual PC, (2) the other individuals capacity to ma-nipulate both its true and its false messages, (3) thedecision-makers own mindreading capacity, and (4)the decision-makers own level of generalized mistrust.Figure 1 uses illustrative values to show how the PD vs.ALT decision is made.

    Sender is attempting to manipulate Receiver intobelieving its 100 I will always cooperate messages(although, of course, both individuals play both rolessimultaneously). In this example, Senders PC is .75,meaning that it is sending 75 true messages and 25lies, with each set of messages distributed (normally)around Senders manipulative capacities for true andfalse messages, respectively; in the example, those ca-pacities are .40 for true messages and .30 for false ones,making our example Sender somewhat better at per-suading with respect to truths than to lies. The bestoutcome for Receiver would be to reject all Senders

    6 In natural circumstances there is likely to be some doubt aboutall these values, and an adaptive problem in our ancestral pastas nowis to evaluate accurately just what is at stake in variousgames, as well as what particular game is a possibility here. Indeed,it seems likely that humans and other social animals have evolvedspecial-purpose, domain-specific cognitive mechanisms for facilitat-ing accurate such evaluations (Gigerenzer 1997). Given our interestin the relationship between Machiavellian intelligence and coopera-tive dispositions, however, we can make the simplifying assumptionof perfect knowledge in these respects.

    FIGURE 1. An Example of How Cognitive andDispositional Attributes Interact in ReceiversDecision-Making

    Note: Although this Senders 75 true statements are somewhatmore persuasive than its 25 false ones (.40, in comparison with.30), with no mindreading being exercised no message in ei-ther its true or its false sets falls above Receivers .60 mistrustthreshold. Receiver would, then, assess Senders PC as zero.After mindreading is exercised, mostalthough not allof thisSenders true statements fall above Receivers Mistrust thresh-old. Those that do are used to define Receivers estimate ofSenders PC; in this case, that estimate is somewhat less thanthe true .75. Of course, each of these attributes is likely to varyamong individuals, with varying consequences for the accuracyof such estimates.

    lies while accepting all its truths, but this receivers mis-trust threshold is set at .60, meaning that it will onlyaccept messages that are above that threshold (absentmindreading, in the example, none). Absent mindread-ing, therefore, this receiver would underestimate thissenders PC, believing it to be zero rather than the truevalue of .75.

    In general, any positive mindreading capacity on Re-ceivers part will lead it to read Senders true messagesas more believable than otherwise (graphically, fartherto the right) and to read its lies as less believable thanotherwise (farther to the left). In the Figure 1 example,Receivers mindreading capacity of .5 leads it to readSenders 75 true messages as having a mean believabil-ity of .7 (.5 of the distance between their manipulatedpersuasiveness of .4 and being fully believable at 1.0)and will lead it to read Senders 25 lies as .15 believable(.5 of the distance between their manipulated persua-siveness of .3 and being fully unbelievable at 0.0). A per-fect mindreader would have done better still, readingall Senders true messages as 1.0 believable and all of itslies as 0.0 believable. In the example, Receivers level ofgeneralized mistrust is set at .60, modestly high but not

    5

  • Machiavellian Intelligence February 2004

    completely mistrustful. After Receiver has exercised its(also modest) mindreading capacity, its mistrust thresh-old is low enough that it accepts as true perhaps 60of Senders 75 true messages, thus estimating SendersPC as .6. Of course, any combination of the variousattributes is possible.

    The bottom line, however, is the decision that Re-ceiver makes using the information that, for better orworse, it has now gathered. Using standard PrisonersDilemma notation (footnote 4) and defining estPC asthe number of Senders 100 mindreading-adjusted mes-sages falling above Receivers own mistrust threshold,if Receiver draws cooperation from its PC it will entera PD when

    estPC(c) + (1 estPC)(s) > ALT,but otherwise will choose ALT. And should it drawdefection, it will enter a PD when

    estPC(t) + (1 estPC)(d) > ALT,but otherwise will choose ALT. In the particular case,of course, Receivers choice will depend importantlyon the PD payoffs and the value of ALT.

    Whether or not Receivers decision will be profitable(adaptive) depends on the quality of the information ithas gathered in this encounter and on the decision thatthe other individualin its role as receivermakes atthe same time. Notice, however, that Receivers deci-sion is modeled as a rational one, granted the quiteprobably imperfect information (Simon 1985) it hasgathered about Senders intentions. Within an evolu-tionary frame where payoffs are in units of fitness, thisis entirely appropriate since rationality equates withfitness maximization, and individuals whose choicesare not geared in that direction will not last long. Butis it inconsistent to model individuals decisions be-tween entering PDs and playing ALT as rational butto model the decisions of those same individuals withinPD games as produced by dispositions with no basisin rationality?

    It is notgranted that our concern is with the con-sequences of natural selection on such dispositions. Asemphasized above, within an evolutionary framework,the importance of such dispositions is whether or notthey contribute to an individuals adaptive success inthe context of decisions about whether or not to playparticular games. Rationality generally assumes that in-dividuals have no innate dispositions, toward either co-operation or defection, basing their decisions entirelyon recognizable gains and losses. But whether cooper-ative dispositions will grow and prosper or be selectedagainst and vanish within a Machiavellian world is pre-cisely what we are investigating here. We will return tothe issue of rationality insofar as it bears on evolution-ary modeling later in the paper.

    SelectionIn the simulation, members of a population encountereach other during a generation, with the gains or lossesfrom those encounters functioning as units of evolu-

    tionary fitness. All the members of a generation dieafter the specified number of encounters has been com-pleted, but the more successful among them reproduce,with their offspring carrying their various cognitive anddispositional attributes into the next generation, sub-ject to mutation. Specifically, individuals whose wealthfalls below zero at the end of their generation die with-out reproducing, those whose wealth is positive butbelow twice the median reproduce once, those whosewealth is between twice the median and below threetimes the median reproduce twiceand so on. Carry-ing capacity of the ecology is limited; should the num-ber of offspring exceed that capacity, a random lotterydetermines which particular agents populate the nextgeneration.7 Generations can involve any number ofencounters, but in the simulations we report below eachagent encounters each other agent twice in their partic-ular generation, once as Alpha, who chooses whetheror not to offer a PD game, and once as Beta, whosechoice is limited to accepting vs. not accepting a PDoffer, should one be made. If Alpha chooses not tooffer a PD, or if Beta rejects such an offer, both makeALT.

    On reproduction, mutation on all a parents disposi-tional and cognitive attributesPC, mindreading, mis-trust, Manipulate True and Manipulate Falseis a pos-sibility. Because all those attributes are modeled asprobabilities, mutation of fixed magnitude could eas-ily have reached a limiting point beyond which furthermutation in a particular direction would not be pos-sible. Our solution was to define each attribute fromtwo constituent integers above zero, one facilitatingand one inhibiting the attribute in question. Thus, forexample, PC is constructed from the integers cooper-ate plus (a positive disposition toward cooperation)and cooperate minus (a negative such disposition),with the individuals actual PC being defined as theproportion

    cooperate plus/(cooperate plus + cooperate minus).Similarly, mistrust is constructed from integers defin-ing a disposition not to believe that others have coop-erative intentions (skepticism) and a disposition tobelieve that they have such intentions (credulity),with the attribute itself being defined as the proportion

    skepticism/(skepticism + credulity).A similar approach is taken with respect to mindread-ing and the two manipulative capacities. Mutation oneach of the integers from which a parents attributesare constructed can thus affect the values of attributesan offspring inherits, with potential consequencesneutral, adaptive, or maladaptivefor that offspringssuccess within its own generations social environment.

    7 Carrying capacity is thus a constraint preventing population sizefrom increasing indefinitelyas could happen if natural selection onsocial success were the only limiting factor. The random lottery is,thus, only conducted among individuals who are already successfulin terms of natural selection. In natural circumstances, of course,immigration to unpopulated areas might have been possible for manyof our early ancestors.

    6

  • American Political Science Review Vol. 98, No. 1

    Mutation happens on such constituent attributeswith the parameters probability and magnitude.That is, there is some fixed probability with which anygiven parental attribute will mutate on being transmit-ted to an offspring, and some fixed magnitude (positiveor negative) of those mutations that do happen. In thesimulations to be reported, probability is set at 0.05 andmagnitude at 5.0 meaning that either of the two inte-gers from which each of an offsprings several attributesis constructed could differ from those of its parent witha probability of 0.05, and by a magnitude of plus or mi-nus 5.0.8 Note that although the number of offspring isinfluenced by a parents relative success in its own gen-eration, the only source of change in attributes passedfrom a successful parent to its offspring is mutation thusspecified. The model does not, therefore, involve anyparent-to-offspring learning and, certainly, no Lamar-ckian inheritance.

    Experimental DesignIn these terms, the objective is to observe howpopulation-level values of the several individual at-tributes change through successive generations and theselective processes by which such changes happen. Inparticular, we are interested in the possibility that se-lection on manipulation and mindreading influences se-lection on cooperative dispositionsand, of course, thefrequency of whatever cooperative behavior follows.

    We start each simulation with a population ofindividuals whose cognitive and dispositional at-tributes militate strongly against PD games beingplayedspecifically, PC = .10, Mistrust = .90, Ma-nipulateTrue = .10; ManipulateFalse = .10, and Min-dread = .10. At least until mutation has had a chanceto change population attributes, agents will send eachother I will always cooperate messages that aremostly lies; they will have little capacity to be persua-sive (either in their many lies or in their few truths);they will have little capacity to penetrate to the truth ofparticular messages; and they will be so generally mis-trustful that even very persuasive claims will seldom beaccepted. Agents in this initial generation are, in fact,substantially nonsocial insofar as they lack the ability toengage in collective action, and they are also substan-tially nonpolitical insofar as their Machiavellian capac-ities are underdeveloped. Any cooperative equilibriathat might emerge from this (perhaps Hobbesian)world will have to be explained, but so will the processby which any initial leap out of this world happens.

    FINDINGSOur initial exploratory runs suggested two possibilities.(1) Selection on Machiavellian intelligence can both

    8 Given that there are 10 components on which mutation mighthappen, these parameters mean that there is a .599 probability ofany given offspring having no mutation on any of the componentsit inherits from its parent. Sensitivity testing shows that increasingor decreasing these parameters does have predictable effects on thespeed with which mean population attributes change but not on thebasic pattern of findings we report.

    produce and sustain transitions from such a coopera-tion and PD unfriendly founding population to laterpopulations with very high mean PC values and veryfrequent mutually cooperative PD games; (2) Such co-operative transitions happen almost exclusively underthe parameters 0 ALT c. Our subsequent analysisfocused more systematically on exploring these possibi-lities and on identifying the processes underlying them.

    Our approach was to run multiple simulationswith PD parameters constant (t = 15; c = 5; d = 5;s = 15) but varying ALT. We first confirmed theabsencemore accurately, very low incidenceof co-operative transitions outside the parameter range 0 ALT c. In retrospect, the reasons for that pattern areclear. When the payoff from ALT is higher than frommutual cooperation, only an individual who intendsdefection and is convinced that a potential partner willcooperate will rationally enter a PD game in preferenceto ALT. Although one competent mindreader mightmake that assessment, both parties must agree beforea PD is joined, and the probability of both deciding thesame thing about the other, though not zero, is slimand becomes still more so as ALT approaches t .

    At the opposite extreme, when ALT 0, even a lowestimate of the others PC might return an expectedvalue for entering a PD that is higher than ALT, andbecause cooperation is dominated by defection, suchan estimate will be particularly likely for intendingdefectorsprobabilistically, individuals with low PCvalues. Estimating the others PC at 0.4, for example,an intending defector would calculate the EV of a PDas 3, whereas an intending cooperator would calculateit at 7. Trapped between two alternatives both offer-ing a loss, the intending cooperatorprobabilistically,an individual with a high PC valuecannot prosper,meaning negative selection on high PC. But a popula-tion of low PC agents produces, for the most part, eitherPDs played to mutual defection (a negative) or ALT(also a negative in this range) and will rapidly die out.9

    With this understanding, we ran 90 simulations, eachconsisting of 20,000 generations, each with the samePD payoffs but including 10 at each 0.5 interval acrossthe range 0 < ALT < 5. The results are reported inTable 1. First, the data confirm that transitions fromoverwhelmingly ALT choices to overwhelmingly PDgames played with mutual cooperation are very fre-quent within that parameter range. Although there areslightly fewer transitions toward the lower end of thatrange, across that whole range transitions fail to happenin only 9 of the 90 simulation runs. Within each of thenine ALT categories, there is considerable variation asto when cooperative transitions begin, but such tran-sitions normally do take several thousand generationsto get under way (column 3). Nevertheless, and despiteoccasional lapses that are reversed with the passage of

    9 With different PD parameters the population might not die out; ifmutual defection were a positive, for example, a population of verylow PC types could survive indefinitely even with a negative ALT. Theimportant point is that, with a negative ALT, cooperative dispositionswill be strongly selected against, and cooperative behavior will beextremely rareat best.

    7

  • Machiavellian Intelligence February 2004

    TABLE 1. Cooperative Transitions within the Parameter Range 0 < ALT < cALT Number of Cooperative Mean Generation Where Mean PC Value After Predicted

    Value Transitions Transition Starts the Transition Threshold PC(1) (2) (3) (4) (5)4.5 9/10 6535 0.981 0.9754.0 10/10 5923 0.965 0.9503.5 10/10 5983 0.940 0.9253.0 9/10 5924 0.926 0.9002.5 9/10 6046 0.898 0.8752.0 10/10 4474 0.897 0.8501.5 8/10 3398 0.858 0.8251.0 8/10 4040 0.833 0.8000.5 8/10 5566 0.815 0.775

    Note: Based on 10 simulations for each value of ALT.

    time, the 81 cooperative transitions all produced quitehigh mean PC values (column 4) that persist in appar-ent equilibrium until the simulations end at the twentythousandth generation.10

    Further, those high PC values are reflected in actualbehavior. We illustrate this with an example when ALTwas set at 4. Figures 2a, 2b, and 2c show, respectively,mean levels of PC across the 20,000 generations of thissimulation; the incidence of ALT outcomes vs. joinedPD games; and, among the PD games that were joined,the incidence of mutual defection, mutual cooperation,and the outcome in which one defecting individual ex-ploits the other. From Figure 2a, the transition to highlevels of PC begins at generation 5,319 after what ap-pear to be a couple of false starts butafter a thou-sand or so generations of some fluctuationhas settleddown to a basically stable state around PC = .97 withonly minor fluctuations from generation to generation.After the transition, in other words, this population isvery disposed to cooperation, granted not perfectly so.

    From Figure 2b, we see that these cooperative dis-positions are paralleled by a willingness to enter PDgames and a corresponding decline in the incidence ofALT outcomes.11 There is some instability from gener-ation to generation and one occasion (around genera-tion 15,000) when the preponderance of PD games inthe ecology is temporarily challenged by ALT, but thereversal in the overall pattern of game choices after thetransition is clear.

    Finally, from Figure 2c we see that the PD gamesthat dominate relationships among individuals after thetransition have, overwhelmingly, mutually cooperativeoutcomes. The one fluctuation that does occur is as-sociated with the incidence of PD games, not with anymajor fluctuation of within-game behavior, and mutualcooperation remains the characteristic outcome to PD

    10 More extensive testing showed that the general pattern of cooper-ative transitions happening only within the specified parameter rangewas robustwhen, for example, the initial world was cooperation-friendly and across wide variation in the absolute values of PDpayoffs.11 The lack of a perfect negative match between PD and ALT out-comes is explained by there being two ways the latter outcome canhappenwhen Alpha (the first mover) rejects the PD option andwhen Beta (the second mover) rejects a PD offer from Alpha. PDgames, on the other hand, require both parties to agree.

    games throughout this period. Importantly, notice thatthe trickle of PD games that we observe after aboutgeneration 1,000 (Figure 2b) is identified here as beinggames played almost exclusively to mutual defection. Infact, only two mutually cooperative games were playedduring the 5,318 generations before the cooperativetransition got under way. We will return to the signifi-cance of this trickle of mutually defecting PD gamesshortly. In general, however, these data show that therapid change in cooperative dispositions is associatedwith a corresponding change in behavior from over-whelmingly ALT choices to overwhelmingly joined PDgames and, within those games, frequent cooperation.This granted, the problem is to explain (a) the originof such cooperative transitions and (b) the processesby which they are maintained at what appears to beequilibrium.

    OriginsTo find answers, we examine mean levels of the threeMachiavellian attributes (ManipulateTrue, Manipu-lateFalse, Mindread), the generalized disposition tomistrust, and cooperative dispositions, as those variedwith PD games played. Although population changesare the ultimate concern of evolutionary analyses, ag-gregate data do not speak directly to the selective pres-sures on individuals. Therefore, we will also report cor-relational data at the individual level during adaptivelyimportant periods.

    Of greatest initial interest, Figure 3 shows, for thecase being followed, a substantial step-level, upwardchange in mean population mindreading values ataround generation 2,000, a change that is plausibly aresponse to the costs of becoming involved with PDgames when the mean PC is very low, thus the proba-bility of taking a negative from such a game is highwhether as a (rare) cooperator or as a (more frequent)defector. A strong test of that hypothesis is provided byobserving the relationship between individuals min-dreading scores (independent variable) and the pro-portion of PD games that, if offered to an individ-ual, were accepted, across the 100 generations between1,990 and 2,090 when the increase in mindreading wasmost marked and the variance in mindreading was also

    8

  • American Political Science Review Vol. 98, No. 1

    FIGURE 2. A Cooperative Transition When ALT == 4Mean PC and Behavioral Outcomes: (a) MeanLevels of PC, (b) ALT vs. PD Outcomes to Encounters, and (c) Outcomes to Joined PD games

    highest.12 The result was = 0.1593 (p < .001), sup-porting the hypothesis that selective pressures in thisnasty period were favoring individuals with mindread-ing capacities sufficient to keep them from acceptinginvitations to enter PD gamesthat, if entered, wouldhave resulted in loss.

    12 Data points are each of the 50 individuals in each of the 100 gen-erations, for an n of 5,000.

    Notice that positive selection on PC values in thisearly, cooperation-unfriendly stage is not likely becauseso few PD games are being playedwith those thatare played resulting in costly mutual defection. Butwhat happens to individuals whose PC does happento mutate upward? We can predict, first, that such co-operative mutants will be spotted as potential victimsby mindreading-equipped low-PC types, and the datasupport this: In the case being followed, for example,

    9

  • Machiavellian Intelligence February 2004

    FIGURE 2Continued.

    the relationship between agents own PC values (in-dependent variable) and the number of PD gamesoffered to them by others for the 100 generations im-mediately prior to the start of the cooperative transi-tion was = 36.3824 (p < .0001). The same mindread-

    FIGURE 3. A Cooperative Transition When ALT == 4Mindreading

    ing capacity that evolved to address the prophylacticfunction of keeping agents out of costly PD games ina cooperation-unfriendly world also equips them torecognize and try to exploit occasional high PC types.But mindreading cuts two ways, and Table 2 reports

    10

  • American Political Science Review Vol. 98, No. 1

    TABLE 2. Multiple Regression Beta Weights:Proportion of Offered PD Games Accepted byMistrust, Mindreading, and PC

    CoefficientIntercept 1.2216Mistrust 0.6821Mindreading 0.9414Probability of cooperating 0.0641Note: One hundred generations prior to the cooperative transi-tion. R2 = .557; P < .001; N = 4,978. Observations are on 50agents in each of the 100 generations; missing cases had zerooffers.

    coefficients from a multiple regression predicting, fromthe PC, mindreading, and mistrust scores of such tar-get agents (independent variables), the proportion ofPD offers that, once made, were accepted by agents.Clearly, the mindreading capacities of such occasionalcooperative mutants makes it possible for them torecognize and avoid such trapsthus, for their rela-tively higher cooperative dispositions to be passed onto any offspring they might have.

    Granted, sustained upward drift of PC values will re-quire low-probability upward mutations on the PC val-ues of at least several members of a lineage, somethingthat is not likely for any particular lineage. But min-dreading does make such cooperative drift at leastpossible and, with the passage of many generations,quite likely on at least some lineages. And this is whatsets the stage for the observed cooperative transitions.

    We have followed the individual-by-individual de-tails across several different values of ALT andwithdue allowance for the variation resulting from the manystochastic elements in the simulationthe basic pat-tern is always the same. At some point, two relativelycooperative agents, both equipped with high mindread-ing, each recognize the other as a good bet for a PDgame, and with one offering and the other accepting,both capture the mutual cooperate payoff. With otheragents all rejecting PD games in favor of ALT (whereALT < c), these two enjoy relative reproductive suc-cess. Their offspring, inheriting the high PC and highmindreading that made their parents successful, havesimilar success in their own generation until, quiterapidly, the whole population is descended from thisoriginal pair.

    Because this process relies on randomly drifting PCvalues, such cooperative transitions are not inevitablewithin this parameter range. Quite possibly, no twoindividuals will ever reach high enough PC levels tobe seen as attractive partners by each other. In fact,as we have pointed out, transitions failed in nine ofthe 90 simulation runs we conducted. By the sametoken, some transitions might take many more gen-erations to happen than others. But the process wehave described makes cooperative transitions likely tohappensooner or later.

    Returning to Table 1, it is apparent that the level atwhich PC equilibrates varies monotonically with ALT;as ALT becomes higher within the critical parame-ter range, mean PC values after the transition also

    become higher. Because natural selection can be as-sumed to discover attribute levels that are optimalfor particular environments, it appears that optimal lev-els of cooperative dispositions are positively related toALT. This is perhaps surprising, as one might expectcooperative dispositions to be higher as the relativeadvantage of mutual cooperation over other ways ofmaking a living becomes greater, but just the oppositehappens.

    The issue is resolved by recognizing that two crite-ria must be met in order to maximize returns in thisworld. First, an agent must persuade another to enter aPD game rather than to play ALT; and, second, underthat constraint, it must maximize its own returns withinany PD game that might be joined. Given the potentialvictims (now) probably high mindreading capacities,meeting the first criterion requires the agent to havegenuinely high cooperative dispositions; when ALT ishigh, a potential victim can afford to be fussy aboutthe cooperative dispositions of would-be victimizers.But the would-be victimizers cooperative dispositionsshould be no higher than minimally necessary to getthe potential victim into play. Being any more coop-eratively disposed will result in lost opportunities forprofitable free riding, placing the agent at an evolution-ary disadvantage relative to others who do manage toenter PD games with cooperative individuals but whodo, nevertheless, defect.13

    This analysis suggests: Natural selection will dis-cover levels of cooperative dispositions that are simulta-neously high enough to ensure that prospective partnersassess the expected value of entering PD games as greaterthan ALT, but low enough to maximize the possibility ofexploiting partners should a PD game be joined. For apopulation of perfect mindreaders, this logic predicts athreshold PC for each value of ALT (just) above whicha populations PC values will equilibratespecificallywhere PC(c) + (1 PC)(s) = ALT. In these terms, col-umn 5 in Table 1 specifies such thresholds for the ninevalues of ALT that we have examined across the param-eter range 0 < ALT < c, and Column 4 shows that, aspredicted, the mean levels at which cooperative disposi-tions equilibrate after transitions are (1) monotonicallyrelated to values of ALT within this range and (2) onlyslightly above the threshold in each casethe slight er-ror in each case being, presumably, a consequence ofminor mistakes in mindreading that still do happen.14

    13 Of course, in this situation two individuals encountering each otherare both would-be victimizers and potential victims.14 Although transitions do frequently happen across all values ofALT within this parameter range, case-by-case and generation-by-generation micro analyses of the process by which they happen showthat there are many more false starts at the lower values of ALTwhere several lineages normally appear about to dominate the entirepopulation but die out before doing so. Whereas one normally does,sooner or later, do so, the reason for such false starts is the permissive-ness that low values of ALT provide for relatively defection-inclinedindividuals who are more likely to attempt exploitation of each otherthan their (necessarily) more cooperative counterparts when ALT ishigh. Correspondingly, the permissiveness provided for defection atvery low values of ALT makes cooperative transitions, when they dohappen, more unstable than at higher such values, in a very smallnumber of cases allowing them to collapse altogether.

    11

  • Machiavellian Intelligence February 2004

    FIGURE 4. A Cooperative Transition When ALT == 4; Mean Mistrust and MeanMindreading-Adjusted Truths and Lies

    Maintenance

    What, then, sustains such cooperative transitions oncethey have happened? From Figure 4, we observe a rapiddrop in mistrust as the cooperative transition is gettingunder way (generation 5,319). This is explained in termsof competition among individuals in the newly cooper-ative population. In this nice world, advantage will goto those who enter more PD games than their adaptivecompetitors, and with mindreading already high, thismeans selective advantage to those who have relativelylow levels of mistrust. The result is that, for a few thou-sand generations, the posttransition population is notonly highly disposed to cooperate, but also very low inmistrust.

    But such trusting dispositions in combination withhigh PC values make this population doubly ripe forexploitation by low PC types, thus for an eventual re-version to generally low PC values. The reason whysuch a reversion does not happen involves the relation-ship between mean levels of mistrust and mindreading.First, from Figure 4 we can see that the drop to verylow levels of mistrust only lasts a few thousand gen-erations after the transition has happened. Althoughthat drop did facilitate individuals entering now gen-erally cooperative relationships, it also opened the doorfor relatively less cooperative individuals to prosperand spread, with that, in turn, selecting for individualswith somewhat higher mistrust, as reflected in the meanvalues in Figure 4. The adaptive problem is to find anoptimal level of mistrust, given widespread mindread-ing capacities but also the substantial gains that are

    now available from entering PD games in this generallycooperative world.

    The solution discovered by natural selection is abalance between mistrust and mindreading, with mis-trust high enough to ensure that lies are rejected butlow enough not to endanger acceptance of the true mes-sages that predominate in this posttransition, generallycooperative world. Figure 4 illustrates how this hap-pens using the case we have been following. Grantedthe elevated mindreading that is now normal, receiversaccept senders true messages with a higher probabil-ity than the persuasiveness of those messages wouldimply, but they also accept Senders lies with a lowerprobability than the persuasiveness of those messageswould imply. Critically, mean levels of mistrust evolveupward to a point that is high enough to ensure thatSenders lies are rejected, thus precluding invasion bylower PC types. There are two brief periods where themean mindreading-adjusted persuasiveness of liesmoves above mean mistrust, and the consequences ofthat are visible in the brief drop in the incidence ofcooperatively played PD games (Figures 2b and 2c).But in both cases this drop is only temporary and thecooperative equilibrium returns.

    The reason for this equilibrium is that the most adap-tive configuration of cognitive and dispositional at-tributes is one that will maintain frequent access to PDgames played with cooperators. But in general, advan-tage will go to those whose mistrust is high enough, inconjunction with their mindreading capacity, to ensurerejection of most, if not all, lies that are directed towardthem, while being low enough to ensure acceptance of

    12

  • American Political Science Review Vol. 98, No. 1

    most, if not all, the truths that are also directed towardthem. Those who best strike such an optimal balancewill be best positioned to prosper within a populationof individuals who are, in general, strongly disposed tocooperative behavior but who nevertheless sometimesdo defect.

    In summary, in the initial cooperation-unfriendlyworld, individuals can occasionally be led to enter PDgames that are played to costly mutual defect outcomes,thus selecting for high mindreading capacities. Once inplace, those capacities permit upward drift on coopera-tive dispositions. Individuals with such dispositions arethe frequent targets of exploitative attempts by thosewho are more defection-inclined, but their mindread-ing allows them to survive (via ALT choices) no dif-ferently from others in the population. At some point,however, two such high PC individuals are likely to playa PD game with each other and to do so cooperativelyand, given that ALT < c, they will prosper by compar-ison with the rest of the population. Quite rapidly, thedescendants of these individuals populate the entireecology. The equilibrium level of cooperative disposi-tions after such a transition is high enough to attractskilled mindreaders into PD games but low enoughto extract some gains from exploiting them once theyhave entered. Transitions are sustained by selection infavor of individuals who discover an optimal mixtureof mindreading and mistrustsufficient to ensure thatthey accept the true I will always cooperate messagescharacteristic of this environment but low enough toensure that they reject the false messages that are, nev-ertheless, still being sent.

    DISCUSSIONAs is often pointed out, humans ancestors were almostcertainly cooperative animals well before the point atwhich our line diverged from that of other large pri-mates (e.g., Caporael et al. 1989), meaning that ourcooperation-unfriendly starting world is best under-stood as a convenient analytic device for showing howcooperative dispositions could evolve despite adaptivepressures to the contrary. And if cooperative transitionsresembling those we have discussed ever did happen,there can be no implication that they happened withanything approximating the speed across natural gen-erations that we observe in the generations of oursimulation. Any natural cooperative transitions couldwell have taken hundreds of thousands of years.

    All we claim to have shown is that dispositions tocooperate can evolve and be sustained at equilibriumas a direct consequence of selection on Machiavelliancapacities for manipulating the content of messagessent to others and for mindreading to the truth un-derlying others attempts at manipulation. As sketchedabove, the literature on social evolution has identi-fied a number of plausible evolutionary paths to co-operative behavior, the best known being kin selection(Hamilton 1964), reciprocity (Trivers 1971), and groupselection (Sober and Wilson 1998). And the broadidea that humans and other highly social animals haveevolved Machiavellian capacities supporting their

    capacity to successfully negotiate the competitive chal-lenges of group life is a widespread and peculiarly fer-tile one. To our knowledge, however, there has beenno model of how our cooperative dispositions and ourMachiavellian capacities for manipulation and min-dreading might be functionally related, and that is thegap we have attempted to fill here.

    Although the problem of cooperation has a longanalytic history in political science, the idea ofMachiavellian intelligencethat we are a political an-imal in our cognitive design as well as in the fact thatwe relate to each other in groupsis less well known,and we believe that it deserves greater currency. This iscertainly the case if, as our data suggest, Machiavellianintelligence and cooperative dispositions have an inter-twined evolutionary history.

    Our model is based on natural selection acting onindividual cognitive and dispositional attributes, butthe processes it identifies could interact with cul-tural evolution as modeled, for example, by Boyd andRicherson (1985). As we have shown, cooperative tran-sitions occur when a pair of agents with high cooper-ative dispositions and well-developed mindreading ca-pacities recognize each other as offering a good PD betand, joining such a game, both cooperate and prosperaccordingly. In our pared-down world, however, agentsdo not have the capacity to learn from others expe-rience, to adapt their behavior by observing otherssuccess, but such a capacity certainly does exist amonghumans and, indeed, among some other primates (see,e.g., de Waal 2001). Incorporating that capacity, wemight see the success of cooperative behavior amongothers being recognized by individuals not personallydisposed in that direction, with the behavior spreadingby imitationindependent of the cognitive and dispo-sitional evolution that concerns us here.

    This raises many further possibilities. For example,populations could evolve to include some individu-als whose cooperation is a consequence of geneticallybased cooperative dispositions and others whose coop-eration is a consequence of social learning coupled witha strategic recognition that cooperative behavior willwork for them as it has for others. More interesting,perhaps, we might see individuals appearing whose co-operative behavior is a product of both innate dispo-sitions and such social learning. Social learning couldalso reveal opportunities for Machiavellian exploita-tion that would not be recognized in its absence, fur-ther accelerating the evolutionary arms race betweenmanipulation and mindreading. And benefits receivedas a consequence of social learning could select forthe (presumably genetically based) capacity for sociallearningperhaps in an upward spiral involving thatcapacity, cooperative dispositions, and Machiavellianmanipulation and mindreading. But these are specula-tions, taking us well beyond our more modest presentconcern with the evolutionary relationship among justthe latter two attributes.

    The idea of rationality is presently under disputein the discipline (Green and Shapiro 1994), but theevolutionary model proposed here allows us to dis-tinguish between two versions of rationality, perhaps

    13

  • Machiavellian Intelligence February 2004

    contributing to a reconciliation between the idea ofrationality and the empirical data on cooperationmentioned abovethat has featured in that dispute.

    Rationality in action: This is the standard way inwhich rationality is employed in political science andrelated disciplines, notably economics. Essentially, indi-viduals choose so as to maximize their private welfareunder a variety of constraints, most importantly, on in-formation. Within our simulation, this is how we havemodeled individual actors choosing between enteringPD games and playing ALT. Their choices are con-strained by others success at manipulation and by theirown limits with respect to mindreading, but they do thebest they can. And, we believe, that is how individualsshould be modeled in an evolutionary simulation suchas ours. As Alchian (1950) long ago pointed out, theidea that individuals choose as if they were rationalin this classic sense can be employed as a profitablefiction as long as interest resides in equilibria that areproduced through an evolutionary process that selectsfor fitness maximizing behavior. Remembering thatpayoffs in our evolutionary simulation are units offitness, individuals who do not act in such a mannerwould rapidly be selected out, making it reasonableto start from this model of action, even if it were inno way parallel to the actual deliberative or cognitiveprocesses that ancestral populations employed in theirsocial decision making.

    Yet, as we have pointed out, this is not how wehave modeled choices by those same individuals withinany PD games that might be joined. Those are mod-eled as random draws from a probability distribu-tion in order to capture the idea of cooperativedispositions that have no necessary basis in the indi-viduals calculation of self-interest or, indeed, in itsactual self-interest. By this emphasis on dispositionsunrelated to calculations of current interest, we aredeparting from the tradition in political science andelsewhere that seeks solutions to the problem of coop-eration in processes that are founded in rationality inaction.

    Rationality in Design: Here rationality refers to theadaptive fit between some designed apparatus and theenvironmental problems that apparatus is intended tosolvean idea usefully captured within an evolution-ary context by Tooby and Cosmides (1992) metaphorof an appropriately designed key being one that opensa particular lock. As these authors propose, a focuson design requires asking a series of engineering ques-tions, most importantly about the correspondence be-tween the problem to be solved and the mechanismdesigned by natural selection to solve it. Granted thata particular adaptive problem was a repetitive part ofthe ancestral environment, what design solutions hasnatural selection produced in response? And thinkingas a reverse engineer (Dennett 1995), What adaptivepressures in the ancient environment are most likely tohave led natural selection to design particular com-plex structurespresumed adaptationsthat we ob-serve today?

    This rationality in design approach lets us addressthe empirical fact of frequent cooperative behavior that

    is anomalous in terms of rationality in action. Our sim-ulation suggests that the most adaptive configuration ofcognitive and dispositional attributesthe most rationaldesign response by natural selection to the problems ofgroup livingis strongly but not perfectly cooperativedispositions, a modest but not particularly high level ofmistrust, and a substantial ability to mindread as wellas to manipulate. Such a model not only is consistentwith the laboratory data about cooperation, but alsohas the advantage of explaining cooperative disposi-tions squarely within the broader intellectual (and in-terdisciplinary) enterprise seeking an evolutionary un-derstanding of humans cognitive adaptations for sociallife, including Machiavellian capacities.

    This model is not inconsistent with some individu-als having only very weak dispositions to cooperate,just as it is not inconsistent with some having unusu-ally strong, even perfect, such dispositions. Individ-ual differences aside, however, the model addressesspecies-typical attributes, suggesting that, as a default,we should expect social animalsmost interestingly,of course, humansto be quite strongly disposed tocooperative behavior, as well as modestly trusting andreasonably adept at both manipulation and mindread-ing. How individuals cooperative dispositions might beundermined or reinforced by the incentive attributes ofthe present situation is, of course, a different matter.

    The availability of an alternative to playing prisonersdilemma games as a way of gathering resources is criti-cal to our model, and our finding that cooperative tran-sitions only happen within the finite range 0 < ALT < cinvites further study. Whereas we believe that the logicby which transitions are so confined is clear, the moredifficult question concerns the natural world circum-stances that might embody that logicor, more ac-curately perhaps, might have embodied that logic inthe ancient environment. Most interesting, we think, isthe finding that cooperative dispositions evolve to theirhighest level when the payoff from alternative ways ofgathering resources most closely approaches the payofffrom mutual cooperation without actually exceeding it.This appears counter-intuitive at first; one might expectcooperation to evolve to its highest levels when there isthe most to be gained from mutual cooperation relativeto alternative courses of action. But as the argumentabout particular equilibria of cooperative dispositionssuggests, a greater distance between ALT and c justallows more room for noncooperative dispositionsto evolve.

    We recognize, of course, that the ancestral environ-ment that shaped our modern cognitive apparatus didnot involve a single set of payoff parameters (as doparticular runs of our simulation), but a substantial di-versity of particular parameters, making that environ-ment best thought of as a statistical average across thatdiversity (Daly and Wilson 1999; Tooby and Cosmides1990). Nevertheless, our model does provide a basis forhypothesizing that sociality itselfa willingness to en-ter PD-type games coupled with a strong disposition toplay such games in a cooperative mannerevolvedto its highest levels when there were only marginal gainsto be had from jointly cooperative action in comparison

    14

  • American Political Science Review Vol. 98, No. 1

    with going it alone. The problems associated withsociality became more acute as relatively higher pay-offs became available from cooperative activities, thuspermitting the evolution of at least somewhat more am-bivalent cooperative dispositions.

    REFERENCESAlchian, Armen. 1950. Uncertainty, Evolution and Economic The-

    ory. Journal of Political Economy 58 (3 June): 21121.Axelrod, Robert. 1981. Emergence of Cooperation among Egoists.

    American Political Science Review 75 (June): 30618.Axelrod, Robert. 1984. The Evolution of Cooperation. New York:

    Basic Books.Boehm, Christopher. 1997. Egalitarian Behavior and the Evolution

    of Political Intelligence. In Machiavellian Intelligence II, ed. R. W.Byrne and A. Whiten. Cambridge: Cambridge University Press,34164.

    Bowlby, J. 1969. Attachment and Loss. New York: Basic Books.Boyd, Robert, and Peter J. Richerson. 1985. Culture and the Evolu-

    tionary Process. Chicago: University of Chicago Press.Boyd, Robert, Herbert Gintis, Samuel Bowles, and Peter J.

    Richerson. 2003. The Evolution of Altruistic Punishment. Pro-ceedings of the National Academy of Sciences of the United Statesof America 1000 (March 11): 3531.

    Brothers, Leslie. 1997. Fridays Footprint: How Society Shapes theHuman Mind. New York: Oxford University Press.

    Byrne, Richard W., and Andrew Whiten, eds. 1988. MachiavellianIntelligence; Social Expertise and the Evolution of Intellect in Mon-keys, Apes, and Humans. New York: New York: Oxford UniversityPress.

    Camerer, C. F., and R. Thaler. 1995. AnomaliesUltimatums, Dic-tators, and Manners. Journal of Economic Perspectives 9 (Spring):20919.

    Caporael, Linnda, Robyn Dawes, John Orbell, and Alphons vande Kragt. 1989. Selfishness Examined: Cooperation in the Ab-sence of Egoistic Incentives. Behavioral and Brain Science 12(December): 68399.

    Daly, Martin, and Margo Wilson. 1999. Human Evolutionary Psy-chology and Animal Behavior. Animal Behaviour 57 (March):50919.

    Dawes, Robyn. 1975. Formal Models of Dilemmas in SocialDecision-Making. In Human Judgment and Decision Processes,ed. M. F. Kaplan and S. Schwartz. New York: Academic Press,87108.

    Dawkins, Richard. 1976. The Selfish Gene. New York: Oxford Uni-versity Press.

    Dawkins, Richard, and John R. Krebs. 1978. Animal Signals: Infor-mation or Manipulation? In Behavioural Ecology: An Evolution-ary Approach, ed. J. R. Krebs and N. B. Davies. Oxford: BlackwellScientific, 282309.

    de Waal, Frans. 2001. The Ape and the Sushi Master; Cultural Reflec-tions by a Primatologist. New York: Basic Books.

    Dennett, Daniel. 1995. Darwins Dangerous Idea: Evolution and theMeanings of Life. New York: Simon & Schuster.

    Dobzhansky, Theodosius. 1973. Nothing in Biology Makes SenseExcept in the Light of Evolution. American Biology Teacher 35(March): 12529.

    Dunbar, Robin I. M., Chris Knight, and Camilla Power, eds. 1999.The Evolution of Culture. New Brunswick, NJ: Rutgers UniversityPress.

    Field, Alexander J. 2001. Altruistically Inclined? The Behavioral Sci-ences, Evolutionary Theory, and the Origins of Reciprocity. Editedby T. Kuran, Economics, Cognition, and Society. Ann Arbor:University of Michigan Press.

    Frohlich, N., and J. Oppenheimer. 1996. Experiencing Impartial-ity to Invoke Fairness in the N-Pd: Some Experimental Results.Public Choice 86 (January): 11735.

    Fukuyama, Francis. 1995. Trust: The Social Virtues and the Creationof Prosperity. New York: Free Press.

    Gigerenzer, Gerd. 1997. The Modularity of Social Intelligence.In Machiavellian Intelligence II: Extensions and Evaluations, ed.A. Whiten and R. W. Byrne. Cambridge: Cambridge UniversityPress, 26488.

    Green, Donald P., and Ian Shapiro. 1994. Pathologies of RationalChoice Theory; A Critique of Applications in Political Science. NewHaven, CT: Yale University Press.

    Hamilton, W. D. 1964. The Genetical Evolution of Social Behaviour.I and II. Journal of Theoretical Biology 7 (July): 152.

    Hardin, Garrett. 1977. The Limits of Altruism; An Ecologists Viewof Survival. Bloomington: Indiana University Press.

    Hardin, Russell. 1991. Trusting Persons, Trusting Institutions. InThe Strategy of Choice, ed. R. J. Zeckhauser. Cambridge, MA:MIT Press, 185209.

    Hobbes, Thomas. [1651] 1947. Leviathan. New York: E. P. Dutton.Humphrey, Nicholas K. 1976. The Social Function of Intellect. In

    Growing Points in Ethology, ed. P. P. G. Bateson and R. A. Hinde.Cambridge: Cambridge University Press, 30317.

    Jolly, Allison. 1966. Lemur Social Behavior and Primate Intelli-gence. Science 153 (July): 5016.

    Mithen, Steven. 1996. The Prehistory of the Mind; The CognitiveOrigins of Art, Religion and Science. London: Thames & Hudson.

    Nesse, Randolph M. 2001. Evolution and the Capacity for Commit-ment. ed. K. S. Cook, M. Levi and R. Hardin, The Russell SageFoundation Series on Trust. New York: Russell Sage Foundation.

    Olson, Mancur. 1965. The Logic of Collective Action. Cambridge,MA: Harvard University Press.

    Orbell, John, and Robyn Dawes. 1993. Social Welfare, CooperatorsAdvantage and the Option of Not Playing the Game. AmericanSociological Review 58 (December): 787800.

    Orbell, John, Robyn Dawes, and Peregrine Schwartz-Shea. 1994.Trust, Social Categories and Individuals: The Case of Gender.Motivation and Emotion 18 (June):10928.

    Orbell, John, Robyn Dawes, Randy Simmons, and Alphons van deKragt. 1986. Organizing Groups for Collective Action. AmericanPolitical Science Review 80 (December): 117185.

    Orbell, John, Tomonori Morikawa, and Nicholas Allen. 2002. TheEvolution of Political Intelligence: Simulation Results. BritishJournal of Political Science 32 (October): 61339.

    Ostrom, Elinor. 1998. A Behavioral Approach to the RationalChoice Theory of Collective Action. American Political ScienceReview 92 (March): 122.

    Ostrom, Elinor, James Walker, and Roy Gardner. 1992. Covenantswith and without the Sword: Self-Governance Is Possible. Amer-ican Political Science Review 86 (June): 40417.

    Putnam, Robert. 2000. Bowling Alone: The Collapse and Revival ofAmerican Community. New York: Simon & Schuster.

    Rapoport, Anatol, Melvin J. Guyer, and David G. Gordon. 1976. The2 2 Game. Ann Arbor: University of Michigan Press.

    Reeve, Hudson Kern. 2000. Multi-Level Selection and Human Co-operation. Evolution and Human Behavior 21 (January): 6572.

    Simon, Herbert. 1985. Human Nature in Politics; The Dialogue ofPolitical Science with Psychology. American Political Science Re-view June (2): 293304.

    Sober, Elliott, and David Sloan Wilson. 1998. Unto Others: The Evo-lution and Psychology of Unselfish Behavior. Cambridge, MA:Harvard University Press.

    Tinbergen, N. 1963. On Aims and Methods of Ethology. Zeitschriftfur Angewandte Zoologie 20: 41033.

    Tooby, John, and Leda Cosmides. 1990. The Past Explains thePresent: Emotional Adaptations and the Structure of Ances-tral Environments. Ethology and Sociobiology 11 (45): 375424.

    Tooby, John, and Leda Cosmides. 1992. The Psychological Founda-tions of Culture. In The Adapted Mind: Evolutionary Psychologyand the Generation of Culture, ed. J. Barkow, L. Cosmides, andJ. Tooby. New York: Oxford University Press, 19136.

    Trivers, Robert. 1971. The Evolution of Reciprocal Altruism. Quar-terly Review of Biology 46 (March): 3557.

    Trivers, Robert. 1985. Social Evolution. Menlo Park, CA:Benjamin/Cummings.

    Whiten, Andrew, and Richard W. Byrne, eds. 1997. Machiavellian In-telligence II: Extensions and Evaluations. Cambridge: CambridgeUniversity Press.

    Williams, George C. 1966. Adaptation and Natural Selection: ACritique of Some Current Evolutionary Thought. Princeton, NJ:Princeton University Press.

    Wynn, Thomas. 2002. Archaeology and Cognitive Evolution.Behavioral and Brain Science 25 (June): 389438.

    15