Search This Blog

Wednesday, December 15, 2021

Evolutionarily stable strategy

From Wikipedia, the free encyclopedia
 
Evolutionarily stable strategy
A solution concept in game theory
Relationship
Subset ofNash equilibrium
Superset ofStochastically stable equilibrium, Stable Strong Nash equilibrium
Intersects withSubgame perfect equilibrium, Trembling hand perfect equilibrium, Perfect Bayesian equilibrium
Significance
Proposed byJohn Maynard Smith and George R. Price
Used forBiological modeling and Evolutionary game theory
ExampleHawk-dove

An evolutionarily stable strategy (ESS) is a strategy (or set of strategies) that is impermeable when adopted by a population in adaptation to a specific environment, that is to say it cannot be displaced by an alternative strategy (or set of strategies) which may be novel or initially rare. Introduced by John Maynard Smith and George R. Price in 1972/3, it is an important concept in behavioural ecology, evolutionary psychology, mathematical game theory and economics, with applications in other fields such as anthropology, philosophy and political science.

In game-theoretical terms, an ESS is an equilibrium refinement of the Nash equilibrium, being a Nash equilibrium that is also "evolutionarily stable." Thus, once fixed in a population, natural selection alone is sufficient to prevent alternative (mutant) strategies from replacing it (although this does not preclude the possibility that a better strategy, or set of strategies, will emerge in response to selective pressures resulting from environmental change).

History

Evolutionarily stable strategies were defined and introduced by John Maynard Smith and George R. Price in a 1973 Nature paper. Such was the time taken in peer-reviewing the paper for Nature that this was preceded by a 1972 essay by Maynard Smith in a book of essays titled On Evolution. The 1972 essay is sometimes cited instead of the 1973 paper, but university libraries are much more likely to have copies of Nature. Papers in Nature are usually short; in 1974, Maynard Smith published a longer paper in the Journal of Theoretical Biology. Maynard Smith explains further in his 1982 book Evolution and the Theory of Games. Sometimes these are cited instead. In fact, the ESS has become so central to game theory that often no citation is given, as the reader is assumed to be familiar with it.

Maynard Smith mathematically formalised a verbal argument made by Price, which he read while peer-reviewing Price's paper. When Maynard Smith realized that the somewhat disorganised Price was not ready to revise his article for publication, he offered to add Price as co-author.

The concept was derived from R. H. MacArthur and W. D. Hamilton's work on sex ratios, derived from Fisher's principle, especially Hamilton's (1967) concept of an unbeatable strategy. Maynard Smith was jointly awarded the 1999 Crafoord Prize for his development of the concept of evolutionarily stable strategies and the application of game theory to the evolution of behaviour.

Uses of ESS:

Motivation

The Nash equilibrium is the traditional solution concept in game theory. It depends on the cognitive abilities of the players. It is assumed that players are aware of the structure of the game and consciously try to predict the moves of their opponents and to maximize their own payoffs. In addition, it is presumed that all the players know this (see common knowledge). These assumptions are then used to explain why players choose Nash equilibrium strategies.

Evolutionarily stable strategies are motivated entirely differently. Here, it is presumed that the players' strategies are biologically encoded and heritable. Individuals have no control over their strategy and need not be aware of the game. They reproduce and are subject to the forces of natural selection, with the payoffs of the game representing reproductive success (biological fitness). It is imagined that alternative strategies of the game occasionally occur, via a process like mutation. To be an ESS, a strategy must be resistant to these alternatives.

Given the radically different motivating assumptions, it may come as a surprise that ESSes and Nash equilibria often coincide. In fact, every ESS corresponds to a Nash equilibrium, but some Nash equilibria are not ESSes.

Nash equilibrium

An ESS is a refined or modified form of a Nash equilibrium. (See the next section for examples which contrast the two.) In a Nash equilibrium, if all players adopt their respective parts, no player can benefit by switching to any alternative strategy. In a two player game, it is a strategy pair. Let E(S,T) represent the payoff for playing strategy S against strategy T. The strategy pair (S, S) is a Nash equilibrium in a two player game if and only if for both players, for any strategy T:

E(S,S) ≥ E(T,S)

In this definition, a strategy TS can be a neutral alternative to S (scoring equally well, but not better). A Nash equilibrium is presumed to be stable even if T scores equally, on the assumption that there is no long-term incentive for players to adopt T instead of S. This fact represents the point of departure of the ESS.

Maynard Smith and Price specify two conditions for a strategy S to be an ESS. For all TS, either

  1. E(S,S) > E(T,S), or
  2. E(S,S) = E(T,S) and E(S,T) > E(T,T)

The first condition is sometimes called a strict Nash equilibrium. The second is sometimes called "Maynard Smith's second condition". The second condition means that although strategy T is neutral with respect to the payoff against strategy S, the population of players who continue to play strategy S has an advantage when playing against T.

There is also an alternative, stronger definition of ESS, due to Thomas. This places a different emphasis on the role of the Nash equilibrium concept in the ESS concept. Following the terminology given in the first definition above, this definition requires that for all TS

  1. E(S,S) ≥ E(T,S), and
  2. E(S,T) > E(T,T)

In this formulation, the first condition specifies that the strategy is a Nash equilibrium, and the second specifies that Maynard Smith's second condition is met. Note that the two definitions are not precisely equivalent: for example, each pure strategy in the coordination game below is an ESS by the first definition but not the second.

In words, this definition looks like this: The payoff of the first player when both players play strategy S is higher than (or equal to) the payoff of the first player when he changes to another strategy T and the second player keeps his strategy S and the payoff of the first player when only his opponent changes his strategy to T is higher than his payoff in case that both of players change their strategies to T.

This formulation more clearly highlights the role of the Nash equilibrium condition in the ESS. It also allows for a natural definition of related concepts such as a weak ESS or an evolutionarily stable set.

Examples of differences between Nash equilibria and ESSes


Cooperate Defect
Cooperate 3, 3 1, 4
Defect 4, 1 2, 2
Prisoner's Dilemma

A B
A 2, 2 1, 2
B 2, 1 2, 2
Harm thy neighbor

In most simple games, the ESSes and Nash equilibria coincide perfectly. For instance, in the prisoner's dilemma there is only one Nash equilibrium, and its strategy (Defect) is also an ESS.

Some games may have Nash equilibria that are not ESSes. For example, in harm thy neighbor (whose payoff matrix is shown here) both (A, A) and (B, B) are Nash equilibria, since players cannot do better by switching away from either. However, only B is an ESS (and a strong Nash). A is not an ESS, so B can neutrally invade a population of A strategists and predominate, because B scores higher against B than A does against B. This dynamic is captured by Maynard Smith's second condition, since E(A, A) = E(B, A), but it is not the case that E(A,B) > E(B,B).


C D
C 2, 2 1, 2
D 2, 1 0, 0
Harm everyone

Swerve Stay
Swerve 0,0 −1,+1
Stay +1,−1 −20,−20
Chicken

Nash equilibria with equally scoring alternatives can be ESSes. For example, in the game Harm everyone, C is an ESS because it satisfies Maynard Smith's second condition. D strategists may temporarily invade a population of C strategists by scoring equally well against C, but they pay a price when they begin to play against each other; C scores better against D than does D. So here although E(C, C) = E(D, C), it is also the case that E(C,D) > E(D,D). As a result, C is an ESS.

Even if a game has pure strategy Nash equilibria, it might be that none of those pure strategies are ESS. Consider the Game of chicken. There are two pure strategy Nash equilibria in this game (Swerve, Stay) and (Stay, Swerve). However, in the absence of an uncorrelated asymmetry, neither Swerve nor Stay are ESSes. There is a third Nash equilibrium, a mixed strategy which is an ESS for this game (see Hawk-dove game and Best response for explanation).

This last example points to an important difference between Nash equilibria and ESS. Nash equilibria are defined on strategy sets (a specification of a strategy for each player), while ESS are defined in terms of strategies themselves. The equilibria defined by ESS must always be symmetric, and thus have fewer equilibrium points.

Vs. evolutionarily stable state

In population biology, the two concepts of an evolutionarily stable strategy (ESS) and an evolutionarily stable state are closely linked but describe different situations.

In an evolutionarily stable strategy, if all the members of a population adopt it, no mutant strategy can invade. Once virtually all members of the population use this strategy, there is no 'rational' alternative. ESS is part of classical game theory.

In an evolutionarily stable state, a population's genetic composition is restored by selection after a disturbance, if the disturbance is not too large. An evolutionarily stable state is a dynamic property of a population that returns to using a strategy, or mix of strategies, if it is perturbed from that initial state. It is part of population genetics, dynamical system, or evolutionary game theory. This is now called convergent stability.

B. Thomas (1984) applies the term ESS to an individual strategy which may be mixed, and evolutionarily stable population state to a population mixture of pure strategies which may be formally equivalent to the mixed ESS.

Whether a population is evolutionarily stable does not relate to its genetic diversity: it can be genetically monomorphic or polymorphic.

Stochastic ESS

In the classic definition of an ESS, no mutant strategy can invade. In finite populations, any mutant could in principle invade, albeit at low probability, implying that no ESS can exist. In an infinite population, an ESS can instead be defined as a strategy which, should it become invaded by a new mutant strategy with probability p, would be able to counterinvade from a single starting individual with probability >p, as illustrated by the evolution of bet-hedging.

Prisoner's dilemma


Cooperate Defect
Cooperate 3, 3 1, 4
Defect 4, 1 2, 2
Prisoner's Dilemma

A common model of altruism and social cooperation is the Prisoner's dilemma. Here a group of players would collectively be better off if they could play Cooperate, but since Defect fares better each individual player has an incentive to play Defect. One solution to this problem is to introduce the possibility of retaliation by having individuals play the game repeatedly against the same player. In the so-called iterated Prisoner's dilemma, the same two individuals play the prisoner's dilemma over and over. While the Prisoner's dilemma has only two strategies (Cooperate and Defect), the iterated Prisoner's dilemma has a huge number of possible strategies. Since an individual can have different contingency plan for each history and the game may be repeated an indefinite number of times, there may in fact be an infinite number of such contingency plans.

Three simple contingency plans which have received substantial attention are Always Defect, Always Cooperate, and Tit for Tat. The first two strategies do the same thing regardless of the other player's actions, while the latter responds on the next round by doing what was done to it on the previous round—it responds to Cooperate with Cooperate and Defect with Defect.

If the entire population plays Tit-for-Tat and a mutant arises who plays Always Defect, Tit-for-Tat will outperform Always Defect. If the population of the mutant becomes too large — the percentage of the mutant will be kept small. Tit for Tat is therefore an ESS, with respect to only these two strategies. On the other hand, an island of Always Defect players will be stable against the invasion of a few Tit-for-Tat players, but not against a large number of them. If we introduce Always Cooperate, a population of Tit-for-Tat is no longer an ESS. Since a population of Tit-for-Tat players always cooperates, the strategy Always Cooperate behaves identically in this population. As a result, a mutant who plays Always Cooperate will not be eliminated. However, even though a population of Always Cooperate and Tit-for-Tat can coexist, if there is a small percentage of the population that is Always Defect, the selective pressure is against Always Cooperate, and in favour of Tit-for-Tat. This is due to the lower payoffs of cooperating than those of defecting in case the opponent defects.

This demonstrates the difficulties in applying the formal definition of an ESS to games with large strategy spaces, and has motivated some to consider alternatives.

Human behavior

The fields of sociobiology and evolutionary psychology attempt to explain animal and human behavior and social structures, largely in terms of evolutionarily stable strategies. Sociopathy (chronic antisocial or criminal behavior) may be a result of a combination of two such strategies.

Evolutionarily stable strategies were originally considered for biological evolution, but they can apply to other contexts. In fact, there are stable states for a large class of adaptive dynamics. As a result, they can be used to explain human behaviours that lack any genetic influences.

Evolutionary game theory

From Wikipedia, the free encyclopedia

Evolutionary game theory (EGT) is the application of game theory to evolving populations in biology. It defines a framework of contests, strategies, and analytics into which Darwinian competition can be modelled. It originated in 1973 with John Maynard Smith and George R. Price's formalisation of contests, analysed as strategies, and the mathematical criteria that can be used to predict the results of competing strategies.

Evolutionary game theory differs from classical game theory in focusing more on the dynamics of strategy change. This is influenced by the frequency of the competing strategies in the population.

Evolutionary game theory has helped to explain the basis of altruistic behaviours in Darwinian evolution. It has in turn become of interest to economists, sociologists, anthropologists, and philosophers.

History

Classical game theory

Classical non-cooperative game theory was conceived by John von Neumann to determine optimal strategies in competitions between adversaries. A contest involves players, all of whom have a choice of moves. Games can be a single round or repetitive. The approach a player takes in making his moves constitutes his strategy. Rules govern the outcome for the moves taken by the players, and outcomes produce payoffs for the players; rules and resulting payoffs can be expressed as decision trees or in a payoff matrix. Classical theory requires the players to make rational choices. Each player must consider the strategic analysis that his opponents are making to make his own choice of moves.

The problem of ritualized behaviour

The mathematical biologist John Maynard Smith modelled evolutionary games.

Evolutionary game theory started with the problem of how to explain ritualized animal behaviour in a conflict situation; "why are animals so 'gentlemanly or ladylike' in contests for resources?" The leading ethologists Niko Tinbergen and Konrad Lorenz proposed that such behaviour exists for the benefit of the species. John Maynard Smith considered that incompatible with Darwinian thought, where selection occurs at an individual level, so self-interest is rewarded while seeking the common good is not. Maynard Smith, a mathematical biologist, turned to game theory as suggested by George Price, though Richard Lewontin's attempts to use the theory had failed.

Adapting game theory to evolutionary games

Maynard Smith realised that an evolutionary version of game theory does not require players to act rationally—only that they have a strategy. The results of a game show how good that strategy was, just as evolution tests alternative strategies for the ability to survive and reproduce. In biology, strategies are genetically inherited traits that control an individual's action, analogous with computer programs. The success of a strategy is determined by how good the strategy is in the presence of competing strategies (including itself), and of the frequency with which those strategies are used. Maynard Smith described his work in his book Evolution and the Theory of Games.

Participants aim to produce as many replicas of themselves as they can, and the payoff is in units of fitness (relative worth in being able to reproduce). It is always a multi-player game with many competitors. Rules include replicator dynamics, in other words how the fitter players will spawn more replicas of themselves into the population and how the less fit will be culled, in a replicator equation. The replicator dynamics models heredity but not mutation, and assumes asexual reproduction for the sake of simplicity. Games are run repetitively with no terminating conditions. Results include the dynamics of changes in the population, the success of strategies, and any equilibrium states reached. Unlike in classical game theory, players do not choose their strategy and cannot change it: they are born with a strategy and their offspring inherit that same strategy.

Evolutionary games

Models

Evolutionary game theory analyses Darwinian mechanisms with a system model with three main components – population, game, and replicator dynamics. The system process has four phases:

1) The model (as evolution itself) deals with a population (Pn). The population will exhibit variation among competing individuals. In the model this competition is represented by the game.

2) The game tests the strategies of the individuals under the rules of the game. These rules produce different payoffs – in units of fitness (the production rate of offspring). The contesting individuals meet in pairwise contests with others, normally in a highly mixed distribution of the population. The mix of strategies in the population affects the payoff results by altering the odds that any individual may meet up in contests with various strategies. The individuals leave the game pairwise contest with a resulting fitness determined by the contest outcome, represented in a payoff matrix.

3) Based on this resulting fitness each member of the population then undergoes replication or culling determined by the exact mathematics of the replicator dynamics process. This overall process then produces a new generation P(n+1). Each surviving individual now has a new fitness level determined by the game result.

4) The new generation then takes the place of the previous one and the cycle repeats. The population mix may converge to an evolutionarily stable state that cannot be invaded by any mutant strategy.

Evolutionary game theory encompasses Darwinian evolution, including competition (the game), natural selection (replicator dynamics), and heredity. Evolutionary game theory has contributed to the understanding of group selection, sexual selection, altruism, parental care, co-evolution, and ecological dynamics. Many counter-intuitive situations in these areas have been put on a firm mathematical footing by the use of these models.

The common way to study the evolutionary dynamics in games is through replicator equations. These show the growth rate of the proportion of organisms using a certain strategy and that rate is equal to the difference between the average payoff of that strategy and the average payoff of the population as a whole. Continuous replicator equations assume infinite populations, continuous time, complete mixing and that strategies breed true. The attractors (stable fixed points) of the equations are equivalent with evolutionarily stable states. A strategy which can survive all "mutant" strategies is considered evolutionarily stable. In the context of animal behavior, this usually means such strategies are programmed and heavily influenced by genetics, thus making any player or organism's strategy determined by these biological factors.

Evolutionary games are mathematical objects with different rules, payoffs, and mathematical behaviours. Each "game" represents different problems that organisms have to deal with, and the strategies they might adopt to survive and reproduce. Evolutionary games are often given colourful names and cover stories which describe the general situation of a particular game. Representative games include hawk-dove, war of attrition, stag hunt, producer-scrounger, tragedy of the commons, and prisoner's dilemma. Strategies for these games include hawk, dove, bourgeois, prober, defector, assessor, and retaliator. The various strategies compete under the particular game's rules, and the mathematics are used to determine the results and behaviours.

Hawk dove

Solution of the hawk dove game for V=2, C=10 and fitness starting base B=4. The fitness of a hawk for different population mixes is plotted as a black line, that of dove in red. An ESS (a stationary point) will exist when hawk and dove fitness are equal: Hawks are 20% of population and doves are 80% of the population.

The first game that Maynard Smith analysed is the classic hawk dove game. It was conceived to analyse Lorenz and Tinbergen's problem, a contest over a shareable resource. The contestants can be either a hawk or a dove. These are two subtypes or morphs of one species with different strategies. The hawk first displays aggression, then escalates into a fight until it either wins or is injured (loses). The dove first displays aggression, but if faced with major escalation runs for safety. If not faced with such escalation, the dove attempts to share the resource.

Payoff matrix for hawk dove game

meets hawk meets dove
if hawk V/2 − C/2 V
if dove 0 V/2

Given that the resource is given the value V, the damage from losing a fight is given cost C:

  • If a hawk meets a dove, the hawk gets the full resource V
  • If a hawk meets a hawk, half the time they win, half the time they lose...so the average outcome is then V/2 minus C/2
  • If a dove meets a hawk, the dove will back off and get nothing – 0
  • If a dove meets a dove, both share the resource and get V/2

The actual payoff, however, depends on the probability of meeting a hawk or dove, which in turn is a representation of the percentage of hawks and doves in the population when a particular contest takes place. That, in turn, is determined by the results of all of the previous contests. If the cost of losing C is greater than the value of winning V (the normal situation in the natural world) the mathematics ends in an evolutionarily stable strategy (ESS), a mix of the two strategies where the population of hawks is V/C. The population regresses to this equilibrium point if any new hawks or doves make a temporary perturbation in the population. The solution of the hawk dove game explains why most animal contests involve only ritual fighting behaviours in contests rather than outright battles. The result does not at all depend on "good of the species" behaviours as suggested by Lorenz, but solely on the implication of actions of so-called selfish genes.

War of attrition

In the hawk dove game the resource is shareable, which gives payoffs to both doves meeting in a pairwise contest. Where the resource is not shareable, but an alternative resource might be available by backing off and trying elsewhere, pure hawk or dove strategies are less effective. If an unshareable resource is combined with a high cost of losing a contest (injury or possible death) both hawk and dove payoffs are further diminished. A safer strategy of lower cost display, bluffing and waiting to win, is then viable – a bluffer strategy. The game then becomes one of accumulating costs, either the costs of displaying or the costs of prolonged unresolved engagement. It is effectively an auction; the winner is the contestant who will swallow the greater cost while the loser gets the same cost as the winner but no resource. The resulting evolutionary game theory mathematics lead to an optimal strategy of timed bluffing.

War of attrition for different values of resource. Note the time it takes for an accumulation of 50% of the contestants to quit vs. the value (V) of resource contested for.

This is because in the war of attrition any strategy that is unwavering and predictable is unstable, because it will ultimately be displaced by a mutant strategy which relies on the fact that it can best the existing predictable strategy by investing an extra small delta of waiting resource to ensure that it wins. Therefore, only a random unpredictable strategy can maintain itself in a population of bluffers. The contestants in effect choose an acceptable cost to be incurred related to the value of the resource being sought, effectively making a random bid as part of a mixed strategy (a strategy where a contestant has several, or even many, possible actions in their strategy). This implements a distribution of bids for a resource of specific value V, where the bid for any specific contest is chosen at random from that distribution. The distribution (an ESS) can be computed using the Bishop-Cannings theorem, which holds true for any mixed-strategy ESS. The distribution function in these contests was determined by Parker and Thompson to be:

The result is that the cumulative population of quitters for any particular cost m in this "mixed strategy" solution is:

as shown in the adjacent graph. The intuitive sense that greater values of resource sought leads to greater waiting times is borne out. This is observed in nature, as in male dung flies contesting for mating sites, where the timing of disengagement in contests is as predicted by evolutionary theory mathematics.

Asymmetries that allow new strategies

Dung fly (Scatophaga stercoraria) – a war of attrition player
 
The mantis shrimp guarding its home with the bourgeois strategy
 
Animal strategy examples: by examining the behaviours, then determining both the costs and the values of resources attained in a contest the strategy of an organism can be verified

In the war of attrition there must be nothing that signals the size of a bid to an opponent, otherwise the opponent can use the cue in an effective counter-strategy. There is however a mutant strategy which can better a bluffer in the war of attrition game if a suitable asymmetry exists, the bourgeois strategy. Bourgeois uses an asymmetry of some sort to break the deadlock. In nature one such asymmetry is possession of a resource. The strategy is to play a hawk if in possession of the resource, but to display then retreat if not in possession. This requires greater cognitive capability than hawk, but bourgeois is common in many animal contests, such as in contests among mantis shrimps and among speckled wood butterflies.

Social behaviour

Alternatives for game theoretic social interaction

Games like hawk dove and war of attrition represent pure competition between individuals and have no attendant social elements. Where social influences apply, competitors have four possible alternatives for strategic interaction. This is shown on the adjacent figure, where a plus sign represents a benefit and a minus sign represents a cost.

  • In a cooperative or mutualistic relationship both "donor" and "recipient" are almost indistinguishable as both gain a benefit in the game by co-operating, i.e. the pair are in a game-wise situation where both can gain by executing a certain strategy, or alternatively both must act in concert because of some encompassing constraints that effectively puts them "in the same boat".
  • In an altruistic relationship the donor, at a cost to themself provides a benefit to the recipient. In the general case the recipient will have a kin relationship to the donor and the donation is one-way. Behaviours where benefits are donated alternatively (in both directions) at a cost, are often called "altruistic", but on analysis such "altruism" can be seen to arise from optimised "selfish" strategies.
  • Spite is essentially a "reversed" form of altruism where an ally is aided by damaging the ally's competitors. The general case is that the ally is kin related and the benefit is an easier competitive environment for the ally. Note: George Price, one of the early mathematical modellers of both altruism and spite, found this equivalence particularly disturbing at an emotional level.
  • Selfishness is the base criteria of all strategic choice from a game theory perspective – strategies not aimed at self-survival and self-replication are not long for any game. Critically however, this situation is impacted by the fact that competition is taking place on multiple levels – i.e. at a genetic, an individual and a group level.

Contests of selfish genes

Female Belding's ground squirrels risk their lives giving loud alarm calls, protecting closely related female colony members; males are less closely related and do not call.

At first glance it may appear that the contestants of evolutionary games are the individuals present in each generation who directly participate in the game. But individuals live only through one game cycle, and instead it is the strategies that really contest with one another over the duration of these many-generation games. So it is ultimately genes that play out a full contest – selfish genes of strategy. The contesting genes are present in an individual and to a degree in all of the individual's kin. This can sometimes profoundly affect which strategies survive, especially with issues of cooperation and defection. William Hamilton, known for his theory of kin selection, explored many of these cases using game-theoretic models. Kin-related treatment of game contests helps to explain many aspects of the behaviour of social insects, the altruistic behaviour in parent-offspring interactions, mutual protection behaviours, and co-operative care of offspring. For such games, Hamilton defined an extended form of fitness – inclusive fitness, which includes an individual's offspring as well as any offspring equivalents found in kin.

The mathematics of kin selection
The concept of kin selection is that:
inclusive fitness=own contribution to fitness + contribution of all relatives.

Fitness is measured relative to the average population; for example, fitness=1 means growth at the average rate for the population, fitness < 1 means having a decreasing share in the population (dying out), fitness > 1 means an increasing share in the population (taking over).

The inclusive fitness of an individual wi is the sum of its specific fitness of itself ai plus the specific fitness of each and every relative weighted by the degree of relatedness which equates to the summation of all rj*bj....... where rj is relatedness of a specific relative and bj is that specific relative's fitness – producing:

If individual ai sacrifices their "own average equivalent fitness of 1" by accepting a fitness cost C, and then to "get that loss back", wi must still be 1 (or greater than 1)...and using R*B to represent the summation results in:

1< (1-C)+RB ....or rearranging..... R>C/B.

Hamilton went beyond kin relatedness to work with Robert Axelrod, analysing games of co-operation under conditions not involving kin where reciprocal altruism came into play.

Eusociality and kin selection

Meat ant workers (always female) are related to a parent by a factor of 0.5, to a sister by 0.75, to a child by 0.5 and to a brother by 0.25. Therefore, it is significantly more advantageous to help produce a sister (0.75) than to have a child (0.5).

Eusocial insect workers forfeit reproductive rights to their queen. It has been suggested that kin selection, based on the genetic makeup of these workers, may predispose them to altruistic behaviours. Most eusocial insect societies have haplodiploid sexual determination, which means that workers are unusually closely related.

This explanation of insect eusociality has, however, been challenged by a few highly-noted evolutionary game theorists (Nowak and Wilson) who have published a controversial alternative game theoretic explanation based on a sequential development and group selection effects proposed for these insect species.

Prisoner's dilemma

A difficulty of the theory of evolution, recognised by Darwin himself, was the problem of altruism. If the basis for selection is at an individual level, altruism makes no sense at all. But universal selection at the group level (for the good of the species, not the individual) fails to pass the test of the mathematics of game theory and is certainly not the general case in nature. Yet in many social animals, altruistic behaviour exists. The solution to this problem can be found in the application of evolutionary game theory to the prisoner's dilemma game – a game which tests the payoffs of cooperating or in defecting from cooperation. It is the most studied game in all of game theory.

The analysis of the prisoner's dilemma is as a repetitive game. This affords competitors the possibility of retaliating for defection in previous rounds of the game. Many strategies have been tested; the best competitive strategies are general cooperation, with a reserved retaliatory response if necessary. The most famous and one of the most successful of these is tit-for-tat with a simple algorithm.

def tit_for_tat(last_move_by_opponent):
    if last_move_by_opponent == defect:
        defect()
    else:
        cooperate()

The pay-off for any single round of the game is defined by the pay-off matrix for a single round game (shown in bar chart 1 below). In multi-round games the different choices – co-operate or defect – can be made in any particular round, resulting in a certain round payoff. It is, however, the possible accumulated pay-offs over the multiple rounds that count in shaping the overall pay-offs for differing multi-round strategies such as tit-for-tat.

Payoffs in two varieties of prisoner's dilemma game
Prisoner's dilemma: co-operate or defect
Payoff (temptation in defecting vs. co-operation) > Payoff (mutual co-operation) > Payoff(joint defection) > Payoff(sucker co-operates but opponent defects)

Example 1: The straightforward single round prisoner's dilemma game. The classic prisoner's dilemma game payoffs gives a player a maximum payoff if they defect and their partner co-operates (this choice is known as temptation). If, however, the player co-operates and their partner defects, they get the worst possible result (the suckers payoff). In these payoff conditions the best choice (a Nash equilibrium) is to defect.

Example 2: Prisoner's dilemma played repeatedly. The strategy employed is tit-for-tat which alters behaviours based on the action taken by a partner in the previous round – i.e. reward co-operation and punish defection. The effect of this strategy in accumulated payoff over many rounds is to produce a higher payoff for both players' co-operation and a lower payoff for defection. This removes the temptation to defect. The suckers payoff also becomes less, although "invasion" by a pure defection strategy is not entirely eliminated.

Routes to altruism

Altruism takes place when one individual, at a cost (C) to itself, exercises a strategy that provides a benefit (B) to another individual. The cost may consist of a loss of capability or resource which helps in the battle for survival and reproduction, or an added risk to its own survival. Altruism strategies can arise through:

Type Applies to: Situation Mathematical effect
Kin selection – (inclusive fitness of related contestants) Kin – genetically related individuals Evolutionary game participants are genes of strategy. The best payoff for an individual is not necessarily the best payoff for the gene. In any generation the player gene is not only in one individual, it is in a kin-group. The highest fitness payoff for the kin group is selected by natural selection. Therefore, strategies that include self-sacrifice on the part of individuals are often game winners – the evolutionarily stable strategy. Animals must live in kin-groups during part of the game for the opportunity for this altruistic sacrifice ever to take place. Games must take into account inclusive fitness. Fitness function is the combined fitness of a group of related contestants – each weighted by the degree of relatedness – relative to the total genetic population. The mathematical analysis of this gene-centric view of the game leads to Hamilton's rule, that the relatedness of the altruistic donor must exceed the cost-benefit ratio of the altruistic act itself:
R>c/b R is relatedness, c the cost, b the benefit
Direct reciprocity Contestants that trade favours in paired relationships A game theoretic embodiment of "I'll scratch your back if you scratch mine". A pair of individuals exchange favours in a multi-round game. The individuals are recognisable to one another as partnered. The term "direct" applies because the return favour is specifically given back to the pair partner only. The characteristics of the multi-round game produce a danger of defection and the potentially lesser payoffs of cooperation in each round, but any such defection can lead to punishment in a following round – establishing the game as a repeated prisoner's dilemma. Therefore, the family of tit-for-tat strategies come to the fore.
Indirect reciprocity Related or non related contestants trade favours but without partnering. A return favour is "implied" but with no specific identified source who is to give it. The return favour is not derived from any particular established partner. The potential for indirect reciprocity exists for a specific organism if it lives in a cluster of individuals who can interact over an extended period of time.

It has been argued that human behaviours in establishing moral systems as well as the expending of significant energies in human society for tracking individual reputations is a direct effect of societies' reliance on strategies of indirect reciprocation.

The game is highly susceptible to defection, as direct retaliation is impossible. Therefore, indirect reciprocity will not work without keeping a social score, a measure of past co-operative behaviour. The mathematics lead to a modified version of Hamilton's rule where:
q>c/b where q (the probability of knowing the social score) must be greater than the cost benefit ratio

Organisms that use social score are termed Discriminators, and require a higher level of cognition than strategies of simple direct reciprocity. As evolutionary biologist David Haig put it – "For direct reciprocity you need a face; for indirect reciprocity you need a name".

The evolutionarily stable strategy

The payoff matrix for the hawk dove game, with the addition of the assessor strategy. This "studies its opponent", behaving as a hawk when matched with an opponent it judges "weaker", like a dove when the opponent seems bigger and stronger. Assessor is an ESS, since it can invade both hawk and dove populations, and can withstand invasion by either hawk or dove mutants.

The evolutionarily stable strategy (ESS) is akin to the Nash equilibrium in classical game theory, but with mathematically extended criteria. Nash equilibrium is a game equilibrium where it is not rational for any player to deviate from their present strategy, provided that the others adhere to their strategies. An ESS is a state of game dynamics where, in a very large population of competitors, another mutant strategy cannot successfully enter the population to disturb the existing dynamic (which itself depends on the population mix). Therefore, a successful strategy (with an ESS) must be both effective against competitors when it is rare – to enter the previous competing population, and successful when later in high proportion in the population – to defend itself. This in turn means that the strategy must be successful when it contends with others exactly like itself.

An ESS is not:

  • An optimal strategy: that would maximize fitness, and many ESS states are far below the maximum fitness achievable in a fitness landscape. (See hawk dove graph above as an example of this.)
  • A singular solution: often several ESS conditions can exist in a competitive situation. A particular contest might stabilize into any one of these possibilities, but later a major perturbation in conditions can move the solution into one of the alternative ESS states.
  • Always present: it is possible for there to be no ESS. An evolutionary game with no ESS is "rock-scissors-paper", as found in species such as the side-blotched lizard (Uta stansburiana).
  • An unbeatable strategy: the ESS is only an uninvadeable strategy.
Female funnel web spiders (Agelenopsis aperta) contest with one another for the possession of their desert spider webs using the assessor strategy.

The ESS state can be solved for by exploring either the dynamics of population change to determine an ESS, or by solving equations for the stable stationary point conditions which define an ESS. For example, in the hawk dove game we can look for whether there is a static population mix condition where the fitness of doves will be exactly the same as fitness of hawks (therefore both having equivalent growth rates – a static point).

Let the chance of meeting a hawk=p so therefore the chance of meeting a dove is (1-p)

Let Whawk equal the payoff for hawk.....

Whawk=payoff in the chance of meeting a dove + payoff in the chance of meeting a hawk

Taking the payoff matrix results and plugging them into the above equation:

Whawk= V·(1-p)+(V/2-C/2)·p

Similarly for a dove:

Wdove= V/2·(1-p)+0·(p)

so....

Wdove= V/2·(1-p)

Equating the two fitnesses, hawk and dove

V·(1-p)+(V/2-C/2)·p= V/2·(1-p)

... and solving for p

p= V/C

so for this "static point" where the population percent is an ESS solves to be ESS(percent Hawk)=V/C

Similarly, using inequalities, it can be shown that an additional hawk or dove mutant entering this ESS state eventually results in less fitness for their kind – both a true Nash and an ESS equilibrium. This example shows that when the risks of contest injury or death (the cost C) is significantly greater than the potential reward (the benefit value V), the stable population will be mixed between aggressors and doves, and the proportion of doves will exceed that of the aggressors. This explains behaviours observed in nature.

Unstable games, cyclic patterns

Rock paper scissors

Rock paper scissors
Mutant invasion for rock paper scissors payoff matrix – an endless cycle
A computer simulation of the rock scissors paper game. The associated RPS game payoff matrix is shown. Starting with an arbitrary population the percentage of the three morphs builds up into a continuously cycling pattern.

Rock paper scissors incorporated into an evolutionary game has been used for modelling natural processes in the study of ecology. Using experimental economics methods, scientists have used RPS games to test human social evolutionary dynamical behaviours in laboratories. The social cyclic behaviours, predicted by evolutionary game theory, have been observed in various laboratory experiments.

Side-blotched lizard plays the RPS, and other cyclical games

The first example of RPS in nature was seen in the behaviours and throat colours of a small lizard of western North America. The side-blotched lizard (Uta stansburiana) is polymorphic with three throat-colour morphs that each pursue a different mating strategy

The side-blotched lizard effectively uses a rock-paper-scissors mating strategy
  • The orange throat is very aggressive and operates over a large territory – attempting to mate with numerous females within this larger area
  • The unaggressive yellow throat mimics the markings and behavior of female lizards, and "sneakily" slips into the orange throat's territory to mate with the females there (thereby taking over the population)
  • The blue throat mates with, and carefully guards, one female – making it impossible for the sneakers to succeed and therefore overtakes their place in a population

However the blue throats cannot overcome the more aggressive orange throats. Later work showed that the blue males are altruistic to other blue males, with three key traits: they signal with blue color, they recognize and settle next to other (unrelated) blue males, and they will even defend their partner against orange, to the death. This is the hallmark of another game of cooperation that involves a green-beard effect.

The females in the same population have the same throat colours, and this affects how many offspring they produce and the size of the progeny, which generates cycles in density, yet another game - the r-K game. Here, r is the Malthusian parameter governing exponential growth, and K is the carrying capacity of the population. Orange females have larger clutches and smaller offspring and do well at low density. Yellow females (and blue) have smaller clutches and larger offspring and do better when the population exceeds carrying capacity and the population crashes to low density. The orange then takes over and this generates perpetual cycles of orange and yellow tightly tied to population density. The idea of cycles due to density regulation of two strategies originated with Dennis Chitty, who worked on rodents, ergo these kinds of games lead to "Chitty cycles". There are games within games within games embedded in natural populations. These drive RPS cycles in the males with a periodicity of four years and r-K cycles in females with a periodicity of two years.

The overall situation corresponds to the rock, scissors, paper game, creating a four-year population cycle. The RPS game in male side-blotched lizards does not have an ESS, but it has a Nash equilibrium (NE) with endless orbits around the NE attractor. Since that time many other three-strategy polymorphisms have been discovered in lizards and some of these have RPS dynamics merging the male game and density regulation game in a single sex (males). More recently, mammals have been shown to harbour the same RPS game in males and r-K game in females, with coat-colour polymorphisms and behaviours that drive cycles. This game is also linked to the evolution of male care in rodents, and monogamy, and drives speciation rates. There are r-K strategy games linked to rodent population cycles (and lizard cycles).

When he read that these lizards were essentially engaged in a game with a rock-paper-scissors structure, John Maynard Smith is said to have exclaimed "They have read my book!".

Signalling, sexual selection and the handicap principle

The peacock's tail may be an instance of the handicap principle in action

Aside from the difficulty of explaining how altruism exists in many evolved organisms, Darwin was also bothered by a second conundrum – why a significant number of species have phenotypical attributes that are patently disadvantageous to them with respect to their survival – and should by the process of natural section be selected against – e.g. the massive inconvenient feather structure found in a peacock's tail. Regarding this issue Darwin wrote to a colleague "The sight of a feather in a peacock's tail, whenever I gaze at it, makes me sick." It is the mathematics of evolutionary game theory, which has not only explained the existence of altruism, but also explains the totally counterintuitive existence of the peacock's tail and other such biological encumbrances.

On analysis, problems of biological life are not at all unlike the problems that define economics – eating (akin to resource acquisition and management), survival (competitive strategy) and reproduction (investment, risk and return). Game theory was originally conceived as a mathematical analysis of economic processes and indeed this is why it has proven so useful in explaining so many biological behaviours. One important further refinement of the evolutionary game theory model that has economic overtones rests on the analysis of costs. A simple model of cost assumes that all competitors suffer the same penalty imposed by the game costs, but this is not the case. More successful players will be endowed with or will have accumulated a higher "wealth reserve" or "affordability" than less-successful players. This wealth effect in evolutionary game theory is represented mathematically by "resource holding potential (RHP)" and shows that the effective cost to a competitor with a higher RHP are not as great as for a competitor with a lower RHP. As a higher RHP individual is a more desirable mate in producing potentially successful offspring, it is only logical that with sexual selection RHP should have evolved to be signalled in some way by the competing rivals, and for this to work this signalling must be done honestly. Amotz Zahavi has developed this thinking in what is known as the "handicap principle", where superior competitors signal their superiority by a costly display. As higher RHP individuals can properly afford such a costly display this signalling is inherently honest, and can be taken as such by the signal receiver. In nature this is illustrated than in the costly plumage of the peacock. The mathematical proof of the handicap principle was developed by Alan Grafen using evolutionary game-theoretic modelling.

Coevolution

Two types of dynamics:

  • Evolutionary games which lead to a stable situation or point of stasis for contending strategies which result in an evolutionarily stable strategy
  • Evolutionary games which exhibit a cyclic behaviour (as with RPS game) where the proportions of contending strategies continuously cycle over time within the overall population
Competitive Coevolution - The rough-skinned newt (Tarricha granulosa) is highly toxic, due to an evolutionary arms race with a predator, the common garter snake (Thamnophis sirtalis), which in turn is highly tolerant of the poison. The two are locked in a Red Queen arms race.
Mutualistic Coevolution - Darwin's orchid (Angraecum sesquipedale) and the moth Morgan's sphinx (Xanthopan morgani) have a mutual relationship where the moth gains pollen and the flower is pollinated.

A third, coevolutionary, dynamic, combines intra-specific and inter-specific competition. Examples include predator-prey competition and host-parasite co-evolution, as well as mutualism. Evolutionary game models have been created for pairwise and multi-species coevolutionary systems. The general dynamic differs between competitive systems and mutualistic systems.

In competitive (non-mutualistic) inter-species coevolutionary system the species are involved in an arms race – where adaptations that are better at competing against the other species tend to be preserved. Both game payoffs and replicator dynamics reflect this. This leads to a Red Queen dynamic where the protagonists must "run as fast as they can to just stay in one place".

A number of evolutionary game theory models have been produced to encompass coevolutionary situations. A key factor applicable in these coevolutionary systems is the continuous adaptation of strategy in such arms races. Coevolutionary modelling therefore often includes genetic algorithms to reflect mutational effects, while computers simulate the dynamics of the overall coevolutionary game. The resulting dynamics are studied as various parameters are modified. Because several variables are simultaneously at play, solutions become the province of multi-variable optimisation. The mathematical criteria of determining stable points are Pareto efficiency and Pareto dominance, a measure of solution optimality peaks in multivariable systems.

Carl Bergstrom and Michael Lachmann apply evolutionary game theory to the division of benefits in mutualistic interactions between organisms. Darwinian assumptions about fitness are modeled using replicator dynamics to show that the organism evolving at a slower rate in a mutualistic relationship gains a disproportionately high share of the benefits or payoffs.

Extending the model

A mathematical model analysing the behaviour of a system needs initially to be as simple as possible to aid in developing a base understanding the fundamentals, or “first order effects”, pertaining to what is being studied. With this understanding in place it is then appropriate to see if other, more subtle, parameters (second order effects) further impact the primary behaviours or shape additional behaviours in the system. Following Maynard Smith's seminal work in evolutionary game theory, the subject has had a number of very significant extensions which have shed more light on understanding evolutionary dynamics, particularly in the area of altruistic behaviors. Some of these key extensions to evolutionary game theory are:

A Spatial Game
In a spatial evolutionary game contestants meet in contests at fixed grid positions and only interact with immediate neighbors. Shown here are the dynamics of a Hawk Dove contest, showing Hawk and Dove contestants as well as the changes of strategy taking place in the various cells

Spatial Games

Geographic factors in evolution include gene flow and horizontal gene transfer. Spatial game models represent geometry by putting contestants in a lattice of cells: contests take place only with immediate neighbours. Winning strategies take over these immediate neighbourhoods and then interact with adjacent neighbourhoods. This model is useful in showing how pockets of co-operators can invade and introduce altruism in the Prisoners Dilemma game, where Tit for Tat (TFT) is a Nash Equilibrium but NOT also an ESS. Spatial structure is sometimes abstracted into a general network of interactions. This is the foundation of evolutionary graph theory.

Effects of having information

In evolutionary game theory as in conventional Game Theory the effect of Signalling (the acquisition of information) is of critical importance, as in Indirect Reciprocity in Prisoners Dilemma (where contests between the SAME paired individuals are NOT repetitive). This models the reality of most normal social interactions which are non-kin related. Unless a probability measure of reputation is available in Prisoners Dilemma only direct reciprocity can be achieved. With this information indirect reciprocity is also supported.

Alternatively, agents might have access to an arbitrary signal initially uncorrelated to strategy but becomes correlated due to evolutionary dynamics. This is the green-beard effect (see side-blotched lizards, above) or evolution of ethnocentrism in humans. Depending on the game, it can allow the evolution of either cooperation or irrational hostility.

From molecular to multicellular level, a signaling game model with information asymmetry between sender and receiver might be appropriate, such as in mate attraction or evolution of translation machinery from RNA strings.

Finite populations

Many evolutionary games have been modelled in finite populations to see the effect this may have, for example in the success of mixed strategies.

Geodesic

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Geodesic   Klein quartic with 28 geodesics...