A one-dimensional cyclic cellular automaton with n = 4, run for 300 steps from a random initial configuration.
A cyclic cellular automaton is a kind of cellular automaton rule developed by David Griffeath
and studied by several other cellular automaton researchers. In this
system, each cell remains unchanged until some neighboring cell has a modular
value exactly one unit larger than that of the cell itself, at which
point it copies its neighbor's value. One-dimensional cyclic cellular
automata can be interpreted as systems of interacting particles, while
cyclic cellular automata in higher dimensions exhibit complex spiraling
behavior.
Rules
As with any
cellular automaton, the cyclic cellular automaton consists of a regular
grid of cells in one or more dimensions. The cells can take on any of states, ranging from to .
The first generation starts out with random states in each of the
cells. In each subsequent generation, if a cell has a neighboring cell
whose value is the successor of the cell's value, the cell is "consumed"
and takes on the succeeding value. (Note that is the successor of ; see also modular arithmetic.) More general forms of this type of rule also include a threshold parameter, and only allow a cell to be consumed when the number of neighbors with the successor value exceeds this threshold.
One dimension
The one-dimensional cyclic cellular automaton has been extensively studied by Robert Fisch, a student of Griffeath.
Starting from a random configuration with n = 3 or n = 4,
this type of rule can produce a pattern which, when presented as a
time-space diagram, shows growing triangles of values competing for
larger regions of the grid.
The boundaries between these regions can be viewed as moving
particles which collide and interact with each other. In the three-state
cyclic cellular automaton, the boundary between regions with values i and i + 1 (mod n)
can be viewed as a particle that moves either leftwards or rightwards
depending on the ordering of the regions; when a leftward-moving
particle collides with a rightward-moving one, they annihilate each other, leaving two fewer particles in the system. This type of ballistic annihilation process occurs in several other cellular automata and related systems, including Rule 184, a cellular automaton used to model traffic flow.
In the n = 4 automaton, the same two types of particles
and the same annihilation reaction occur. Additionally, a boundary
between regions with values i and i + 2 (mod n) can
be viewed as a third type of particle, that remains stationary. A
collision between a moving and a stationary particle results in a single
moving particle moving in the opposite direction.
However, for n ≥ 5, random initial configurations tend to
stabilize quickly rather than forming any non-trivial long-range
dynamics. Griffeath has nicknamed this dichotomy between the long-range
particle dynamics of the n = 3 and n = 4 automata on the one hand, and the static behavior of the n ≥ 5 automata on the other hand, "Bob's dilemma", after Bob Fisch.
Two or more dimensions
A two-dimensional cyclic cellular automaton with n = 16, for 1300 steps starting from a random initial configuration.
In two dimensions, with no threshold and the von Neumann neighborhood or Moore neighborhood,
this cellular automaton generates three general types of patterns
sequentially, from random initial conditions on sufficiently large
grids, regardless of n.
At first, the field is purely random. As cells consume their neighbors
and get within range to be consumed by higher-ranking cells, the
automaton goes to the consuming phase, where there are blocks of color
advancing against remaining blocks of randomness. Important in further
development are objects called demons, which are cycles of adjacent
cells containing one cell of each state, in the cyclic order; these
cycles continuously rotate and generate waves that spread out in a spiral
pattern centered at the cells of the demon. The third stage, the demon
stage, is dominated by these cycles. The demons with shorter cycles
consume demons with longer cycles until, almost surely, every cell of the automaton eventually enters a repeating cycle of states, where the period of the repetition is either n or (for automata with n odd and the von Neumann neighborhood) n
+ 1. The same eventually-periodic behavior occurs also in higher
dimensions. Small structures can also be constructed with any even
period between n and 3n/2. Merging these structures, configurations can be built with a global super-polynomial period.
For larger neighborhoods, similar spiraling behavior occurs for
low thresholds, but for sufficiently high thresholds the automaton
stabilizes in the block of color stage without forming spirals. At
intermediate values of the threshold, a complex mix of color blocks and
partial spirals, called turbulence, can form.
For appropriate choices of the number of states and the size of the
neighborhood, the spiral patterns formed by this automaton can be made
to resemble those of the Belousov–Zhabotinsky reaction in chemistry, or other systems of autowaves, although other cellular automata more accurately model the excitable medium that leads to this reaction.
The term gambler's ruin is a statistical concept expressed in a variety of forms:
The original meaning is that a persistent gambler
who raises his bet to a fixed fraction of bankroll when he wins, but
does not reduce it when he loses, will eventually and inevitably go
broke, even if he has a positive expected value on each bet.
Another common meaning is that a persistent gambler with finite
wealth, playing a fair game (that is, each bet has expected value zero
to both sides) will eventually and inevitably go broke against an
opponent with infinite wealth. Such a situation can be modeled by a random walk
on the real number line. In that context it is provable that the agent
will return to his point of origin or go broke and is ruined an
infinite number of times if the random walk continues forever.
The result above is a corollary of a general theorem by Christiaan Huygens
which is also known as gambler's ruin. That theorem shows how to
compute the probability of each player winning a series of bets that
continues until one's entire initial stake is lost, given the initial
stakes of the two players and the constant probability of winning. This
is the oldest mathematical idea that goes by the name gambler's ruin, but not the first idea to which the name was applied.
The most common use of the term today is that a gambler playing a
negative expected value game will eventually go broke, regardless of
betting system. This is another corollary to Huygens's result.
The concept may be stated as an ironic paradox:
Persistently taking beneficial chances is never beneficial at the end.
This paradoxical form of gambler's ruin should not be confused with the
gambler's fallacy, a different concept.
The concept has specific relevance for gamblers; however it also leads to mathematical theorems with wide application and many related results in probability and statistics. Huygens's result in particular led to important advances in the mathematical theory of probability.
History
The earliest known mention of the gambler's ruin problem is a letter from Blaise Pascal to Pierre Fermat in 1656 (two years after the more famous correspondence on the problem of points). Pascal's version was summarized in a 1656 letter from Pierre de Carcavi to Huygens:
Let two men play with three dice, the first player
scoring a point whenever 11 is thrown, and the second whenever 14 is
thrown. But instead of the points accumulating in the ordinary way, let a
point be added to a player's score only if his opponent's score is nil,
but otherwise let it be subtracted from his opponent's score. It is as
if opposing points form pairs, and annihilate each other, so that the
trailing player always has zero points. The winner is the first to reach
twelve points; what are the relative chances of each player winning?
Huygens reformulated the problem and published it in De ratiociniis in ludo aleae ("On Reasoning in Games of Chance", 1657):
Problem (2-1) Each player starts with 12 points, and a
successful roll of the three dice for a player (getting an 11 for the
first player or a 14 for the second) adds one to that player's score and
subtracts one from the other player's score; the loser of the game is
the first to reach zero points. What is the probability of victory for
each player?
This is the classic gambler's ruin formulation: two players begin with
fixed stakes, transferring points until one or the other is "ruined" by
getting to zero points. However, the term "gambler's ruin" was not
applied until many years later.
Reasons for the four results
Let "bankroll" be the amount of money a gambler has at his disposal at any moment, and let N be any positive integer. Suppose that he raises his stake to
when he wins, but does not reduce his stake when he loses. This general
pattern is not uncommon among real gamblers, and casinos encourage it
by "chipping up" winners (giving them higher denomination chips). Under this betting scheme, it will take at most N
losing bets in a row to bankrupt him. If his probability of winning
each bet is less than 1 (if it is 1, then he is no gambler), he will
eventually lose N bets in a row, however big N is. It is
not necessary that he follow the precise rule, just that he increase his
bet fast enough as he wins. This is true even if the expected value of
each bet is positive.
The gambler playing a fair game (with 0.5 probability of winning)
will eventually either go broke or double his wealth. Let's define that
the game ends upon either event. These events are equally likely, or
the game would not be fair. So he has a 0.5 chance of going broke before
doubling his money. Given he doubles his money, a new game begins and
he again has a 0.5 chance of doubling his money before going broke.
After the second game there is a 1/2 x 1/2 chance that he has not gone
broke in the first and second games. Continuing this way, his chance of
not going broke after n successive games is 1/2 x 1/2 x 1/2 x . . .
1/2^n which approaches 0. His chance of going broke after n successive
games is 0.5 + 0.25 + 0.125 + . . . 1 - 1/2^n which approaches 1.
Huygens's result is illustrated in the next section.
The eventual fate of a player at a negative expected value game cannot be better than the player at a fair game, so he will go broke as well.
Example of Huygens's result
Fair coin flipping
Consider
a coin-flipping game with two players where each player has a 50%
chance of winning with each flip of the coin. After each flip of the
coin the loser transfers one penny to the winner. The game ends when one
player has all the pennies.
If there are no other limitations on the number of flips, the
probability that the game will eventually end this way is 1. (One way to
see this is as follows. Any given finite string of heads and tails will
eventually be flipped with certainty: the probability of not seeing
this string, while high at first, decays exponentially. In particular,
the players would eventually flip a string of heads as long as the total
number of pennies in play, by which time the game must have already
ended.)
If player one has n1 pennies and player two n2 pennies, the probabilities P1 and P2 that players one and two, respectively, will end penniless are:
Two examples of this are if one player has more pennies than the other; and if both players have the same number of pennies.
In the first case say player one has 8 pennies and player two () were to have 5 pennies then the probability of each losing is:
It follows that even with equal odds of winning the player that starts with fewer pennies is more likely to fail.
In the second case where both players have the same number of pennies (in this case 6) the likelihood of each losing is:
Unfair coin flipping
In the event of an unfair coin, where player one wins each toss with probability p, and player two wins with probability q = 1 − p, then the probability of each ending penniless is:
This can be shown as follows: Consider the probability of player 1 experiencing gamblers ruin having started with amount of money, . Then, using the Law of Total Probability, we have
where W denotes the event that player 1 wins the first bet. Then clearly and . Also is the probability that player 1 experiences gambler's ruin having started with amount of money: ; and is the probability that player 1 experiences gambler's ruin having started with amount of money: .
Denoting , we get the linear homogeneous recurrence relation
which we can solve using the fact that (i.e. the probability of gambler's ruin given that player 1 starts with no money is 1), and
(i.e. the probability of gambler's ruin given that player 1 starts with
all the money is 0.) For a more detailed description of the method see
e.g. Feller (1970), An introduction to probability theory and its applications, 3rd ed.
N-player ruin problem
The above described problem (2 players) is a special case of the so-called N-Player ruin problem.
Here players with initial capital
dollars, respectively,
play a sequence of (arbitrary) independent games and win and lose
certain amounts of dollars from/to each other according to fixed rules.
The sequence of games ends as soon as at least one player is ruined.
Standard Markov chain methods can be applied to
solve in principle this more general problem, but the computations quickly become prohibitive as soon as the number of players
or their initial capital increase. For and large initial capitals
the solution can be well approximated by using two-dimensional Brownian motion. (For this is not possible.)
In practice the true problem is to find the solution for the typical cases of and limited initial capital.
Swan (2006) proposed an algorithm based on Matrix-analytic methods (Folding algorithm for ruin problems) which significantly
reduces the order of the computational task in such cases.
The word probability has been used in a variety of ways since it was first applied to the mathematical study of games of chance.
Does probability measure the real, physical tendency of something to
occur or is it a measure of how strongly one believes it will occur, or
does it draw on both these elements? In answering such questions,
mathematicians interpret the probability values of probability theory.
There are two broad categories of probability interpretations which can be called "physical" and "evidential" probabilities. Physical probabilities, which are also called objective or frequency probabilities,
are associated with random physical systems such as roulette wheels,
rolling dice and radioactive atoms. In such systems, a given type of
event (such as a die yielding a six) tends to occur at a persistent
rate, or "relative frequency", in a long run of trials. Physical
probabilities either explain, or are invoked to explain, these stable
frequencies. The two main kinds of theory of physical probability are frequentist accounts (such as those of Venn, Reichenbach and von Mises) and propensity accounts (such as those of Popper, Miller, Giere and Fetzer).
Evidential probability, also called Bayesian probability,
can be assigned to any statement whatsoever, even when no random
process is involved, as a way to represent its subjective plausibility,
or the degree to which the statement is supported by the available
evidence. On most accounts, evidential probabilities are considered to
be degrees of belief, defined in terms of dispositions to gamble at
certain odds. The four main evidential interpretations are the
classical (e.g. Laplace's) interpretation, the subjective interpretation (de Finetti and Savage), the epistemic or inductive interpretation (Ramsey, Cox) and the logical interpretation (Keynes and Carnap).
There are also evidential interpretations of probability covering
groups, which are often labelled as 'intersubjective' (proposed by Gillies and Rowbottom).
Some interpretations of probability are associated with approaches to statistical inference, including theories of estimation and hypothesis testing. The physical interpretation, for example, is taken by followers of "frequentist" statistical methods, such as Ronald Fisher, Jerzy Neyman and Egon Pearson. Statisticians of the opposing Bayesian
school typically accept the existence and importance of physical
probabilities, but also consider the calculation of evidential
probabilities to be both valid and necessary in statistics. This
article, however, focuses on the interpretations of probability rather
than theories of statistical inference.
The terminology of this topic is rather confusing, in part
because probabilities are studied within a variety of academic fields.
The word "frequentist" is especially tricky. To philosophers it refers
to a particular theory of physical probability, one that has more or
less been abandoned. To scientists, on the other hand, "frequentist probability"
is just another name for physical (or objective) probability. Those who
promote Bayesian inference view "frequentist statistics" as an approach
to statistical inference that recognises only physical probabilities.
Also the word "objective", as applied to probability, sometimes means
exactly what "physical" means here, but is also used of evidential
probabilities that are fixed by rational constraints, such as logical
and epistemic probabilities.
It is unanimously agreed that
statistics depends somehow on probability. But, as to what probability
is and how it is connected with statistics, there has seldom been such
complete disagreement and breakdown of communication since the Tower of
Babel. Doubtless, much of the disagreement is merely terminological and
would disappear under sufficiently sharp analysis.
— (Savage, 1954, p 2)
Philosophy
The philosophy of probability presents problems chiefly in matters of epistemology and the uneasy interface between mathematical concepts and ordinary language as it is used by non-mathematicians.
Probability theory is an established field of study in mathematics. It has its origins in correspondence discussing the mathematics of games of chance between Blaise Pascal and Pierre de Fermat in the seventeenth century, and was formalized and rendered axiomatic as a distinct branch of mathematics by Andrey Kolmogorov
in the twentieth century. In axiomatic form, mathematical statements
about probability theory carry the same sort of epistemological
confidence within the philosophy of mathematics as are shared by other mathematical statements.
The mathematical analysis originated in observations of the behaviour of game equipment such as playing cards and dice, which are designed specifically to introduce random and equalized elements; in mathematical terms, they are subjects of indifference. This is not the only way probabilistic statements are used in ordinary human language: when people say that "it will probably rain",
they typically do not mean that the outcome of rain versus not-rain is a
random factor that the odds currently favor; instead, such statements
are perhaps better understood as qualifying their expectation of rain
with a degree of confidence. Likewise, when it is written that "the
most probable explanation" of the name of Ludlow, Massachusetts "is that it was named after Roger Ludlow",
what is meant here is not that Roger Ludlow is favored by a random
factor, but rather that this is the most plausible explanation of the
evidence, which admits other, less likely explanations.
Thomas Bayes attempted to provide a logic that could handle varying degrees of confidence; as such, Bayesian probability
is an attempt to recast the representation of probabilistic statements
as an expression of the degree of confidence by which the beliefs they
express are held.
The first attempt at mathematical rigour in the field of probability, championed by Pierre-Simon Laplace, is now known as the classical definition. Developed from studies of games of chance (such as rolling dice)
it states that probability is shared equally between all the possible
outcomes, provided these outcomes can be deemed equally likely.
The theory of chance consists in
reducing all the events of the same kind to a certain number of cases
equally possible, that is to say, to such as we may be equally undecided
about in regard to their existence, and in determining the number of
cases favorable to the event whose probability is sought. The ratio of
this number to that of all the cases possible is the measure of this
probability, which is thus simply a fraction whose numerator is the
number of favorable cases and whose denominator is the number of all the
cases possible.
— Pierre-Simon Laplace, A Philosophical Essay on Probabilities
The classical definition of probability works well for situations with only a finite number of equally-likely outcomes.
This can be represented mathematically as follows:
If a random experiment can result in N mutually exclusive and equally likely outcomes and if NA of these outcomes result in the occurrence of the event A, the probability of A is defined by
There are two clear limitations to the classical definition.
Firstly, it is applicable only to situations in which there is only a
'finite' number of possible outcomes. But some important random
experiments, such as tossing a coin until it rises heads, give rise to an infinite
set of outcomes. And secondly, you need to determine in advance that
all the possible outcomes are equally likely without relying on the
notion of probability to avoid circularity—for instance, by symmetry
considerations.
Frequentism
For
frequentists, the probability of the ball landing in any pocket can be
determined only by repeated trials in which the observed result
converges to the underlying probability in the long run.
Frequentists posit that the probability of an event is its relative frequency over time,
(3.4) i.e., its relative frequency of occurrence after repeating a
process a large number of times under similar conditions. This is also
known as aleatory probability. The events are assumed to be governed by some random physical phenomena, which are either phenomena that are predictable, in principle, with sufficient information (see determinism); or phenomena which are essentially unpredictable. Examples of the first kind include tossing dice or spinning a roulette wheel; an example of the second kind is radioactive decay.
In the case of tossing a fair coin, frequentists say that the
probability of getting a heads is 1/2, not because there are two equally
likely outcomes but because repeated series of large numbers of trials
demonstrate that the empirical frequency converges to the limit 1/2 as
the number of trials goes to infinity.
If we denote by the number of occurrences of an event in trials, then if we say that .
The frequentist view has its own problems. It is of course
impossible to actually perform an infinity of repetitions of a random
experiment to determine the probability of an event. But if only a
finite number of repetitions of the process are performed, different
relative frequencies will appear in different series of trials. If these
relative frequencies are to define the probability, the probability
will be slightly different every time it is measured. But the real
probability should be the same every time. If we acknowledge the fact
that we only can measure a probability with some error of measurement
attached, we still get into problems as the error of measurement can
only be expressed as a probability, the very concept we are trying to
define. This renders even the frequency definition circular; see for
example “What is the Chance of an Earthquake?”
Subjectivism
Gambling odds reflect the average bettor's 'degree of belief' in the outcome.
Subjectivists, also known as Bayesians or followers of epistemic probability,
give the notion of probability a subjective status by regarding it as a
measure of the 'degree of belief' of the individual assessing the
uncertainty of a particular situation. Epistemic or subjective probability is sometimes called credence, as opposed to the term chance for a propensity probability.
Some examples of epistemic probability are to assign a
probability to the proposition that a proposed law of physics is true,
and to determine how probable it is that a suspect committed a crime,
based on the evidence presented.
Gambling odds don't reflect the bookies' belief in a likely
winner, so much as the other bettors' belief, because the bettors are
actually betting against one another. The odds are set based on how
many people have bet on a possible winner, so that even if the high odds
players always win, the bookies will always make their percentages
anyway.
The use of Bayesian probability raises the philosophical debate as to whether it can contribute valid justifications of belief.
Bayesians point to the work of Ramsey (p 182) and de Finetti (p 103) as proving that subjective beliefs must follow the laws of probability if they are to be coherent. Evidence casts doubt that humans will have coherent beliefs.
The use of Bayesian probability involves specifying a prior probability.
This may be obtained from consideration of whether the required prior
probability is greater or lesser than a reference probability associated with an urn model or a thought experiment.
The issue is that for a given problem, multiple thought experiments
could apply, and choosing one is a matter of judgement: different people
may assign different prior probabilities, known as the reference class problem.
The "sunrise problem" provides an example.
Propensity
Propensity theorists think of probability as a physical propensity,
or disposition, or tendency of a given type of physical situation to
yield an outcome of a certain kind or to yield a long run relative
frequency of such an outcome. This kind of objective probability is sometimes called 'chance'.
Propensities, or chances, are not relative frequencies, but
purported causes of the observed stable relative frequencies.
Propensities are invoked to explain why repeating a certain kind of
experiment will generate given outcome types at persistent rates, which
are known as propensities or chances. Frequentists are unable to take
this approach, since relative frequencies do not exist for single tosses
of a coin, but only for large ensembles or collectives (see "single
case possible" in the table above). In contrast, a propensitist is able to use the law of large numbers
to explain the behaviour of long-run frequencies. This law, which is a
consequence of the axioms of probability, says that if (for example) a
coin is tossed repeatedly many times, in such a way that its probability
of landing heads is the same on each toss, and the outcomes are
probabilistically independent, then the relative frequency of heads will
be close to the probability of heads on each single toss. This law
allows that stable long-run frequencies are a manifestation of invariant
single-case probabilities. In addition to explaining the
emergence of stable relative frequencies, the idea of propensity is
motivated by the desire to make sense of single-case probability
attributions in quantum mechanics, such as the probability of decay of a particular atom at a particular time.
The main challenge facing propensity theories is to say exactly
what propensity means. (And then, of course, to show that propensity
thus defined has the required properties.) At present, unfortunately,
none of the well-recognised accounts of propensity comes close to
meeting this challenge.
A propensity theory of probability was given by Charles Sanders Peirce. A later propensity theory was proposed by philosopher Karl Popper, who had only slight acquaintance with the writings of C. S. Peirce, however.
Popper noted that the outcome of a physical experiment is produced by a
certain set of "generating conditions". When we repeat an experiment,
as the saying goes, we really perform another experiment with a (more or
less) similar set of generating conditions. To say that a set of
generating conditions has propensity p of producing the outcome E means that those exact conditions, if repeated indefinitely, would produce an outcome sequence in which E occurred with limiting relative frequency p.
For Popper then, a deterministic experiment would have propensity 0 or
1 for each outcome, since those generating conditions would have same
outcome on each trial. In other words, non-trivial propensities (those
that differ from 0 and 1) only exist for genuinely nondeterministic
experiments.
A number of other philosophers, including David Miller and Donald A. Gillies, have proposed propensity theories somewhat similar to Popper's.
Other propensity theorists (e.g. Ronald Giere)
do not explicitly define propensities at all, but rather see propensity
as defined by the theoretical role it plays in science. They argued,
for example, that physical magnitudes such as electrical charge
cannot be explicitly defined either, in terms of more basic things, but
only in terms of what they do (such as attracting and repelling other
electrical charges). In a similar way, propensity is whatever fills the
various roles that physical probability plays in science.
What roles does physical probability play in science? What are
its properties? One central property of chance is that, when known, it
constrains rational belief to take the same numerical value. David
Lewis called this the Principal Principle,
(3.3 & 3.5) a term that philosophers have mostly adopted. For
example, suppose you are certain that a particular biased coin has
propensity 0.32 to land heads every time it is tossed. What is then the
correct price for a gamble that pays $1 if the coin lands heads, and
nothing otherwise? According to the Principal Principle, the fair price
is 32 cents.
Logical, epistemic, and inductive probability
It is widely recognized that the term "probability" is sometimes used
in contexts where it has nothing to do with physical randomness.
Consider, for example, the claim that the extinction of the dinosaurs
was probably caused by a large meteorite hitting the earth.
Statements such as "Hypothesis H is probably true" have been interpreted
to mean that the (presently available) empirical evidence (E, say) supports H to a high degree. This degree of support of H by E has been called the logical probability of H given E, or the epistemic probability of H given E, or the inductive probability of H given E.
The differences between these interpretations are rather small,
and may seem inconsequential. One of the main points of disagreement
lies in the relation between probability and belief. Logical
probabilities are conceived (for example in Keynes' Treatise on Probability)
to be objective, logical relations between propositions (or sentences),
and hence not to depend in any way upon belief. They are degrees of
(partial) entailment, or degrees of logical consequence, not degrees of belief. (They do, nevertheless, dictate proper degrees of belief, as is discussed below.) Frank P. Ramsey,
on the other hand, was skeptical about the existence of such objective
logical relations and argued that (evidential) probability is "the logic
of partial belief". (p 157) In other words, Ramsey held that epistemic probabilities simply are degrees of rational belief, rather than being logical relations that merely constrain degrees of rational belief.
Another point of disagreement concerns the uniqueness of evidential probability, relative to a given state of knowledge. Rudolf Carnap
held, for example, that logical principles always determine a unique
logical probability for any statement, relative to any body of evidence.
Ramsey, by contrast, thought that while degrees of belief are subject
to some rational constraints (such as, but not limited to, the axioms of
probability) these constraints usually do not determine a unique value.
Rational people, in other words, may differ somewhat in their degrees
of belief, even if they all have the same information.
Prediction
An alternative account of probability emphasizes the role of prediction
– predicting future observations on the basis of past observations, not
on unobservable parameters. In its modern form, it is mainly in the
Bayesian vein. This was the main function of probability before the 20th
century,
but fell out of favor compared to the parametric approach, which
modeled phenomena as a physical system that was observed with error,
such as in celestial mechanics.
The modern predictive approach was pioneered by Bruno de Finetti, with the central idea of exchangeability – that future observations should behave like past observations. This view came to the attention of the Anglophone world with the 1974 translation of de Finetti's book, and has
since been propounded by such statisticians as Seymour Geisser.
Axiomatic probability
The
mathematics of probability can be developed on an entirely axiomatic
basis that is independent of any interpretation: see the articles on probability theory and probability axioms for a detailed treatment.