The commutator of two elements, g and h, of a groupG, is the element
[g, h] = g−1h−1gh.
It is equal to the group's identity if and only if g and h commute (i.e., if and only if gh = hg). The subgroup of Ggenerated by all commutators is called the derived group or the commutator subgroup of G.
Note that one must consider the subgroup generated by the set of
commutators because in general the set of commutators is not closed
under the group operation. Commutators are used to define nilpotent and solvable groups.
The above definition of the commutator is used by some group
theorists, as well as throughout this article. However, many other group
theorists define the commutator as
[g, h] = ghg−1h−1.
Identities (group theory)
Commutator identities are an important tool in group theory. The expression ax denotes the conjugate of a by x, defined as x−1ax.
and
and
and
Identity (5) is also known as the Hall–Witt identity, after Philip Hall and Ernst Witt. It is a group-theoretic analogue of the Jacobi identity for the ring-theoretic commutator (see next section).
N.B., the above definition of the conjugate of a by x is used by some group theorists. Many other group theorists define the conjugate of a by x as xax−1. This is often written . Similar identities hold for these conventions.
Many identities are used that are true modulo certain subgroups. These can be particularly useful in the study of solvable groups and nilpotent groups. For instance, in any group, second powers behave well:
It is zero if and only if a and b commute. In linear algebra,
if two endomorphisms of a space are represented by commuting matrices
in terms of one basis, then they are so represented in terms of every
basis. By using the commutator as a Lie bracket, every associative algebra can be turned into a Lie algebra.
The anticommutator of two elements a and b of a ring or an associative algebra is defined by
Sometimes the brackets [ ]+ are also used to denote anticommutators, while [ ]− is then used for commutators. The anticommutator is used less often than the commutator, but can be used for example to define Clifford algebras, Jordan algebras and is utilised to derive the Dirac equation in particle physics.
An additional identity may be found for this last expression, in the form:
If A is a fixed element of a ring R, the first additional identity can be interpreted as a Leibniz rule for the map given by . In other words, the map adA defines a derivation on the ring R.
The second and third identities represent Leibniz rules for more than
two factors that are valid for any derivation. Identities 4–6 can also
be interpreted as Leibniz rules for a certain derivation.
By contrast, it is not always an algebra homomorphism; it does not hold in general:
Examples
General Leibniz rule
The general Leibniz rule, expanding repeated derivatives of a product, can be written abstractly using the adjoint representation:
Replacing x by the differentiation operator , and y by the multiplication operator , we get , and applying both sides to a function g, the identity becomes the general Leibniz rule for .
The transactional interpretation of quantum mechanics (TIQM) takes the psi and psi* wave functions of the standard quantum formalism to be retarded (forward in time) and advanced (backward in time) waves that form a quantum interaction as a Wheeler–Feynman handshake or transaction. It was first proposed in 1986 by John G. Cramer,
who argues that it helps in developing intuition for quantum processes.
He also suggests that it avoids the philosophical problems with the Copenhagen interpretation and the role of the observer, and also resolves various quantum paradoxes. TIQM formed a minor plot point in his science fiction novel Einstein's Bridge.
In TIQM, the source emits a usual (retarded) wave forward in time, but it also emits an advanced wave
backward in time; furthermore, the receiver, who is later in time, also
emits an advanced wave backward in time and a retarded wave forward in
time. A quantum event occurs when a "handshake" exchange of advanced and
retarded waves triggers the formation of a transaction in which energy,
momentum, angular momentum, etc. are transferred. The quantum
mechanism behind transaction formation has been demonstrated explicitly
for the case of a photon transfer between atoms in Sect. 5.4 of Carver Mead's book Collective Electrodynamics. In this interpretation, the collapse of the wavefunction
does not happen at any specific point in time, but is "atemporal" and
occurs along the whole transaction, and the emission/absorption process
is time-symmetric. The waves are seen as physically real, rather than a
mere mathematical device to record the observer's knowledge as in some
other interpretations of quantum mechanics. Philosopher and writer Ruth Kastner
argues that the waves exist as possibilities outside of physical
spacetime and that therefore it is necessary to accept such
possibilities as part of reality.
TIQM is explicitly non-local and, as a consequence, logically consistent with counterfactual definiteness (CFD), the minimum realist assumption. As such it incorporates the non-locality demonstrated by the Bell test experiments and eliminates the observer dependent reality that plagues the Copenhagen Interpretation.
Greenberger–Horne–Zeilinger state the key advance over Everett's Relative State Interpretation[6] is to regard the conjugate state vector of the Dirac formalism
as ontologically real, incorporating a part of the formalism that,
prior to TIQM, had been interpretationally neglected. Having interpreted
the conjugate state vector as an advanced wave, it is shown that the
origins of the Born rule follow naturally from the description of a transaction.
The transactional interpretation is superficially similar to the two-state vector formalism (TSVF) which has its origin in work by Yakir Aharonov, Peter Bergmann and Joel Lebowitz of 1964.[8][9]
However, it has important differences—the TSVF is lacking the
confirmation and therefore cannot provide a physical referent for the
Born Rule (as TI does). Kastner has criticized some other time-symmetric
interpretations, including TSVF, as making ontologically inconsistent
claims.
Kastner has developed a new Relativistic Transactional Interpretation (RTI) also called Possibilist Transactional Interpretation
(PTI) in which space-time itself emerges by a way of transactions. It
has been argued that this relativistic transactional interpretation can
provide the quantum dynamics for the causal sets program.
In his book, The Quantum Handshake, Cramer has added a
hierarchy to the description of pseudo-time to deal with Maudlin's
objection and has pointed out that some of Maudlin's arguments are based
on the inappropriate application of Heisenberg's knowledge
interpretation to the transactional description.
Transactional Interpretation faces criticisms. The following is partial list and some replies:
1. “TI does not generate new predictions / is not testable / has not been tested.”
TI is an exact interpretation of QM and so its predictions must be the same as QM. Like the many-worlds interpretation
(MWI), TI is a "pure" interpretation in that it does not add anything
ad hoc but provides a physical referent for a part of the formalism that
has lacked one (the advanced states implicitly appearing in the Born rule).
Thus the demand often placed on TI for new predictions or testability
is a mistaken one that misconstrues the project of interpretation as one
of theory modification.
2. “It is not made clear where in spacetime a transaction occurs.”
One clear account is given in Cramer (1986), which pictures a
transaction as a four-vector standing wave whose endpoints are the
emission and absorption events.
3. “Maudlin (1996, 2002) has demonstrated that TI is inconsistent.”
Maudlin's probability criticism confused the transactional
interpretation with Heisenberg's knowledge interpretation. However, he
raised a valid point concerning causally connected possible outcomes,
which led Cramer to add hierarchy to the pseudo-time description of
transaction formation. Kastner has extended TI to the relativistic domain, and in light of this
expansion of the interpretation, it can be shown that the Maudlin
Challenge cannot even be mounted, and is therefore nullified; there is
no need for the 'hierarchy' proposal of Cramer.
Maudlin has also claimed that all the dynamics of TI is deterministic
and therefore there can be no 'collapse.' But this appears to disregard
the response of absorbers, which is the whole innovation of the model.
Specifically, the linearity of the Schrödinger evolution is broken by
the response of absorbers; this directly sets up the non-unitary
measurement transition, without any need for ad hoc modifications to the
theory. The non-unitarity is discussed, for example in Chapter 3 of
Kastner's book The Transactional Interpretation of Quantum Mechanics:
The Reality of Possibility (CUP, 2012).
4. "It is not clear how the transactional interpretation handles the quantum mechanics of more than one particle."
This issue is addressed in Cramer's 1986 paper, in which he gives
many examples of the application of TIQM to multi-particle quantum
systems. However, if the question is about the existence of
multi-particle wave functions in normal 3D space, Cramer's 2015 book
goes into some detail in justifying multi-particle wave functions in 3D
space.
A criticism of Cramer's 2015 account of dealing with multi-particle
quantum systems is found in Kastner 2016, "An Overview of the
Transactional Interpretation and its Evolution into the 21st Century,
Philosophy Compass (2016)
.
It observes in particular that the account in Cramer 2015 is
necessarily anti-realist about the multi-particle states: if they are
only part of a 'map,' then they are not real, and in this form TI
becomes an instrumentalist interpretation, contrary to its original
spirit. Thus the so-called "retreat" to Hilbert space (criticized also
below in the lengthy discussion of note)
can instead be seen as a needed expansion of the ontology, rather than a
retreat to anti-realism/instrumentalism about the multi-particle
states. The vague statement (under
that "Offer waves are somewhat ephemeral three-dimensional space
objects" indicates the lack of clear definition of the ontology when one
attempts to keep everything in 3+1 spacetime.
The framework of quantum mechanics requires a careful definition of measurement. The issue of measurement lies at the heart of the problem of the interpretation of quantum mechanics, for which there is currently no consensus. The question of how the operational process measurement affects the ontological state of the observed system is unresolved, and called the measurement problem.
Measurement from a practical point of view
Measurement plays an important role in quantum mechanics, and it is viewed in different ways among various interpretations of quantum mechanics. In spite of considerable philosophical differences, different views of measurement almost universally agree on the practical question of what results from a routine quantum-physics laboratory measurement. To understand this, the Copenhagen interpretation, which has been commonly used, is employed in this article.
Qualitative overview
In classical mechanics, a simple system consisting of only one single particle is fully described by the position and momentum of the particle. As an analogue, in quantum mechanics a system is described by its quantum state,
which contains the probabilities of possible positions and momenta. In
mathematical language, all possible pure states of a system form an
abstract vector space called Hilbert space, which is typically infinite-dimensional. A pure state is represented by a state vector in the Hilbert space.
Once a quantum system has been prepared in laboratory, some
measurable quantity such as position or energy is measured. For
pedagogic reasons, the measurement is usually assumed to be ideally
accurate. The state of a system after measurement is assumed to "collapse" into an eigenstate of the operator
corresponding to the measurement. Repeating the same measurement
without any evolution of the quantum state will lead to the same result.
If the preparation is repeated, subsequent measurements will likely
lead to different results.
The predicted values of the measurement are described by a probability distribution, or an "average" (or "expectation") of the measurement operator based on the quantum state of the prepared system. The probability distribution is either continuous (such as position and momentum) or discrete (such as spin), depending on the quantity being measured.
The measurement process is often considered as random and indeterministic. Nonetheless, there is considerable dispute over this issue. In some interpretations of quantum mechanics, the result merely appears
random and indeterministic, whereas in other interpretations the
indeterminism is core and irreducible. A significant element in this
disagreement is the issue of "collapse of the wave function"
associated with the change in state following measurement. There are
many philosophical issues and stances (and some mathematical variations)
taken—and near universal agreement that we do not yet fully understand
quantum reality. In any case, our descriptions of dynamics involve
probabilities, not certainties.
Quantitative details
The
mathematical relationship between the quantum state and the probability
distribution is, again, widely accepted among physicists, and has been
experimentally confirmed countless times. This section summarizes this
relationship, which is stated in terms of the mathematical formulation of quantum mechanics.
Measurable quantities ("observables") as operators
It is a postulate of quantum mechanics that all measurements have an associated operator (called an observable operator, or just an observable), with the following properties:
Thus, the observable's eigenvectors (called an eigenbasis) form an orthonormalbasis that span the state space in which that observable exists. Any quantum state can be represented as a superposition of the eigenstates of an observable.
Hermitian operators' eigenvalues are real. The possible outcomes of a measurement are precisely the eigenvalues of the given observable.
For each eigenvalue there are one or more corresponding eigenvectors (eigenstates).
A measurement results in the system being in the eigenstate
corresponding to the eigenvalue result of the measurement. If the
eigenvalue determined from the measurement corresponds to more than one
eigenstate ("degeneracy"), instead of being in a definite state, the
system is in a sub-space of the measurement operator corresponding to
all the states having that eigenvalue.
The position operator is given by (in the position basis), or (in the momentum basis).
Operators can be noncommuting.
Two Hermitian operators commute if (and only if) there is at least one
basis of vectors such that each of which is an eigenvector of both
operators (this is sometimes called a simultaneous eigenbasis). Noncommuting observables are said to be incompatible and cannot in general be measured simultaneously. In fact, they are related by an uncertainty principle as discovered by Werner Heisenberg.
Measurement probabilities and wave function collapse
There
are a few possible ways to mathematically describe the measurement
process (both the probability distribution and the collapsed wave
function). The most convenient description depends on the spectrum (i.e., set of eigenvalues) of the observable.
Discrete, nondegenerate spectrum
Let be an observable. By assumption, has discrete eigenstates with corresponding distinct eigenvalues . That is, the states are nondegenerate.
Consider a system prepared in state . Since the eigenstates of the observable form a complete basis called eigenbasis, the state vector can be written in terms of the eigenstates as
,
where are complex numbers in general. The eigenvalues are all possible values of the measurement. The corresponding probabilities are given by
Usually is assumed to be normalized, i.e. . Therefore, the expression above is reduced to
If the result of the measurement is , then the system (after measurement) is in pure state . That is,
so any repeated measurement of will yield the same result .
When there is a discontinuous change in state due to a measurement that involves discrete eigenvalues, that is called wave function collapse.
For some, this is simply a description of a reasonably accurate
discontinuous change in a mathematical representation of physical
reality; for others, depending on philosophical orientation, this is a
fundamentally serious problem with quantum theory; others see this as
statistically-justified approximation resulting from the fact that the
entity performing this measurement has been excluded from the
state-representation. In particular, multiple measurements of certain
physically extended systems demonstrate predicted statistical
correlations which would not be possible under classical assumptions.
Continuous, nondegenerate spectrum
Let be an observable. By assumption, has continuous eigenstate , with corresponding distinct eigenvalue . The eigenvalue forms a continuous spectrum filling the interval (a,b).
Consider a system prepared in state . Since the eigenstates of the observable form a complete basis called eigenbasis, the state vector can be written in terms of the eigenstates as
,
where is a complex-valued function. The eigenvalue that fills up the interval is the possible value of measurement. The corresponding probability is described by a probability function given by
where . Usually is assumed to be normalized, i.e. . Therefore, the expression above is reduced to
If the result of the measurement is , then the system (after measurement) is in pure state . That is,
Alternatively, it is often possible and convenient to analyze a continuous-spectrum measurement by taking it to be the limit of a different measurement with a discrete spectrum. For example, an analysis of scattering involves a continuous spectrum of energies, but by adding a "box" potential (which bounds the volume in which the particle can be found), the spectrum becomes discrete.
By considering larger and larger boxes, this approach need not involve
any approximation, but rather can be regarded as an equally valid
formalism in which this problem can be analyzed.
Degenerate spectra
If there are multiple eigenstates with the same eigenvalue (called degeneracies),
the analysis is a bit less simple to state, but not essentially
different. In the discrete case, for example, instead of finding a
complete eigenbasis, it is a bit more convenient to write the Hilbert
space as a direct sum of multiple eigenspaces. The probability of measuring a particular eigenvalue is the squared component of the state vector in the corresponding eigenspace, and the new state after measurement is the projection of the original state vector into the appropriate eigenspace.
Density matrix formulation
Instead of performing quantum-mechanics computations in terms of wave functions (kets), it is sometimes necessary to describe a quantum-mechanical system in terms of a density matrix.
The analysis in this case is formally slightly different, but the
physical content is the same, and indeed this case can be derived from
the wave function formulation above. The result for the discrete,
degenerate case, for example, is as follows:
Assume the system is prepared in the state described by the density matrix ρ. Then measuring can yield any of the results , with corresponding probabilities given by
where denotes trace. If the result of the measurement is n, then the new density matrix will be
Alternatively, one can say that the measurement process results in the new density matrix
where the difference is that is the density matrix describing the entire ensemble, whereas is the density matrix describing the sub-ensemble whose measurement result was .
Statistics of measurement
As
detailed above, the result of measuring a quantum-mechanical system is
described by a probability distribution. Some properties of this
distribution are as follows:
Suppose we take a measurement corresponding to observable , on a state whose quantum state is .
These are direct consequences of the above formulas for measurement probabilities.
Example
Suppose that we have a particle in a 1-dimensional box, set up initially in the ground state . As can be computed from the time-independent Schrödinger equation, the energy of this state is (where m is the particle's mass and L is the box length), and the spatial wave function is . If the energy is now measured, the result will always certainly be , and this measurement will not affect the wave function.
Next suppose that the particle's position is measured. The position x will be measured with probability density
If the measurement result was x=S, then the wave function after measurement will be the position eigenstate . If the particle's position is immediately measured again, the same position will be obtained.
The new wave function
can, like any wave function, be written as a superposition of
eigenstates of any observable. In particular, using energy eigenstates, , we have
If we now leave this state alone, it will smoothly evolve in time according to the Schrödinger equation. But suppose instead that an energy measurement is immediately taken. Then the possible energy values will be measured with relative probabilities:
and moreover if the measurement result is , then the new state will be the energy eigenstate .
So in this example, due to the process of wave function collapse, a particle initially in the ground state can end up in any energy level, after just two subsequent non-commuting measurements are made.
Wave function collapse
According to the Copenhagen interpretation the process in which a quantum state becomes one of the eigenstates of the operator corresponding to the measured observable is called "collapse", or "wave function collapse". The final eigenstate appears randomly with a probability equal to the square of its overlap with the original state. The process of collapse has been studied in many experiments, most famously in the double-slit experiment. The wave function collapse raises serious questions regarding "the measurement problem", as well as questions of determinism and locality, as demonstrated in the EPR paradox and later in GHZ entanglement.
In the last few decades, major advances have been made toward a
theoretical understanding of the collapse process. This new theoretical
framework, called quantum decoherence, supersedes previous notions of instantaneous collapse and provides an explanation for the absence of quantum coherence
after measurement. Decoherence correctly predicts the form and
probability distribution of the final eigenstates, and explains the
apparent randomness of the choice of final state in terms of einselection.
von Neumann measurement scheme
The von Neumann measurement scheme, the ancestor of quantum decoherence theory, describes measurements by taking into account the measuring apparatus which is also treated as a quantum object.
"Measurement" of the first kind — premeasurement without detection
Let the quantum state be in the superposition , where are eigenstates
of the operator for the so-called "measurement" prior to von Neumann's
second apparatus. In order to make the "measurement", the system
described by needs to interact with the measuring apparatus described by the quantum state . The total wave function before the interaction with the second apparatus is then . During the interaction of object and measuring instrument, the unitary evolution is supposed to realize the following transition from the initial to the final total wave function:
where are orthonormal states of the measuring apparatus. The unitary evolution above is referred to as premeasurement. The relation with wave function collapse is established by calculating the final density operator of the object
from the final total wave function. This density operator is
interpreted by von Neumann as describing an ensemble of objects being
after the measurement with probability in the state
The transition
is often referred to as weak von Neumann projection, the wave function collapse or strong von Neumann projection
being thought to correspond to an additional selection of a subensemble by means of observation.
In case the measured observable has a degenerate spectrum, weak von Neumann projection is generalized to Lüders projection
in which the vectors for fixed n are the degenerate eigenvectors of the measured observable. For an arbitrary state described by a density operator
Lüders projection is given by
Measurement of the second kind — with irreversible detection
In a measurement of the second kind the unitary evolution during the interaction of object and measuring instrument is supposed to be given by
in which the states
of the object are determined by specific properties of the interaction
between object and measuring instrument. They are normalized but not
necessarily mutually orthogonal. The relation with wave function collapse is analogous to that obtained for measurements of the first kind, the final state of the object now being with probability Note that many measurement procedures are measurements of the second kind, some even functioning correctly only as a consequence of being of the second kind.
For instance, a photon counter, detecting a photon by absorbing and
hence annihilating it, thus ideally leaving the electromagnetic field in
the vacuum state rather than in the state corresponding to the number
of detected photons; also the Stern–Gerlach experiment would not function at all if it really were a measurement of the first kind.
Decoherence in quantum measurement
One can also introduce the interaction with the environment , so that, in a measurement of the first kind, after the interaction the total wave function takes a form
which is related to the phenomenon of decoherence.
The above is completely described by the Schrödinger equation and
there are not any interpretational problems with this. Now the
problematic wave function collapse does not need to be understood as a process on the level of the measured system, but can also be understood as a process on the level of the measuring apparatus, or as a process on the level of the environment. Studying these processes provides considerable insight into the measurement problem
by avoiding the arbitrary boundary between the quantum and classical
worlds, though it does not explain the presence of randomness in the
choice of final eigenstate. If the set of states
, , or
represents a set of states that do not overlap in space, the appearance of collapse can be generated by either the Bohm interpretation or the Everett interpretation
which both deny the reality of wave function collapse. Both of these
are stated to predict the same probabilities for collapses to various
states as the conventional interpretation by their supporters. The Bohm
interpretation is held to be correct only by a small minority of
physicists, since there are difficulties with the generalization for use
with relativistic quantum field theory.
However, there is no proof that the Bohm interpretation is inconsistent
with quantum field theory, and work to reconcile the two is ongoing.
The Everett interpretation easily accommodates relativistic quantum field theory.
Quotes
A measurement always causes the system to jump into an eigenstate
of the dynamical variable that is being measured, the eigenvalue this
eigenstate belongs to being equal to the result of the measurement.