In computer algebra
one often wishes to encode mathematical expressions using an expression
tree. But there are often multiple equivalent expression trees. The
question naturally arises of whether there is an algorithm which, given
as input two expressions, decides whether they represent the same
element. Such an algorithm is called a solution to the word problem. For example, imagine that are symbols representing real numbers - then a relevant solution to the word problem would, given the input , produce the output EQUAL, and similarly produce NOT_EQUAL from .
The most direct solution to a word problem takes the form of a normal form theorem and algorithm which maps every element in an equivalence class of expressions to a single encoding known as the normal form - the word problem is then solved by comparing these normal forms via syntactic equality. For example one might decide that is the normal form of , , and ,
and devise a transformation system to rewrite those expressions to that
form, in the process proving that all equivalent expressions will be
rewritten to the same normal form.
But not all solutions to the word problem use a normal form theorem -
there are algebraic properties which indirectly imply the existence of
an algorithm.
While the word problem asks whether two terms containing constants are equal, a proper extension of the word problem known as the unification problem asks whether two terms containing variables have instances that are equal, or in other words whether the equation has any solutions. As a common example, is a word problem in the integer group,
while is a unification problem in the same group; since the former terms happen to be equal in , the latter problem has the substitution as a solution.
History
One of the most deeply studied cases of the word problem is in the theory of semigroups and groups. A timeline of papers relevant to the Novikov-Boone theorem is as follows:
1910: Axel Thue
poses a general problem of term rewriting on tree-like structures. He
states "A solution of this problem in the most general case may perhaps
be connected with unsurmountable difficulties".
1912: Dehn presents Dehn's algorithm, and proves it solves the word problem for the fundamental groups of closed orientable two-dimensional manifolds of genus greater than or equal to 2. Subsequent authors have greatly extended it to a wide range of group-theoretic decision problems.
1914: Axel Thue poses the word problem for finitely presented semigroups.
1930 – 1938: The Church-Turing thesis emerges, defining formal notions of computability and undecidability.
1947: Emil Post and Andrey Markov Jr. independently construct finitely presented semigroups with unsolvable word problem. Post's construction is built on Turing machines while Markov's uses Post's normal systems.
1950: Alan Turing shows the word problem for cancellation semigroups is unsolvable, by furthering Post's construction. The proof is difficult to follow but marks a turning point in the word problem for groups.
1955: Pyotr Novikov gives the first published proof that the word problem for groups is unsolvable, using Turing's cancellation semigroup result. The proof contains a "Principal Lemma" equivalent to Britton's Lemma.
1954 – 1957: William Boone independently shows the word problem for groups is unsolvable, using Post's semigroup construction.
1957 – 1958: John Britton
gives another proof that the word problem for groups is unsolvable,
based on Turing's cancellation semigroups result and some of Britton's
earlier work. An early version of Britton's Lemma appears.
1958 – 1959: Boone publishes a simplified version of his construction.
1961: Graham Higman characterises the subgroups of finitely presented groups with Higman's embedding theorem,
connecting recursion theory with group theory in an unexpected way and
giving a very different proof of the unsolvability of the word problem.
1961 – 1963: Britton presents a greatly simplified version of Boone's 1959 proof that the word problem for groups is unsolvable. It uses a group-theoretic approach, in particular Britton's Lemma. This proof has been used in a graduate course, although more modern and condensed proofs exist.
1977: Gennady Makanin proves that the existential theory of equations over free monoids is solvable.
The word problem for semi-Thue systems
The accessibility problem for string rewriting systems (semi-Thue systems or semigroups) can be stated as follows: Given a semi-Thue system and two words (strings) , can be transformed into by applying rules from ?
Note that the rewriting here is one-way. The word problem is the
accessibility problem for symmetric rewrite relations, i.e. Thue
systems.
The accessibility and word problems are undecidable, i.e. there is no general algorithm for solving this problem.
This even holds if we limit the systems to have finite presentations,
i.e. a finite set of symbols and a finite set of relations on those
symbols. Even the word problem restricted to ground terms is not decidable for certain finitely presented semigroups.
Given a presentation for a group G, the word problem is the algorithmic problem of deciding, given as input two words in S, whether they represent the same element of G. The word problem is one of three algorithmic problems for groups proposed by Max Dehn in 1911. It was shown by Pyotr Novikov in 1955 that there exists a finitely presented group G such that the word problem for G is undecidable.
The word problem in combinatorial calculus and lambda calculus
One of the earliest proofs that a word problem is undecidable was for combinatory logic: when are two strings of combinators equivalent? Because combinators encode all possible Turing machines,
and the equivalence of two Turing machines is undecidable, it follows
that the equivalence of two strings of combinators is undecidable. Alonzo Church observed this in 1936.
Likewise, one has essentially the same problem in (untyped) lambda calculus: given two distinct lambda expressions, there is no algorithm which can discern whether they are equivalent or not; equivalence is undecidable. For several typed variants of the lambda calculus, equivalence is decidable by comparison of normal forms.
The word problem for abstract rewriting systems
The word problem for an abstract rewriting system (ARS) is quite succinct: given objects x and y are they equivalent under ? The word problem for an ARS is undecidable in general. However, there is a computable
solution for the word problem in the specific case where every object
reduces to a unique normal form in a finite number of steps (i.e. the
system is convergent): two objects are equivalent under if and only if they reduce to the same normal form.
The Knuth-Bendix completion algorithm can be used to transform a set of equations into a convergent term rewriting system.
The word problem in universal algebra
In universal algebra one studies algebraic structures consisting of a generating setA, a collection of operations on A
of finite arity, and a finite set of identities that these operations
must satisfy. The word problem for an algebra is then to determine,
given two expressions (words) involving the generators and operations,
whether they represent the same element of the algebra modulo the
identities. The word problems for groups and semigroups can be phrased
as word problems for algebras.
The word problem on free Heyting algebras is difficult.
The only known results are that the free Heyting algebra on one generator is infinite, and that the free complete Heyting algebra on one generator exists (and has one more element than the free Heyting algebra).
The word problem for free lattices
Example computation of x∧z ~ x∧z∧(x∨y)
x∧z∧(x∨y)
≤~
x∧z
by 5.
since
x∧z
≤~
x∧z
by 1.
since
x∧z
=
x∧z
x∧z
≤~
x∧z∧(x∨y)
by 7.
since
x∧z
≤~
x∧z
and
x∧z
≤~
x∨y
by 1.
since
x∧z
=
x∧z
by 6.
since
x∧z
≤~
x
by 5.
since
x
≤~
x
by 1.
since
x
=
x
The word problem on free lattices and more generally free bounded lattices has a decidable solution. Bounded lattices are algebraic structures with the two binary operations ∨ and ∧ and the two constants (nullary operations) 0 and 1. The set of all well-formed expressions that can be formulated using these operations on elements from a given set of generators X will be called W(X). This set of words contains many expressions that turn out to denote equal values in every lattice. For example, if a is some element of X, then a ∨ 1 = 1 and a ∧ 1 = a. The word problem for free bounded lattices is the problem of determining which of these elements of W(X) denote the same element in the free bounded lattice FX, and hence in every bounded lattice.
The word problem may be resolved as follows. A relation ≤~ on W(X) may be defined inductively by setting w ≤~vif and only if one of the following holds:
w = v (this can be restricted to the case where w and v are elements of X),
w = 0,
v = 1,
w = w1 ∨ w2 and both w1 ≤~v and w2 ≤~v hold,
w = w1 ∧ w2 and either w1 ≤~v or w2 ≤~v holds,
v = v1 ∨ v2 and either w ≤~v1 or w ≤~v2 holds,
v = v1 ∧ v2 and both w ≤~v1 and w ≤~v2 hold.
This defines a preorder ≤~ on W(X), so an equivalence relation can be defined by w ~ v when w ≤~v and v ≤~w. One may then show that the partially orderedquotient setW(X)/~ is the free bounded lattice FX. The equivalence classes of W(X)/~ are the sets of all words w and v with w ≤~v and v ≤~w. Two well-formed words v and w in W(X) denote the same value in every bounded lattice if and only if w ≤~v and v ≤~w;
the latter conditions can be effectively decided using the above
inductive definition. The table shows an example computation to show
that the words x∧z and x∧z∧(x∨y)
denote the same value in every bounded lattice. The case of lattices
that are not bounded is treated similarly, omitting rules 2 and 3 in the
above construction of ≤~.
Example: A term rewriting system to decide the word problem in the free group
Bläsius and Bürckert demonstrate the Knuth–Bendix algorithm on an axiom set for groups.
The algorithm yields a confluent and noetherianterm rewrite system that transforms every term into a unique normal form.
The rewrite rules are numbered incontiguous since some rules became
redundant and were deleted during the algorithm run.
The equality of two terms follows from the axioms if and only if both
terms are transformed into literally the same normal form term. For
example, the terms
, and
share the same normal form, viz. ; therefore both terms are equal in every group.
As another example, the term and has the normal form and ,
respectively. Since the normal forms are literally different, the
original terms cannot be equal in every group. In fact, they are usually
different in non-abelian groups.
Group axioms used in Knuth–Bendix completion
A1
A2
A3
Term rewrite system obtained from Knuth–Bendix completion
In computability theory and computational complexity theory, a decision problem is a computational problem that can be posed as a yes–no question of the input values. An example of a decision problem is deciding by means of an algorithm whether a given natural number is prime. Another is the problem "given two numbers x and y, does x evenly divide y?". The answer is either 'yes' or 'no' depending upon the values of x and y. A method for solving a decision problem, given in the form of an algorithm, is called a decision procedure for that problem. A decision procedure for the decision problem "given two numbers x and y, does x evenly divide y?" would give the steps for determining whether x evenly divides y. One such algorithm is long division.
If the remainder is zero the answer is 'yes', otherwise it is 'no'. A
decision problem which can be solved by an algorithm is called decidable.
Decision problems typically appear in mathematical questions of decidability, that is, the question of the existence of an effective method to determine the existence of some object or its membership in a set; some of the most important problems in mathematics are undecidable.
The field of computational complexity categorizes decidable decision problems by how difficult they are to solve. "Difficult", in this sense, is described in terms of the computational resources needed by the most efficient algorithm for a certain problem. The field of recursion theory, meanwhile, categorizes undecidable decision problems by Turing degree, which is a measure of the noncomputability inherent in any solution.
Definition
A decision problem is a yes-or-no question on an infinite set
of inputs. It is traditional to define the decision problem as the set
of possible inputs together with the set of inputs for which the answer
is yes.
These inputs can be natural numbers, but can also be values of some other kind, like binary strings or strings over some other alphabet. The subset of strings for which the problem returns "yes" is a formal language, and often decision problems are defined as formal languages.
Using an encoding such as Gödel numbering,
any string can be encoded as a natural number, via which a decision
problem can be defined as a subset of the natural numbers. Therefore,
the algorithm of a decision problem is to compute the characteristic function of a subset of the natural numbers.
Examples
A
classic example of a decidable decision problem is the set of prime
numbers. It is possible to effectively decide whether a given natural
number is prime by testing every possible nontrivial factor. Although
much more efficient methods of primality testing are known, the existence of any effective method is enough to establish decidability.
A decision problem is decidable or effectively solvable if the set of inputs (or natural numbers) for which the answer is yes is a recursive set. A problem is partially decidable, semidecidable, solvable, or provable if the set of inputs (or natural numbers) for which the answer is yes is a recursively enumerable set. Problems that are not decidable are undecidable. For those it is not possible to create an algorithm, efficient or otherwise, that solves them.
Decision problems are closely related to function problems,
which can have answers that are more complex than a simple 'yes' or
'no'. A corresponding function problem is "given two numbers x and y, what is x divided by y?".
A function problem consists of a partial functionf; the informal "problem" is to compute the values of f on the inputs for which it is defined.
Every function problem can be turned into a decision problem; the
decision problem is just the graph of the associated function. (The
graph of a function f is the set of pairs (x,y) such that f(x) = y.)
If this decision problem were effectively solvable then the function
problem would be as well. This reduction does not respect computational
complexity, however. For example, it is possible for the graph of a
function to be decidable in polynomial time (in which case running time
is computed as a function of the pair (x,y)) when the function is not computable in polynomial time (in which case running time is computed as a function of x alone). The function f(x) = 2x has this property.
Every decision problem can be converted into the function problem of computing the characteristic function
of the set associated to the decision problem. If this function is
computable then the associated decision problem is decidable. However,
this reduction is more liberal than the standard reduction used in
computational complexity (sometimes called polynomial-time many-one
reduction); for example, the complexity of the characteristic functions
of an NP-complete problem and its co-NP-completecomplement
is exactly the same even though the underlying decision problems may
not be considered equivalent in some typical models of computation.
Unlike decision problems, for which there is only one correct answer
for each input, optimization problems are concerned with finding the best answer to a particular input. Optimization problems arise naturally in many applications, such as the traveling salesman problem and many questions in linear programming.
Function and optimization problems are often transformed into
decision problems by considering the question of whether the output is equal to or less than or equal to
a given value. This allows the complexity of the corresponding decision
problem to be studied; and in many cases the original function or
optimization problem can be solved by solving its corresponding decision
problem. For example, in the traveling salesman problem, the
optimization problem is to produce a tour with minimal weight. The
associated decision problem is: for each N, to decide whether the graph has any tour with weight less than N. By repeatedly answering the decision problem, it is possible to find the minimal weight of a tour.
Because the theory of decision problems is very well developed,
research in complexity theory has typically focused on decision
problems. Optimization problems themselves are still of interest in
computability theory, as well as in fields such as operations research.
In theoretical computer science and mathematics, computational complexity theory focuses on classifying computational problems
according to their resource usage, and relating these classes to each
other. A computational problem is a task solved by a computer. A
computation problem is solvable by mechanical application of
mathematical steps, such as an algorithm.
A problem is regarded as inherently difficult if its solution
requires significant resources, whatever the algorithm used. The theory
formalizes this intuition, by introducing mathematical models of computation to study these problems and quantifying their computational complexity,
i.e., the amount of resources needed to solve them, such as time and
storage. Other measures of complexity are also used, such as the amount
of communication (used in communication complexity), the number of gates in a circuit (used in circuit complexity) and the number of processors (used in parallel computing).
One of the roles of computational complexity theory is to determine the
practical limits on what computers can and cannot do. The P versus NP problem, one of the seven Millennium Prize Problems, is part of the field of computational complexity.
Closely related fields in theoretical computer science are analysis of algorithms and computability theory.
A key distinction between analysis of algorithms and computational
complexity theory is that the former is devoted to analyzing the amount
of resources needed by a particular algorithm to solve a problem,
whereas the latter asks a more general question about all possible
algorithms that could be used to solve the same problem. More precisely,
computational complexity theory tries to classify problems that can or
cannot be solved with appropriately restricted resources. In turn,
imposing restrictions on the available resources is what distinguishes
computational complexity from computability theory: the latter theory
asks what kinds of problems can, in principle, be solved
algorithmically.
Computational problems
Problem instances
A computational problem can be viewed as an infinite collection of instances together with a set (possibly empty) of solutions
for every instance. The input string for a computational problem is
referred to as a problem instance, and should not be confused with the
problem itself. In computational complexity theory, a problem refers to
the abstract question to be solved. In contrast, an instance of this
problem is a rather concrete utterance, which can serve as the input for
a decision problem. For example, consider the problem of primality testing.
The instance is a number (e.g., 15) and the solution is "yes" if the
number is prime and "no" otherwise (in this case, 15 is not prime and
the answer is "no"). Stated another way, the instance is a particular input to the problem, and the solution is the output corresponding to the given input.
To further highlight the difference between a problem and an
instance, consider the following instance of the decision version of the
travelling salesman problem:
Is there a route of at most 2000 kilometres passing through all of
Germany's 15 largest cities? The quantitative answer to this particular
problem instance is of little use for solving other instances of the
problem, such as asking for a round trip through all sites in Milan
whose total length is at most 10 km. For this reason, complexity theory
addresses computational problems and not particular problem instances.
Representing problem instances
When considering computational problems, a problem instance is a string over an alphabet. Usually, the alphabet is taken to be the binary alphabet (i.e., the set {0,1}), and thus the strings are bitstrings. As in a real-world computer, mathematical objects other than bitstrings must be suitably encoded. For example, integers can be represented in binary notation, and graphs can be encoded directly via their adjacency matrices, or by encoding their adjacency lists in binary.
Even though some proofs of complexity-theoretic theorems
regularly assume some concrete choice of input encoding, one tries to
keep the discussion abstract enough to be independent of the choice of
encoding. This can be achieved by ensuring that different
representations can be transformed into each other efficiently.
Decision problems as formal languages
Decision problems
are one of the central objects of study in computational complexity
theory. A decision problem is a special type of computational problem
whose answer is either yes or no, or alternately either 1 or 0. A decision problem can be viewed as a formal language,
where the members of the language are instances whose output is yes,
and the non-members are those instances whose output is no. The
objective is to decide, with the aid of an algorithm,
whether a given input string is a member of the formal language under
consideration. If the algorithm deciding this problem returns the answer
yes, the algorithm is said to accept the input string, otherwise it is said to reject the input.
An example of a decision problem is the following. The input is an arbitrary graph. The problem consists in deciding whether the given graph is connected
or not. The formal language associated with this decision problem is
then the set of all connected graphs — to obtain a precise definition of
this language, one has to decide how graphs are encoded as binary
strings.
It is tempting to think that the notion of function problems is
much richer than the notion of decision problems. However, this is not
really the case, since function problems can be recast as decision
problems. For example, the multiplication of two integers can be
expressed as the set of triples (a, b, c) such that the relation a × b = c holds. Deciding whether a given triple is a member of this set corresponds to solving the problem of multiplying two numbers.
Measuring the size of an instance
To
measure the difficulty of solving a computational problem, one may wish
to see how much time the best algorithm requires to solve the problem.
However, the running time may, in general, depend on the instance. In
particular, larger instances will require more time to solve. Thus the
time required to solve a problem (or the space required, or any measure
of complexity) is calculated as a function of the size of the instance.
This is usually taken to be the size of the input in bits. Complexity
theory is interested in how algorithms scale with an increase in the
input size. For instance, in the problem of finding whether a graph is
connected, how much more time does it take to solve a problem for a
graph with 2n vertices compared to the time taken for a graph with n vertices?
If the input size is n, the time taken can be expressed as a function of n. Since the time taken on different inputs of the same size can be different, the worst-case time complexity T(n) is defined to be the maximum time taken over all inputs of size n. If T(n) is a polynomial in n, then the algorithm is said to be a polynomial time algorithm. Cobham's thesis argues that a problem can be solved with a feasible amount of resources if it admits a polynomial-time algorithm.
A Turing machine is a mathematical model of a general computing
machine. It is a theoretical device that manipulates symbols contained
on a strip of tape. Turing machines are not intended as a practical
computing technology, but rather as a general model of a computing
machine—anything from an advanced supercomputer to a mathematician with a
pencil and paper. It is believed that if a problem can be solved by an
algorithm, there exists a Turing machine that solves the problem.
Indeed, this is the statement of the Church–Turing thesis. Furthermore, it is known that everything that can be computed on other models of computation known to us today, such as a RAM machine, Conway's Game of Life, cellular automata, lambda calculus
or any programming language can be computed on a Turing machine. Since
Turing machines are easy to analyze mathematically, and are believed to
be as powerful as any other model of computation, the Turing machine is
the most commonly used model in complexity theory.
A deterministic Turing machine is the most basic Turing machine,
which uses a fixed set of rules to determine its future actions. A
probabilistic Turing machine is a deterministic Turing machine with an
extra supply of random bits. The ability to make probabilistic decisions
often helps algorithms solve problems more efficiently. Algorithms that
use random bits are called randomized algorithms.
A non-deterministic Turing machine is a deterministic Turing machine
with an added feature of non-determinism, which allows a Turing machine
to have multiple possible future actions from a given state. One way to
view non-determinism is that the Turing machine branches into many
possible computational paths at each step, and if it solves the problem
in any of these branches, it is said to have solved the problem.
Clearly, this model is not meant to be a physically realizable model, it
is just a theoretically interesting abstract machine that gives rise to
particularly interesting complexity classes. For examples, see non-deterministic algorithm.
Other machine models
Many machine models different from the standard multi-tape Turing machines have been proposed in the literature, for example random-access machines.
Perhaps surprisingly, each of these models can be converted to another
without providing any extra computational power. The time and memory
consumption of these alternate models may vary. What all these models have in common is that the machines operate deterministically.
However, some computational problems are easier to analyze in
terms of more unusual resources. For example, a non-deterministic Turing
machine is a computational model that is allowed to branch out to check
many different possibilities at once. The non-deterministic Turing
machine has very little to do with how we physically want to compute
algorithms, but its branching exactly captures many of the mathematical
models we want to analyze, so that non-deterministic time is a very important resource in analyzing computational problems.
Complexity measures
For
a precise definition of what it means to solve a problem using a given
amount of time and space, a computational model such as the deterministic Turing machine is used. The time required by a deterministic Turing machine M on input x
is the total number of state transitions, or steps, the machine makes
before it halts and outputs the answer ("yes" or "no"). A Turing machine
M is said to operate within time f(n) if the time required by M on each input of length n is at most f(n). A decision problem A can be solved in time f(n) if there exists a Turing machine operating in time f(n)
that solves the problem. Since complexity theory is interested in
classifying problems based on their difficulty, one defines sets of
problems based on some criteria. For instance, the set of problems
solvable within time f(n) on a deterministic Turing machine is then denoted by DTIME(f(n)).
The complexity of an algorithm is often expressed using big O notation.
Best, worst and average case complexity
The best, worst and average case
complexity refer to three different ways of measuring the time
complexity (or any other complexity measure) of different inputs of the
same size. Since some inputs of size n may be faster to solve than others, we define the following complexities:
Best-case complexity: This is the complexity of solving the problem for the best input of size n.
Average-case complexity: This is the complexity of solving the
problem on an average. This complexity is only defined with respect to a
probability distribution
over the inputs. For instance, if all inputs of the same size are
assumed to be equally likely to appear, the average case complexity can
be defined with respect to the uniform distribution over all inputs of
size n.
Amortized analysis:
Amortized analysis considers both the costly and less costly operations
together over the whole series of operations of the algorithm.
Worst-case complexity: This is the complexity of solving the problem for the worst input of size n.
For example, consider the deterministic sorting algorithm quicksort.
This solves the problem of sorting a list of integers that is given as
the input. The worst-case is when the pivot is always the largest or
smallest value in the list (so the list is never divided). In this case
the algorithm takes time O(n2). If we assume that all possible permutations of the input list are equally likely, the average time taken for sorting is O(n log n). The best case occurs when each pivoting divides the list in half, also needing O(n log n) time.
Upper and lower bounds on the complexity of problems
To
classify the computation time (or similar resources, such as space
consumption), it is helpful to demonstrate upper and lower bounds on the
maximum amount of time required by the most efficient algorithm to
solve a given problem. The complexity of an algorithm is usually taken
to be its worst-case complexity unless specified otherwise. Analyzing a
particular algorithm falls under the field of analysis of algorithms. To show an upper bound T(n) on the time complexity of a problem, one needs to show only that there is a particular algorithm with running time at most T(n).
However, proving lower bounds is much more difficult, since lower
bounds make a statement about all possible algorithms that solve a given
problem. The phrase "all possible algorithms" includes not just the
algorithms known today, but any algorithm that might be discovered in
the future. To show a lower bound of T(n) for a problem requires showing that no algorithm can have time complexity lower than T(n).
Upper and lower bounds are usually stated using the big O notation,
which hides constant factors and smaller terms. This makes the bounds
independent of the specific details of the computational model used. For
instance, if T(n) = 7n2 + 15n + 40, in big O notation one would write T(n) = O(n2).
The model of computation: The most common model of computation is
the deterministic Turing machine, but many complexity classes are based
on non-deterministic Turing machines, Boolean circuits, quantum Turing machines, monotone circuits, etc.
The resource (or resources) that is being bounded and the bound:
These two properties are usually stated together, such as "polynomial
time", "logarithmic space", "constant depth", etc.
Some complexity classes have complicated definitions that do not fit
into this framework. Thus, a typical complexity class has a definition
like the following:
The set of decision problems solvable by a deterministic Turing machine within time f(n). (This complexity class is known as DTIME(f(n)).)
But bounding the computation time above by some concrete function f(n) often yields complexity classes that depend on the chosen machine model. For instance, the language {xx | x is any binary string} can be solved in linear time
on a multi-tape Turing machine, but necessarily requires quadratic time
in the model of single-tape Turing machines. If we allow polynomial
variations in running time, Cobham-Edmonds thesis states that "the time complexities in any two reasonable and general models of computation are polynomially related" (Goldreich 2008, Chapter 1.2). This forms the basis for the complexity class P,
which is the set of decision problems solvable by a deterministic
Turing machine within polynomial time. The corresponding set of function
problems is FP.
Important complexity classes
Many important complexity classes can be defined by bounding the time
or space used by the algorithm. Some important complexity classes of
decision problems defined in this manner are the following:
The logarithmic-space classes (necessarily) do not take into account the space needed to represent the problem.
It turns out that PSPACE = NPSPACE and EXPSPACE = NEXPSPACE by Savitch's theorem.
Other important complexity classes include BPP, ZPP and RP, which are defined using probabilistic Turing machines; AC and NC, which are defined using Boolean circuits; and BQP and QMA, which are defined using quantum Turing machines. #P is an important complexity class of counting problems (not decision problems). Classes like IP and AM are defined using Interactive proof systems. ALL is the class of all decision problems.
For the complexity classes defined in this way, it is desirable to
prove that relaxing the requirements on (say) computation time indeed
defines a bigger set of problems. In particular, although DTIME(n) is contained in DTIME(n2),
it would be interesting to know if the inclusion is strict. For time
and space requirements, the answer to such questions is given by the
time and space hierarchy theorems respectively. They are called
hierarchy theorems because they induce a proper hierarchy on the classes
defined by constraining the respective resources. Thus there are pairs
of complexity classes such that one is properly included in the other.
Having deduced such proper set inclusions, we can proceed to make
quantitative statements about how much more additional time or space is
needed in order to increase the number of problems that can be solved.
The time and space hierarchy theorems form the basis for most
separation results of complexity classes. For instance, the time
hierarchy theorem tells us that P is strictly contained in EXPTIME, and
the space hierarchy theorem tells us that L is strictly contained in
PSPACE.
Many complexity classes are defined using the concept of a reduction.
A reduction is a transformation of one problem into another problem. It
captures the informal notion of a problem being at most as difficult as
another problem. For instance, if a problem X can be solved using an algorithm for Y, X is no more difficult than Y, and we say that Xreduces to Y.
There are many different types of reductions, based on the method of
reduction, such as Cook reductions, Karp reductions and Levin
reductions, and the bound on the complexity of reductions, such as polynomial-time reductions or log-space reductions.
The most commonly used reduction is a polynomial-time reduction.
This means that the reduction process takes polynomial time. For
example, the problem of squaring an integer can be reduced to the
problem of multiplying two integers. This means an algorithm for
multiplying two integers can be used to square an integer. Indeed, this
can be done by giving the same input to both inputs of the
multiplication algorithm. Thus we see that squaring is not more
difficult than multiplication, since squaring can be reduced to
multiplication.
This motivates the concept of a problem being hard for a complexity class. A problem X is hard for a class of problems C if every problem in C can be reduced to X. Thus no problem in C is harder than X, since an algorithm for X allows us to solve any problem in C.
The notion of hard problems depends on the type of reduction being
used. For complexity classes larger than P, polynomial-time reductions
are commonly used. In particular, the set of problems that are hard for
NP is the set of NP-hard problems.
If a problem X is in C and hard for C, then X is said to be complete for C. This means that X is the hardest problem in C. (Since many problems could be equally hard, one might say that X is one of the hardest problems in C.) Thus the class of NP-complete
problems contains the most difficult problems in NP, in the sense that
they are the ones most likely not to be in P. Because the problem P = NP
is not solved, being able to reduce a known NP-complete problem, Π2, to another problem, Π1, would indicate that there is no known polynomial-time solution for Π1. This is because a polynomial-time solution to Π1 would yield a polynomial-time solution to Π2. Similarly, because all NP problems can be reduced to the set, finding an NP-complete problem that can be solved in polynomial time would mean that P = NP.
The complexity class P is often seen as a mathematical abstraction
modeling those computational tasks that admit an efficient algorithm.
This hypothesis is called the Cobham–Edmonds thesis. The complexity class NP,
on the other hand, contains many problems that people would like to
solve efficiently, but for which no efficient algorithm is known, such
as the Boolean satisfiability problem, the Hamiltonian path problem and the vertex cover problem.
Since deterministic Turing machines are special non-deterministic
Turing machines, it is easily observed that each problem in P is also
member of the class NP.
The question of whether P equals NP is one of the most important
open questions in theoretical computer science because of the wide
implications of a solution. If the answer is yes, many important problems can be shown to have more efficient solutions. These include various types of integer programming problems in operations research, many problems in logistics, protein structure prediction in biology, and the ability to find formal proofs of pure mathematics theorems. The P versus NP problem is one of the Millennium Prize Problems proposed by the Clay Mathematics Institute. There is a US$1,000,000 prize for resolving the problem.
Problems in NP not known to be in P or NP-complete
It was shown by Ladner that if P ≠ NP then there exist problems in NP that are neither in P nor NP-complete. Such problems are called NP-intermediate problems. The graph isomorphism problem, the discrete logarithm problem and the integer factorization problem are examples of problems believed to be NP-intermediate. They are some of the very few NP problems not known to be in P or to be NP-complete.
The graph isomorphism problem is the computational problem of determining whether two finite graphs are isomorphic. An important unsolved problem in complexity theory is whether the graph isomorphism problem is in P, NP-complete, or NP-intermediate. The answer is not known, but it is believed that the problem is at least not NP-complete. If graph isomorphism is NP-complete, the polynomial time hierarchy collapses to its second level.
Since it is widely believed that the polynomial hierarchy does not
collapse to any finite level, it is believed that graph isomorphism is
not NP-complete. The best algorithm for this problem, due to László Babai and Eugene Luks has run time for graphs with n vertices, although some recent work by Babai offers some potentially new perspectives on this.
The integer factorization problem is the computational problem of determining the prime factorization
of a given integer. Phrased as a decision problem, it is the problem of
deciding whether the input has a prime factor less than k. No
efficient integer factorization algorithm is known, and this fact forms
the basis of several modern cryptographic systems, such as the RSA algorithm. The integer factorization problem is in NP and in co-NP (and even in UP and co-UP). If the problem is NP-complete, the polynomial time hierarchy will collapse to its first level (i.e., NP will equal co-NP). The best known algorithm for integer factorization is the general number field sieve, which takes time to factor an odd integer n. However, the best known quantum algorithm for this problem, Shor's algorithm,
does run in polynomial time. Unfortunately, this fact doesn't say much
about where the problem lies with respect to non-quantum complexity
classes.
Separations between other complexity classes
Many known complexity classes are suspected to be unequal, but this has not been proved. For instance P ⊆ NP ⊆ PP ⊆ PSPACE, but it is possible that P = PSPACE. If P is not equal to NP, then P is not equal to PSPACE either. Since there are many known complexity classes between P and PSPACE, such as RP, BPP, PP, BQP, MA, PH,
etc., it is possible that all these complexity classes collapse to one
class. Proving that any of these classes are unequal would be a major
breakthrough in complexity theory.
Along the same lines, co-NP is the class containing the complement problems (i.e. problems with the yes/no answers reversed) of NP problems. It is believed that NP is not equal to co-NP; however, it has not yet been proven. It is clear that if these two complexity classes are not equal then P is not equal to NP, since P=co-P. Thus if P=NP we would have co-P=co-NP whence NP=P=co-P=co-NP.
Similarly, it is not known if L (the set of all problems that can be solved in logarithmic space) is strictly contained in P or equal to P. Again, there are many complexity classes between the two, such as NL and NC, and it is not known if they are distinct or equal classes.
It is suspected that P and BPP are equal. However, it is currently open if BPP = NEXP.
A problem that can be solved in theory (e.g. given large but finite resources, especially time), but for which in practice any solution takes too many resources to be useful, is known as an intractable problem. Conversely, a problem that can be solved in practice is called a tractable problem, literally "a problem that can be handled". The term infeasible (literally "cannot be done") is sometimes used interchangeably with intractable, though this risks confusion with a feasible solution in mathematical optimization.
Tractable problems are frequently identified with problems that have polynomial-time solutions (P, PTIME); this is known as the Cobham–Edmonds thesis. Problems that are known to be intractable in this sense include those that are EXPTIME-hard. If NP is not the same as P, then NP-hard problems are also intractable in this sense.
However, this identification is inexact: a polynomial-time
solution with large degree or large leading coefficient grows quickly,
and may be impractical for practical size problems; conversely, an
exponential-time solution that grows slowly may be practical on
realistic input, or a solution that takes a long time in the worst case
may take a short time in most cases or the average case, and thus still
be practical. Saying that a problem is not in P does not imply that all
large cases of the problem are hard or even that most of them are. For
example, the decision problem in Presburger arithmetic
has been shown not to be in P, yet algorithms have been written that
solve the problem in reasonable times in most cases. Similarly,
algorithms can solve the NP-complete knapsack problem over a wide range of sizes in less than quadratic time and SAT solvers routinely handle large instances of the NP-complete Boolean satisfiability problem.
To see why exponential-time algorithms are generally unusable in practice, consider a program that makes 2n operations before halting. For small n, say 100, and assuming for the sake of example that the computer does 1012 operations each second, the program would run for about 4 × 1010 years, which is the same order of magnitude as the age of the universe.
Even with a much faster computer, the program would only be useful for
very small instances and in that sense the intractability of a problem
is somewhat independent of technological progress. However, an
exponential-time algorithm that takes 1.0001n operations is practical until n gets relatively large.
Similarly, a polynomial time algorithm is not always practical. If its running time is, say, n15, it is unreasonable to consider it efficient and it is still useless except on small instances. Indeed, in practice even n3 or n2 algorithms are often impractical on realistic sizes of problems.
Continuous complexity theory
Continuous
complexity theory can refer to complexity theory of problems that
involve continuous functions that are approximated by discretizations,
as studied in numerical analysis. One approach to complexity theory of numerical analysis is information based complexity.
Continuous complexity theory can also refer to complexity theory of the use of analog computation, which uses continuous dynamical systems and differential equations. Control theory
can be considered a form of computation and differential equations are
used in the modelling of continuous-time and hybrid
discrete-continuous-time systems.
History
An early example of algorithm complexity analysis is the running time analysis of the Euclidean algorithm done by Gabriel Lamé in 1844.
Before the actual research explicitly devoted to the complexity
of algorithmic problems started off, numerous foundations were laid out
by various researchers. Most influential among these was the definition
of Turing machines by Alan Turing in 1936, which turned out to be a very robust and flexible simplification of a computer.
The beginning of systematic studies in computational complexity
is attributed to the seminal 1965 paper "On the Computational Complexity
of Algorithms" by Juris Hartmanis and Richard E. Stearns, which laid out the definitions of time complexity and space complexity, and proved the hierarchy theorems. In addition, in 1965 Edmonds suggested to consider a "good" algorithm to be one with running time bounded by a polynomial of the input size.
Earlier papers studying problems solvable by Turing machines with specific bounded resources include John Myhill's definition of linear bounded automata (Myhill 1960), Raymond Smullyan's study of rudimentary sets (1961), as well as Hisao Yamada's paper on real-time computations (1962). Somewhat earlier, Boris Trakhtenbrot (1956), a pioneer in the field from the USSR, studied another specific complexity measure. As he remembers:
However, [my] initial interest [in
automata theory] was increasingly set aside in favor of computational
complexity, an exciting fusion of combinatorial methods, inherited from switching theory,
with the conceptual arsenal of the theory of algorithms. These ideas
had occurred to me earlier in 1955 when I coined the term "signalizing
function", which is nowadays commonly known as "complexity measure".
In 1967, Manuel Blum formulated a set of axioms (now known as Blum axioms)
specifying desirable properties of complexity measures on the set of
computable functions and proved an important result, the so-called speed-up theorem. The field began to flourish in 1971 when Stephen Cook and Leonid Levinproved the existence of practically relevant problems that are NP-complete. In 1972, Richard Karp
took this idea a leap forward with his landmark paper, "Reducibility
Among Combinatorial Problems", in which he showed that 21 diverse combinatorial and graph theoretical problems, each infamous for its computational intractability, are NP-complete.