Search This Blog

Saturday, January 25, 2020

Abstract algebra

From Wikipedia, the free encyclopedia
https://en.wikipedia.org/wiki/Abstract_algebra
 
Picture of a Rubik's Cube
The permutations of Rubik's Cube form a group, a fundamental concept within abstract algebra.
 
In algebra, which is a broad division of mathematics, abstract algebra (occasionally called modern algebra) is the study of algebraic structures. Algebraic structures include groups, rings, fields, modules, vector spaces, lattices, and algebras. The term abstract algebra was coined in the early 20th century to distinguish this area of study from the other parts of algebra.

Algebraic structures, with their associated homomorphisms, form mathematical categories. Category theory is a formalism that allows a unified way for expressing properties and constructions that are similar for various structures.

Universal algebra is a related subject that studies types of algebraic structures as single objects. For example, the structure of groups is a single object in universal algebra, which is called variety of groups.

History

As in other parts of mathematics, concrete problems and examples have played important roles in the development of abstract algebra. Through the end of the nineteenth century, many – perhaps most – of these problems were in some way related to the theory of algebraic equations. Major themes include:
  • Solving of systems of linear equations, which led to linear algebra
  • Attempts to find formulas for solutions of general polynomial equations of higher degree that resulted in discovery of groups as abstract manifestations of symmetry
  • Arithmetical investigations of quadratic and higher degree forms and diophantine equations, that directly produced the notions of a ring and ideal.
Numerous textbooks in abstract algebra start with axiomatic definitions of various algebraic structures and then proceed to establish their properties. This creates a false impression that in algebra axioms had come first and then served as a motivation and as a basis of further study. The true order of historical development was almost exactly the opposite. For example, the hypercomplex numbers of the nineteenth century had kinematic and physical motivations but challenged comprehension. Most theories that are now recognized as parts of algebra started as collections of disparate facts from various branches of mathematics, acquired a common theme that served as a core around which various results were grouped, and finally became unified on a basis of a common set of concepts. An archetypical example of this progressive synthesis can be seen in the history of group theory.

Early group theory

There were several threads in the early development of group theory, in modern language loosely corresponding to number theory, theory of equations, and geometry.

Leonhard Euler considered algebraic operations on numbers modulo an integer—modular arithmetic—in his generalization of Fermat's little theorem. These investigations were taken much further by Carl Friedrich Gauss, who considered the structure of multiplicative groups of residues mod n and established many properties of cyclic and more general abelian groups that arise in this way. In his investigations of composition of binary quadratic forms, Gauss explicitly stated the associative law for the composition of forms, but like Euler before him, he seems to have been more interested in concrete results than in general theory. In 1870, Leopold Kronecker gave a definition of an abelian group in the context of ideal class groups of a number field, generalizing Gauss's work; but it appears he did not tie his definition with previous work on groups, particularly permutation groups. In 1882, considering the same question, Heinrich M. Weber realized the connection and gave a similar definition that involved the cancellation property but omitted the existence of the inverse element, which was sufficient in his context (finite groups). 

Permutations were studied by Joseph-Louis Lagrange in his 1770 paper Réflexions sur la résolution algébrique des équations (Thoughts on the algebraic solution of equations) devoted to solutions of algebraic equations, in which he introduced Lagrange resolvents. Lagrange's goal was to understand why equations of third and fourth degree admit formulas for solutions, and he identified as key objects permutations of the roots. An important novel step taken by Lagrange in this paper was the abstract view of the roots, i.e. as symbols and not as numbers. However, he did not consider composition of permutations. Serendipitously, the first edition of Edward Waring's Meditationes Algebraicae (Meditations on Algebra) appeared in the same year, with an expanded version published in 1782. Waring proved the fundamental theorem of symmetric polynomials, and specially considered the relation between the roots of a quartic equation and its resolvent cubic. Mémoire sur la résolution des équations (Memoire on the Solving of Equations) of Alexandre Vandermonde (1771) developed the theory of symmetric functions from a slightly different angle, but like Lagrange, with the goal of understanding solvability of algebraic equations.
Kronecker claimed in 1888 that the study of modern algebra began with this first paper of Vandermonde. Cauchy states quite clearly that Vandermonde had priority over Lagrange for this remarkable idea, which eventually led to the study of group theory.
Paolo Ruffini was the first person to develop the theory of permutation groups, and like his predecessors, also in the context of solving algebraic equations. His goal was to establish the impossibility of an algebraic solution to a general algebraic equation of degree greater than four. En route to this goal he introduced the notion of the order of an element of a group, conjugacy, the cycle decomposition of elements of permutation groups and the notions of primitive and imprimitive and proved some important theorems relating these concepts, such as
if G is a subgroup of S5 whose order is divisible by 5 then G contains an element of order 5.
However, he got by without formalizing the concept of a group, or even of a permutation group. The next step was taken by Évariste Galois in 1832, although his work remained unpublished until 1846, when he considered for the first time what is now called the closure property of a group of permutations, which he expressed as
if in such a group one has the substitutions S and T then one has the substitution ST.
The theory of permutation groups received further far-reaching development in the hands of Augustin Cauchy and Camille Jordan, both through introduction of new concepts and, primarily, a great wealth of results about special classes of permutation groups and even some general theorems. Among other things, Jordan defined a notion of isomorphism, still in the context of permutation groups and, incidentally, it was he who put the term group in wide use.

The abstract notion of a group appeared for the first time in Arthur Cayley's papers in 1854. Cayley realized that a group need not be a permutation group (or even finite), and may instead consist of matrices, whose algebraic properties, such as multiplication and inverses, he systematically investigated in succeeding years. Much later Cayley would revisit the question whether abstract groups were more general than permutation groups, and establish that, in fact, any group is isomorphic to a group of permutations.

Modern algebra

The end of the 19th and the beginning of the 20th century saw a tremendous shift in the methodology of mathematics. Abstract algebra emerged around the start of the 20th century, under the name modern algebra. Its study was part of the drive for more intellectual rigor in mathematics. Initially, the assumptions in classical algebra, on which the whole of mathematics (and major parts of the natural sciences) depend, took the form of axiomatic systems. No longer satisfied with establishing properties of concrete objects, mathematicians started to turn their attention to general theory. Formal definitions of certain algebraic structures began to emerge in the 19th century. For example, results about various groups of permutations came to be seen as instances of general theorems that concern a general notion of an abstract group. Questions of structure and classification of various mathematical objects came to forefront.

These processes were occurring throughout all of mathematics, but became especially pronounced in algebra. Formal definition through primitive operations and axioms were proposed for many basic algebraic structures, such as groups, rings, and fields. Hence such things as group theory and ring theory took their places in pure mathematics. The algebraic investigations of general fields by Ernst Steinitz and of commutative and then general rings by David Hilbert, Emil Artin and Emmy Noether, building up on the work of Ernst Kummer, Leopold Kronecker and Richard Dedekind, who had considered ideals in commutative rings, and of Georg Frobenius and Issai Schur, concerning representation theory of groups, came to define abstract algebra. These developments of the last quarter of the 19th century and the first quarter of 20th century were systematically exposed in Bartel van der Waerden's Moderne algebra, the two-volume monograph published in 1930–1931 that forever changed for the mathematical world the meaning of the word algebra from the theory of equations to the theory of algebraic structures

Basic concepts

By abstracting away various amounts of detail, mathematicians have defined various algebraic structures that are used in many areas of mathematics. For instance, almost all systems studied are sets, to which the theorems of set theory apply. Those sets that have a certain binary operation defined on them form magmas, to which the concepts concerning magmas, as well those concerning sets, apply. We can add additional constraints on the algebraic structure, such as associativity (to form semigroups); identity, and inverses (to form groups); and other more complex structures. With additional structure, more theorems could be proved, but the generality is reduced. The "hierarchy" of algebraic objects (in terms of generality) creates a hierarchy of the corresponding theories: for instance, the theorems of group theory may be used when studying rings (algebraic objects that have two binary operations with certain axioms) since a ring is a group over one of its operations. In general there is a balance between the amount of generality and the richness of the theory: more general structures have usually fewer nontrivial theorems and fewer applications. 

Examples of algebraic structures with a single binary operation are: 
Examples involving several operations include:

Applications

Because of its generality, abstract algebra is used in many fields of mathematics and science. For instance, algebraic topology uses algebraic objects to study topologies. The Poincaré conjecture, proved in 2003, asserts that the fundamental group of a manifold, which encodes information about connectedness, can be used to determine whether a manifold is a sphere or not. Algebraic number theory studies various number rings that generalize the set of integers. Using tools of algebraic number theory, Andrew Wiles proved Fermat's Last Theorem

In physics, groups are used to represent symmetry operations, and the usage of group theory could simplify differential equations. In gauge theory, the requirement of local symmetry can be used to deduce the equations describing a system. The groups that describe those symmetries are Lie groups, and the study of Lie groups and Lie algebras reveals much about the physical system; for instance, the number of force carriers in a theory is equal to the dimension of the Lie algebra, and these bosons interact with the force they mediate if the Lie algebra is nonabelian.

Logical consequence

From Wikipedia, the free encyclopedia
https://en.wikipedia.org/wiki/Logical_consequence

Logical consequence (also entailment) is a fundamental concept in logic, which describes the relationship between statements that hold true when one statement logically follows from one or more statements. A valid logical argument is one in which the conclusion is entailed by the premises, because the conclusion is the consequence of the premises. The philosophical analysis of logical consequence involves the questions: In what sense does a conclusion follow from its premises? and What does it mean for a conclusion to be a consequence of premises? All of philosophical logic is meant to provide accounts of the nature of logical consequence and the nature of logical truth.

Logical consequence is necessary and formal, by way of examples that explain with formal proof and models of interpretation. A sentence is said to be a logical consequence of a set of sentences, for a given language, if and only if, using only logic (i.e. without regard to any personal interpretations of the sentences) the sentence must be true if every sentence in the set is true.

Logicians make precise accounts of logical consequence regarding a given language , either by constructing a deductive system for or by formal intended semantics for language . The Polish logician Alfred Tarski identified three features of an adequate characterization of entailment: (1) The logical consequence relation relies on the logical form of the sentences, (2) The relation is a priori, i.e. it can be determined with or without regard to empirical evidence (sense experience), and (3) The logical consequence relation has a modal component.

Formal accounts

The most widely prevailing view on how to best account for logical consequence is to appeal to formality. This is to say that whether statements follow from one another logically depends on the structure or logical form of the statements without regard to the contents of that form. 

Syntactic accounts of logical consequence rely on schemes using inference rules. For instance, we can express the logical form of a valid argument as:
All X are Y
All Y are Z
Therefore, all X are Z.
This argument is formally valid, because every instance of arguments constructed using this scheme is valid. 

This is in contrast to an argument like "Fred is Mike's brother's son. Therefore Fred is Mike's nephew." Since this argument depends on the meanings of the words "brother", "son", and "nephew", the statement "Fred is Mike's nephew" is a so-called material consequence of "Fred is Mike's brother's son," not a formal consequence. A formal consequence must be true in all cases, however this is an incomplete definition of formal consequence, since even the argument "P is Q's brother's son, therefore P is Q's nephew" is valid in all cases, but is not a formal argument.

A priori property of logical consequence

If you know that follows logically from , then no information about the possible interpretations of or will affect that knowledge. Our knowledge that is a logical consequence of cannot be influenced by empirical knowledge. Deductively valid arguments can be known to be so without recourse to experience, so they must be knowable a priori. However, formality alone does not guarantee that logical consequence is not influenced by empirical knowledge. So the a priori property of logical consequence is considered to be independent of formality.

Proofs and models

The two prevailing techniques for providing accounts of logical consequence involve expressing the concept in terms of proofs and via models. The study of the syntactic consequence (of a logic) is called (its) proof theory whereas the study of (its) semantic consequence is called (its) model theory.

Syntactic consequence

A formula is a syntactic consequence within some formal system of a set of formulas if there is a formal proof in of from the set .
Syntactic consequence does not depend on any interpretation of the formal system.

Semantic consequence

A formula is a semantic consequence within some formal system of a set of statements
if and only if there is no model in which all members of are true and is false. Or, in other words, the set of the interpretations that make all members of true is a subset of the set of the interpretations that make true. 

Modal accounts

Modal accounts of logical consequence are variations on the following basic idea:
is true if and only if it is necessary that if all of the elements of are true, then is true.
Alternatively (and, most would say, equivalently):
is true if and only if it is impossible for all of the elements of to be true and false.
Such accounts are called "modal" because they appeal to the modal notions of logical necessity and logical possibility. 'It is necessary that' is often expressed as a universal quantifier over possible worlds, so that the accounts above translate as:
is true if and only if there is no possible world at which all of the elements of are true and is false (untrue).
Consider the modal account in terms of the argument given as an example above:
All frogs are green.
Kermit is a frog.
Therefore, Kermit is green.
The conclusion is a logical consequence of the premises because we can't imagine a possible world where (a) all frogs are green; (b) Kermit is a frog; and (c) Kermit is not green. 

Modal-formal accounts

Modal-formal accounts of logical consequence combine the modal and formal accounts above, yielding variations on the following basic idea:
if and only if it is impossible for an argument with the same logical form as / to have true premises and a false conclusion.
 

Warrant-based accounts

The accounts considered above are all "truth-preservational," in that they all assume that the characteristic feature of a good inference is that it never allows one to move from true premises to an untrue conclusion. As an alternative, some have proposed "warrant-preservational" accounts, according to which the characteristic feature of a good inference is that it never allows one to move from justifiably assertible premises to a conclusion that is not justifiably assertible. This is (roughly) the account favored by intuitionists such as Michael Dummett

Non-monotonic logical consequence

The accounts discussed above all yield monotonic consequence relations, i.e. ones such that if is a consequence of , then is a consequence of any superset of . It is also possible to specify non-monotonic consequence relations to capture the idea that, e.g., 'Tweety can fly' is a logical consequence of
{Birds can typically fly, Tweety is a bird}
but not of
{Birds can typically fly, Tweety is a bird, Tweety is a penguin}.

Operator (computer programming)

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Operator_(computer_programmin...