Search This Blog

Tuesday, December 7, 2021

Cognitive development

From Wikipedia, the free encyclopedia
https://en.wikipedia.org/wiki/Cognitive_development

Cognitive development is a field of study in neuroscience and psychology focusing on a child's development in terms of information processing, conceptual resources, perceptual skill, language learning, and other aspects of the developed adult brain and cognitive psychology. Qualitative differences between how a child processes their waking experience and how an adult processes their waking experience are acknowledged (Such as object permanence, the understanding of logical relations, and cause-effect reasoning in school-age children). Cognitive development is defined as the emergence of the ability to consciously cognize, understand, and articulate their understanding in adult terms. Cognitive development is how a person perceives, thinks, and gains understanding of their world through the relations of genetic and learning factors. There are four stages to cognitive information development. They are, reasoning, intelligence, language, and memory. These stages start when the baby is about 18 months old, they play with toys, listen to their parents speak, they watch tv, anything that catches their attention helps build their cognitive development.

Jean Piaget was a major force establishing this field, forming his "theory of cognitive development". Piaget proposed four stages of cognitive development: the sensorimotor, preoperational, concrete operational, and formal operational period. Many of Piaget's theoretical claims have since fallen out of favor. His description of the most prominent changes in cognition with age, is generally still accepted today (e.g., how early perception moves from being dependent on concrete, external actions. Later, abstract understanding of observable aspects of reality can be captured; leading to the discovery of underlying abstract rules and principles, usually starting in adolescence)

In recent years, however, alternative models have been advanced, including information-processing theory, neo-Piagetian theories of cognitive development, which aim to integrate Piaget's ideas with more recent models and concepts in developmental and cognitive science, theoretical cognitive neuroscience, and social-constructivist approaches. Another such model of cognitive development is Bronfenbrenner's Ecological Systems Theory. A major controversy in cognitive development has been "nature versus nurture", i.e, the question if cognitive development is mainly determined by an individual's innate qualities ("nature"), or by their personal experiences ("nurture"). However, it is now recognized by most experts that this is a false dichotomy: there is overwhelming evidence from biological and behavioral sciences that from the earliest points in development, gene activity interacts with events and experiences in the environment.

Historical origins: The history and theory of cognitive development

Jean Piaget is inexorably linked to cognitive development as he was the first to systematically study developmental processes. Despite being the first to develop a systemic study of cognitive development, Piaget was not the first to theorize about cognitive development. A few that are worth mentioning are included in the following chart. It is intended to be a more inclusive list of researchers who have studied the processes of acquiring more complex ways of cognitive development across the lifespan:

Theorist DOB/death Contribution to cognitive development
Jean-Jacques Rousseau 1712–1778 Wrote Emile, or On Education in 1762. He discusses childhood development as happening in three stages. First stage, up to age 12, the child is guided by their emotions and impulses. The second stage, ages 12–16, the child's reason starts to develop. In the third and final stage, age 16 and up, the child develops into an adult.
James Sully 1842–1923 Wrote several books on childhood development, including Studies of Childhood  and Children's Way. He used a detailed observational study method with the children. Contemporary research in child development actually repeats observations, and observational methods, summarized by Sully in Studies of Childhood, such as the mirror technique.
Sigmund Freud 1856–1939 Developed theory of psychosexual development that indicates children must pass through several stages as they develop their cognitive skills.
Lev Vygotsky 1896–1934 Area of specialty was developmental psychology. The main contribution is the somewhat controversial "zone of proximal development" (ZPD) which states that play should be children's main activity as this is their main source of development in terms of emotional, volitional, and cognitive development. ZPD is the link between children's learning and cognitive development.
Maria Montessori 1870–1952 She began her career working with mentally disabled children in 1897, then conducted observation and experimental research in elementary schools. Wrote The Discovery of the Child. Discussed the Four Planes of Development: birth–6, 6–12, 12–18, and 18–24. The Montessori Method now has three developmentally-meaningful age groups: 2–2.5, 2.5–6, and 6–12. She was working on human behavior in older children but only published lecture notes on the subject.
Arnold Gesell 1880–1961 Creator of the Developmental Maturational Theory of development. Gesell says that development occurs due to biological hereditary features such as genetics and children will reach developmental milestones when they are ready to do so in a predictable sequence. Because of his theory of development, he devised a developmental scale that is used today called the Gesell Developmental Schedule (GDS) that provides parents, teachers, doctors, and other pertinent people with an overview of where an infant or child falls on the developmental spectrum.
Jean Piaget 1896–1980 Piaget was the first psychologist and philosopher to brand this type of study as "cognitive development". Other researchers, in multiple disciplines, had studied development in children before, but Piaget is often credited as being the first one to make a systematic study of cognitive development and gave it its name. His main contribution is the stage theory of child cognitive development. He also published his observational studies of cognition in children, and created a series of simple tests to reveal different cognitive abilities in children.
Erik Erikson 1902 –1994 Erik Erikson, a neo-Freudian, focused on how children develop personality and identity. Although a contemporary of Freud, there is a larger focus on social experiences that occur across the lifespan, as opposed to childhood exclusively, that contribute to how personality and identity emerge. His framework uses eight systematic stages that all children must pass through.
Urie Bronfenbrenner 1917–2005 Bronfenbrenner devised the Ecological Systems Theorywhich identifies various levels of a child's environment. The primary focus of this theory focuses on the quality and context of a child's environment. Bronfenbrenner suggests that as a child grows older, their interaction between the various levels of their environment grows more complex due to cognitive abilities expanding.
Lawrence Kohlberg 1927–1987 Wrote the theory of stages of moral development, which extended Piaget's findings of cognitive development and showed that they continue through the lifespan. Kohlberg's six stages follow Piaget's constructivist requirements in that those stages can not be skipped and it is very rare to regress in stages. Notable works: Moral Stages and Moralization: The Cognitive-Development Approach (1976) and Essays on Moral Development (1981)

Piaget's theory of cognitive development

Jean Piaget (1896–1980) believed that people move through stages of development that allow them to think in new, more complex ways.

Sensorimotor stage

The first stage in Piaget's stages of cognitive development is the sensorimotor stage. This stage lasts from birth to two years old. During this stage, behaviors lack a sense of thought and logic. Behaviors gradually move from acting upon inherited reflexes to interacting with the environment with a goal in mind and being able to represent the external world at the end.

The sensorimotor stage has been broken down into six sub stages that explain the gradual development of infants from birth to age 2. Once the child gains the ability to mentally represent reality, the child begins the transition to the preoperational stage of development.

Birth to one month

Each child is born with inherited reflexes that they use to gain knowledge and understanding about their environment. Examples of these reflexes include grasping and sucking.

1–4 months

Children repeat behaviors that happen unexpectedly because of their reflexes. For example, a child's finger comes in contact with the mouth and the child starts sucking on it. If the sensation is pleasurable to the child, then the child will attempt to recreate the behavior. Infants use their initial reflexes (grasping and sucking) to explore their environment and create schemas. Schemas are groups of similar actions or thoughts that are used repeatedly in response to the environment. Once a child begins to create schemas they use accommodation and assimilation to become progressively adapted to the world. Assimilation is when a child responds to a new event in a way that is consistent with an existing schema. For example, an infant may assimilate a new teddy bear into their putting things in their mouth scheme and use their reflexes to make the teddy bear go into their mouth. Accommodation is when a child either modifies an existing schema or forms an entirely new schema to deal with a new object or event. For example, an infant may have to open his or her mouth wider than usual to accommodate the teddy bear's paw.

5–8 months

The child has an experience with an external stimulus that they find pleasurable, so they try to recreate that experience. For example, a child accidentally hits the mobile above the crib and likes to watch it spin. When it stops the child begins to grab at the object to make it spin again. In this stage, habits are formed from general schemas that the infant has created but there is not yet, from the child's point of view, any differentiation between means and ends. Children cannot also focus on multiple tasks at once, and only focus on the task at hand. The child may create a habit of spinning the mobile in its crib, but they are still trying to find out methods to reach the mobile in order to get it to spin in the way that they find pleasurable. Once there is another distraction (say the parent walks in the room) the baby will no longer focus on the mobile. Toys should be given to infants that respond to a child's actions to help foster their investigative instincts. For example, a toy plays a song when you push one button, and then a picture pops up if you push another button.

8–12 months

Behaviors will be displayed for a reason rather than by chance. They begin to understand that one action can cause a reaction. They also begin to understand object permanence, which is the realization that objects continue to exist when removed from view. For example: The baby wants a rattle but the blanket is in the way. The baby moves the blanket to get the rattle. Now that the infant can understand that the object still exists, they can differentiate between the object, and the experience of the object. According to psychologist David Elkind, "An internal representation of the absent object is the earliest manifestation of the symbolic function which develops gradually during the second year of life whose activities dominate the next stage of mental growth."

12–18 months

Actions occur deliberately with some variation. For example, a baby drums on a pot with a wooden spoon, then drums on the floor, then on the table.

18–24 months

Children begin to build mental symbols and start to participate in pretend play. For example, a child is mixing ingredients together but does not have a spoon so they pretend to use one or use another object to replace the spoon. Symbolic thought is a representation of objects and events as mental entities or symbols which helps foster cognitive development and the formation of imagination. According to Piaget, the infant begins to act upon intelligence rather than habit at this point. The end product is established after the infant has pursued for the appropriate means. The means are formed from the schemas that are known by the child. The child is starting to learn how to use what it has learned in the first two years to develop and further explore their environment.

Preoperational stage

Lasts from 2 years of age until 6 or 7. It can be characterized in two somewhat different ways. In his early work, before he had developed his structuralist theory of cognition, Piaget described the child's thoughts during this period as being governed by principles such as egocentrism, animism, and other similar constructs. Egocentrism is when a child can only see a certain situation his or her own way. One cannot comprehend that other people have other views and perceptions of scenarios. Animism is when an individual gives a lifeless object human-like qualities. An individual usually believes that this object has human emotions, thoughts, and intentions. Once he had proposed his structuralist theory, Piaget characterized the preoperational child as lacking the cognitive structures possessed by the concrete operational child. The absence of these structures explains, in part, the behaviors Piaget had previously described as egocentric and animistic, for example, an inability to comprehend that another individual may have different emotional responses to similar experiences. During this stage children also become increasingly adept at using symbols as evidenced by the increase in playing and pretending.

Concrete operational stage

Lasts from 6 or 7 years until about 12 or 13. During this stage, the child's cognitive structures can be characterized by reality. Piaget argues that the same general principles can be discerned in a wide range of behaviors. One of the best-known achievements of this stage is conservation. In a typical conservation experiment a child is asked to judge whether or not two quantities are the same – such as two equal quantities of liquid in a short and tall glass. A preoperational child will typically judge the taller, thinner glass to contain more, while a concrete operational child will judge the amounts still to be the same. The ability to reason in this way reflects the development of a principle of conservation.

Formal operational stage

This stage lasts from 12 or 13 until adulthood, when people are advancing from logical reasoning with concrete examples to abstract examples. The need for concrete examples is no longer necessary because abstract thinking can be used instead. In this stage adolescents are also able to view themselves in the future and can picture the ideal life they would like to pursue. Some theorists believe the formal operational stage can be divided into two sub-categories: early formal operational and late formal operation thought. Early formal operational thoughts may be just fantasies, but as adolescents advance to late formal operational thought the life experiences they have encountered changes those fantasy thoughts to realistic thoughts.

Criticism

Many of Piaget's claims have fallen out of favor. For example, he claimed that young children cannot conserve numbers. However, further experiments showed that children did not really understand what was being asked of them. When the experiment is done with candies, and the children are asked which set they want rather than having to tell an adult which is more, they show no confusion about which group has more items. Piaget argues that the child cannot conserve numbers if they do not understand one-to-one correspondence. There needs to be more information and experiments on whether children understand numbers and quantities the way we do.

Piaget's theory of cognitive development ends at the formal operational stage that is usually developed in early adulthood. It does not take into account later stages of adult cognitive development as described by, for example, Harvard University professor Robert Kegan.

Other theoretical perspectives on cognitive development

Lev Vygotsky's Sociocultural Theory

Lev Vygotsky's (1896-1934) theory is based on social learning as the most important aspect of cognitive development. In Vygotsky's theory, adults are very important for young children's development. They help children learn through mediation, which is modeling and explaining concepts. Together, adults and children master concepts of their culture and activities. Vygotsky believed we get our complex mental activities through social learning. A significant part of Vygotsky's theory is based on the zone of proximal development, which he believes is when the most effective learning takes place. The Zone of proximal development is what a child cannot accomplish alone but can accomplish with the help of an MKO (more knowledgeable other). Vygotsky also believed culture is a very important part of cognitive development such as the language, writing and counting system used in that culture. Another aspect of Vygotsky’s theory is private speech. Private speech is when a person talks to themselves in order to help themselves problem solve. Scaffolding or providing support to a child and then slowly removing support and allowing the child to do more on their own over time is also an aspect of Vygotsky’s theory.

Speculated core systems of cognition

Empiricists study how these skills may be learned in such a short time. The debate is over whether these systems are learned by general-purpose learning devices or domain-specific cognition. Moreover, many modern cognitive developmental psychologists, recognizing that the term "innate" does not square with modern knowledge about epigenesis, neurobiological development, or learning, favor a non-nativist framework. Researchers who discuss "core systems" often speculate about differences in thinking and learning between proposed domains.

Researchers who posit a set of so-called "core domains" suggest that children have an innate sensitivity to specific kinds of patterns of information. Those commonly cited include:

Number

Infants appear to have two systems for dealing with numbers. One deals with small numbers, often called subitizing. Another deals with larger numbers in an approximate fashion.

Space

Very young children appear to have some skill in navigation. This basic ability to infer the direction and distance of unseen locations develops in ways that are not entirely clear. However, there is some evidence that it involves the development of complex language skills between 3 and 5 years. Also, there is evidence that this skill depends importantly on visual experience, because congenitally blind individuals have been found to have impaired abilities to infer new paths between familiar locations.

Visual perception

One of the original nativist versus empiricist debates was over depth perception. There is some evidence that children less than 72 hours old can perceive such complex things as biological motion. However, it is unclear how visual experience in the first few days contributes to this perception. There are far more elaborate aspects of visual perception that develop during infancy and beyond.

Essentialism

Young children seem to be predisposed to think of biological entities (e.g., animals and plants) in an essentialistic way. This means that they expect such entities (as opposed to, e.g., artifacts) to have many traits such as internal properties that are caused by some "essence" (such as, in our modern Western conceptual framework, the genome).

Language acquisition

A major, well-studied process and consequence of cognitive development is language acquisition. The traditional view was that this is the result of deterministic, human-specific genetic structures and processes. Other traditions, however, have emphasized the role of social experience in language learning. However, the relation of gene activity, experience, and language development is now recognized as incredibly complex and difficult to specify. Language development is sometimes separated into learning of phonology (systematic organization of sounds), morphology (structure of linguistic units—root words, affixes, parts of speech, intonation, etc.), syntax (rules of grammar within sentence structure), semantics (study of meaning), and discourse or pragmatics (relation between sentences). However, all of these aspects of language knowledge—which were originally posited by the linguist Noam Chomsky to be autonomous or separate—are now recognized to interact in complex ways.

Bilingualism

It was not until 1962 that bilingualism had been accepted as a contributing factor to cognitive development. There have been a number of studies showing how bilingualism contributes to the executive function of the brain, which is the main center at which cognitive development happens. According to Bialystok in “Bilingualism and the Development of Executive Function: The Role of Attention”, children who are bilingual, have to actively filter through the two different languages to select the one they need to use, which in turn makes the development stronger in that center.

Whorf's hypothesis

Benjamin Whorf (1897–1941), while working as a student of Edward Sapir, posited that a person's thinking depends on the structure and content of their social group's language. Per Whorf, language determines our thoughts and perceptions. For example, it used to be thought that the Greeks, who wrote left to right, thought differently than Egyptians since the Egyptians wrote right to left. Whorf's theory was so strict that he believed if a word is absent in a language, then the individual is unaware of the object's existence. This theory was played out in George Orwell's book, Animal Farm; the pig leaders slowly eliminated words from the citizen's vocabulary so that they were incapable of realizing what they were missing. The Whorfian hypothesis failed to recognize that people can still be aware of the concept or item, even though they lack efficient coding to quickly identify the target information.

Quine's bootstrapping hypothesis

Willard Van Orman Quine (1908–2000) argued that there are innate conceptual biases that enable the acquisition of language, concepts, and beliefs. Quine's theory follows nativist philosophical traditions, such as the European rationalist philosophers, for example Immanuel Kant.

Neo-Piagetian theories of cognitive development

Neo-Piagetian theories of cognitive development emphasized the role of information processing mechanisms in cognitive development, such as attention control and working memory. They suggested that progression along Piagetian stages or other levels of cognitive development is a function of strengthening of control mechanisms and enhancement of working memory storage capacity.

Lev Vygotsky vs. Jean Piaget

Unlike Jean Piaget, who believed development comes before learning, Vygotsky believed that learning comes before development and that one must learn first to be able to develop into a functioning human being. Vygotsky's theory is different from Piaget's theory of cognitive development in four ways. 1. Vygotsky believes culture affects cognitive development more. Piaget thinks that cognitive development is the same across the world, while Vygotsky has the idea that culture makes cognitive development different. 2. Social factors heavily influence cognitive development under Vygotsky's beliefs. Environment and parents the child has will play a big role in a child's cognitive development. The child learns through the Zone of Proximal Development with help from their parent. 3. Vygotsky believes that language is important in cognitive development. While Piaget considers thought as an important role, Vygotsky sees thought and language as different, but eventually coming together. Vygotsky emphasizes the role of inner speech being the first thing to cause cognitive development to form. 4. Cognitive development is strongly influenced by adults. Children observe adults in their life and gain knowledge about their specific culture based on things the adults around them do. They do this through mediation and scaffolding.

Neuroscience

During development, especially the first few years of life, children show interesting patterns of neural development and a high degree of neuroplasticity. Neuroplasticity, as explained by the World Health Organization, can be summed up in three points.

  1. Any adaptive mechanism used by the nervous system to repair itself after injury.
  2. Any means by which the nervous system can repair individually damaged central circuits.
  3. Any means by which the capacity of the central nervous system can adapt to new physiological conditions and environment.

The relation of brain development and cognitive development is extremely complex and, since the 1990s, has been a growing area of research.

Cognitive development and motor development may also be closely interrelated. When a person experiences a neurodevelopmental disorder and their cognitive development is disturbed, we often see adverse effects in motor development as well. Cerebellum, which is the part of brain that is most responsible for motor skills, has been shown to have significant importance in cognitive functions in the same way that prefrontal cortex has important duties in not only cognitive abilities but also development of motor skills. To support this, there is evidence of close co-activation of neocerebellum and dorsolateral prefrontal cortex in functional neuroimaging as well as abnormalities seen in both cerebellum and prefrontal cortex in the same developmental disorder. In this way, we see close interrelation of motor development and cognitive development and they cannot operate in their full capacity when either of them are impaired or delayed.

Cultural influences

From cultural psychologists' view, minds and culture shape each other. In other words, culture can influence brain structures which then influence our interpretation of the culture. These examples reveal cultural variations in neural responses:

Figure-line task (Hedden et al., 2008)

Behavioral research has shown that one's strength in independent or interdependent tasks differ based on their cultural context. In general, East Asian cultures are more interdependent whereas Western cultures are more independent. Hedden et al. assessed functional magnetic resonance imaging (fMRI) responses of East Asians and Americans while they performed independent (absolute) or interdependent (relative) tasks. The study showed that participants used regions of the brain associated with attentional control when they had to perform culturally incongruent tasks. In other words, neural paths used for the same task were different for Americans and East Asians (Hedden et al., 2008).

Transcultural neuroimaging studies (Han s. and Northoff G., 2008)

New studies in transcultural neuroimaging studies have demonstrated that one’s cultural background can influence the neural activity that underlies both high (for example, social cognition) and low (for example, perception) level cognitive functions. Studies demonstrated that groups that come from different cultures or that have been exposed to culturally different stimuli have differences in neural activity. For example, differences were found in that of the pre motor cortex during mental calculation and that of the VMPFC during trait judgements of one’s mother from people with different cultural backgrounds. In conclusion, since differences were found in both high-level and low-level cognition one can assume that our brain’s activity is strongly and, at least in part, constitutionally shaped by its sociocultural context (Han s. and Northoff G., 2008).

Kobayashi et al., 2007

Kobayashi et al. compared American-English monolingual and Japanese-English bilingual children's brain responses in understanding others' intentions through false-belief story and cartoon tasks. They found universal activation of the region bilateral ventromedial prefrontal cortex in theory of mind tasks. However, American children showed greater activity in the left inferior frontal gyrus during the tasks whereas Japanese children had greater activity in the right inferior frontal gyrus during the Japanese Theory of Mind tasks. In conclusion, these examples suggest that the brain's neural activities are not universal but are culture dependent.

Cognitive Development of Underrepresented Groups

Deaf and Hard-of-Hearing

Being Deaf or Hard-of-Hearing has been noted to impact cognitive development as hearing loss impacts social development, language acquisition, and the culture reacts to a deaf child. Cognitive development in academic achievement, reading development, language development, performance on standardized measures of intelligence, visual-spatial and memory skills, development of conceptual skills, and neuropsychological function are dependent upon the child's primary language of communication, either American Sign Language or English, as well as if the child is able to communicate and use the communication modality as a language. There is some research pointing to deficits in development of Theory of mind in children who are Deaf and Hard-of-Hearing which may be due to a lack of early conversational experience. Other research points to lower scores on the Wechsler Intelligence Scale for Children, especially in the Verbal Comprehension Index due differences in cultural knowledge acquisition.

Reason

From Wikipedia, the free encyclopedia

Reason is the capacity of consciously applying logic by drawing conclusions from new or existing information, with the aim of seeking the truth. It is closely associated with such characteristically human activities as philosophy, science, language, mathematics, and art, and is normally considered to be a distinguishing ability possessed by humans. Reason is sometimes referred to as rationality.

Reasoning is associated with the acts of thinking and cognition, and involves using one's intellect. The field of logic studies the ways in which humans can use formal reasoning to produce logically valid arguments. Reasoning may be subdivided into forms of logical reasoning, such as: deductive reasoning, inductive reasoning, and abductive reasoning. Aristotle drew a distinction between logical discursive reasoning (reason proper), and intuitive reasoning, in which the reasoning process through intuition—however valid—may tend toward the personal and the subjectively opaque. In some social and political settings logical and intuitive modes of reasoning may clash, while in other contexts intuition and formal reason are seen as complementary rather than adversarial. For example, in mathematics, intuition is often necessary for the creative processes involved with arriving at a formal proof, arguably the most difficult of formal reasoning tasks.

Reasoning, like habit or intuition, is one of the ways by which thinking moves from one idea to a related idea. For example, reasoning is the means by which rational individuals understand sensory information from their environments, or conceptualize abstract dichotomies such as cause and effect, truth and falsehood, or ideas regarding notions of good or evil. Reasoning, as a part of executive decision making, is also closely identified with the ability to self-consciously change, in terms of goals, beliefs, attitudes, traditions, and institutions, and therefore with the capacity for freedom and self-determination.

In contrast to the use of "reason" as an abstract noun, a reason is a consideration given which either explains or justifies events, phenomena, or behavior. Reasons justify decisions, reasons support explanations of natural phenomena; reasons can be given to explain the actions (conduct) of individuals.

Using reason, or reasoning, can also be described more plainly as providing good, or the best, reasons. For example, when evaluating a moral decision, "morality is, at the very least, the effort to guide one's conduct by reason—that is, doing what there are the best reasons for doing—while giving equal [and impartial] weight to the interests of all those affected by what one does."

Psychologists and cognitive scientists have attempted to study and explain how people reason, e.g. which cognitive and neural processes are engaged, and how cultural factors affect the inferences that people draw. The field of automated reasoning studies how reasoning may or may not be modeled computationally. Animal psychology considers the question of whether animals other than humans can reason.

Etymology and related words

In the English language and other modern European languages, "reason", and related words, represent words which have always been used to translate Latin and classical Greek terms in the sense of their philosophical usage.

  • The original Greek term was "λόγος" logos, the root of the modern English word "logic" but also a word which could mean for example "speech" or "explanation" or an "account" (of money handled).
  • As a philosophical term logos was translated in its non-linguistic senses in Latin as ratio. This was originally not just a translation used for philosophy, but was also commonly a translation for logos in the sense of an account of money.
  • French raison is derived directly from Latin, and this is the direct source of the English word "reason".

The earliest major philosophers to publish in English, such as Francis Bacon, Thomas Hobbes, and John Locke also routinely wrote in Latin and French, and compared their terms to Greek, treating the words "logos", "ratio", "raison" and "reason" as interchangeable. The meaning of the word "reason" in senses such as "human reason" also overlaps to a large extent with "rationality" and the adjective of "reason" in philosophical contexts is normally "rational", rather than "reasoned" or "reasonable". Some philosophers, Thomas Hobbes for example, also used the word ratiocination as a synonym for "reasoning".

Philosophical history

Francisco de Goya, The Sleep of Reason Produces Monsters (El sueño de la razón produce monstruos), c. 1797

The proposal that reason gives humanity a special position in nature has been argued to be a defining characteristic of western philosophy and later western modern science, starting with classical Greece. Philosophy can be described as a way of life based upon reason, and in the other direction, reason has been one of the major subjects of philosophical discussion since ancient times. Reason is often said to be reflexive, or "self-correcting", and the critique of reason has been a persistent theme in philosophy. It has been defined in different ways, at different times, by different thinkers about human nature.

Classical philosophy

For many classical philosophers, nature was understood teleologically, meaning that every type of thing had a definitive purpose that fit within a natural order that was itself understood to have aims. Perhaps starting with Pythagoras or Heraclitus, the cosmos is even said to have reason. Reason, by this account, is not just one characteristic that humans happen to have, and that influences happiness amongst other characteristics. Reason was considered of higher stature than other characteristics of human nature, such as sociability, because it is something humans share with nature itself, linking an apparently immortal part of the human mind with the divine order of the cosmos itself. Within the human mind or soul (psyche), reason was described by Plato as being the natural monarch which should rule over the other parts, such as spiritedness (thumos) and the passions. Aristotle, Plato's student, defined human beings as rational animals, emphasizing reason as a characteristic of human nature. He defined the highest human happiness or well being (eudaimonia) as a life which is lived consistently, excellently, and completely in accordance with reason.

The conclusions to be drawn from the discussions of Aristotle and Plato on this matter are amongst the most debated in the history of philosophy. But teleological accounts such as Aristotle's were highly influential for those who attempt to explain reason in a way that is consistent with monotheism and the immortality and divinity of the human soul. For example, in the neoplatonist account of Plotinus, the cosmos has one soul, which is the seat of all reason, and the souls of all individual humans are part of this soul. Reason is for Plotinus both the provider of form to material things, and the light which brings individuals souls back into line with their source.

Christian and Islamic philosophy

The classical view of reason, like many important Neoplatonic and Stoic ideas, was readily adopted by the early Church as the Church Fathers saw Greek Philosophy as an indispensable instrument given to mankind so that we may understand revelation. For example, the greatest among the early saint Church Fathers and Doctors of the Church such as Augustine of Hippo, Basil of Caesarea and Gregory of Nyssa were as much Neoplatonic philosophers as they were Christian theologians and adopted the Neoplatonic view of human reason together with the associated implications for our relationship to creation, to ourselves and to God. Such Neoplatonist accounts of the rational part of the human soul were also standard amongst medieval Islamic philosophers and remain important in Iranian philosophy. As European intellectualism recovered from the post-Roman dark ages, the Christian Patristic heritage and the influence of the great Islamic scholars such as Averroes and Avicenna produced the Scholastic (see Scholasticism) view of reason from which our modern idea of this concept has developed. Among the Scholastics who relied on the classical concept of reason for the development of their doctrines, none were more influential than Saint Thomas Aquinas, who put this concept at the heart of his Natural Law. In this doctrine, Thomas concludes that because humans have reason and because reason is a spark of the divine, every single human life is invaluable, all humans are equal and every human is born with an intrinsic and permanent set of basic rights. On this foundation, the idea of human rights would later be constructed by Spanish theologians at the School of Salamanca. Other Scholastics, such as Roger Bacon and Albertus Magnus, following the example of Islamic scholars such as Alhazen, emphasised reason an intrinsic human ability to decode the created order and the structures that underlie our experienced physical reality. This interpretation of reason was instrumental to the development of the scientific method in the early Universities of the high middle ages.

Subject-centred reason in early modern philosophy

The early modern era was marked by a number of significant changes in the understanding of reason, starting in Europe. One of the most important of these changes involved a change in the metaphysical understanding of human beings. Scientists and philosophers began to question the teleological understanding of the world. Nature was no longer assumed to be human-like, with its own aims or reason, and human nature was no longer assumed to work according to anything other than the same "laws of nature" which affect inanimate things. This new understanding eventually displaced the previous world view that derived from a spiritual understanding of the universe.

René Descartes

Accordingly, in the 17th century, René Descartes explicitly rejected the traditional notion of humans as "rational animals", suggesting instead that they are nothing more than "thinking things" along the lines of other "things" in nature. Any grounds of knowledge outside that understanding was, therefore, subject to doubt.

In his search for a foundation of all possible knowledge, Descartes deliberately decided to throw into doubt all knowledge – except that of the mind itself in the process of thinking:

At this time I admit nothing that is not necessarily true. I am therefore precisely nothing but a thinking thing; that is a mind, or intellect, or understanding, or reason – words of whose meanings I was previously ignorant.

This eventually became known as epistemological or "subject-centred" reason, because it is based on the knowing subject, who perceives the rest of the world and itself as a set of objects to be studied, and successfully mastered by applying the knowledge accumulated through such study. Breaking with tradition and many thinkers after him, Descartes explicitly did not divide the incorporeal soul into parts, such as reason and intellect, describing them as one indivisible incorporeal entity.

A contemporary of Descartes, Thomas Hobbes described reason as a broader version of "addition and subtraction" which is not limited to numbers. This understanding of reason is sometimes termed "calculative" reason. Similar to Descartes, Hobbes asserted that "No discourse whatsoever, can end in absolute knowledge of fact, past, or to come" but that "sense and memory" is absolute knowledge.

In the late 17th century, through the 18th century, John Locke and David Hume developed Descartes' line of thought still further. Hume took it in an especially skeptical direction, proposing that there could be no possibility of deducing relationships of cause and effect, and therefore no knowledge is based on reasoning alone, even if it seems otherwise.

Hume famously remarked that, "We speak not strictly and philosophically when we talk of the combat of passion and of reason. Reason is, and ought only to be the slave of the passions, and can never pretend to any other office than to serve and obey them." Hume also took his definition of reason to unorthodox extremes by arguing, unlike his predecessors, that human reason is not qualitatively different from either simply conceiving individual ideas, or from judgments associating two ideas, and that "reason is nothing but a wonderful and unintelligible instinct in our souls, which carries us along a certain train of ideas, and endows them with particular qualities, according to their particular situations and relations." It followed from this that animals have reason, only much less complex than human reason.

In the 18th century, Immanuel Kant attempted to show that Hume was wrong by demonstrating that a "transcendental" self, or "I", was a necessary condition of all experience. Therefore, suggested Kant, on the basis of such a self, it is in fact possible to reason both about the conditions and limits of human knowledge. And so long as these limits are respected, reason can be the vehicle of morality, justice, aesthetics, theories of knowledge (epistemology), and understanding.

Substantive and formal reason

In the formulation of Kant, who wrote some of the most influential modern treatises on the subject, the great achievement of reason (German: Vernunft) is that it is able to exercise a kind of universal law-making. Kant was able therefore to reformulate the basis of moral-practical, theoretical and aesthetic reasoning, on "universal" laws.

Here practical reasoning is the self-legislating or self-governing formulation of universal norms, and theoretical reasoning the way humans posit universal laws of nature.

Under practical reason, the moral autonomy or freedom of human beings depends on their ability to behave according to laws that are given to them by the proper exercise of that reason. This contrasted with earlier forms of morality, which depended on religious understanding and interpretation, or nature for their substance.

According to Kant, in a free society each individual must be able to pursue their goals however they see fit, so long as their actions conform to principles given by reason. He formulated such a principle, called the "categorical imperative", which would justify an action only if it could be universalized:

Act only according to that maxim whereby you can, at the same time, will that it should become a universal law.

In contrast to Hume then, Kant insists that reason itself (German Vernunft) has natural ends itself, the solution to the metaphysical problems, especially the discovery of the foundations of morality. Kant claimed that this problem could be solved with his "transcendental logic" which unlike normal logic is not just an instrument, which can be used indifferently, as it was for Aristotle, but a theoretical science in its own right and the basis of all the others.

According to Jürgen Habermas, the "substantive unity" of reason has dissolved in modern times, such that it can no longer answer the question "How should I live?" Instead, the unity of reason has to be strictly formal, or "procedural". He thus described reason as a group of three autonomous spheres (on the model of Kant's three critiques):

  1. Cognitive–instrumental reason is the kind of reason employed by the sciences. It is used to observe events, to predict and control outcomes, and to intervene in the world on the basis of its hypotheses;
  2. Moral–practical reason is what we use to deliberate and discuss issues in the moral and political realm, according to universalizable procedures (similar to Kant's categorical imperative); and
  3. Aesthetic reason is typically found in works of art and literature, and encompasses the novel ways of seeing the world and interpreting things that those practices embody.

For Habermas, these three spheres are the domain of experts, and therefore need to be mediated with the "lifeworld" by philosophers. In drawing such a picture of reason, Habermas hoped to demonstrate that the substantive unity of reason, which in pre-modern societies had been able to answer questions about the good life, could be made up for by the unity of reason's formalizable procedures.

The critique of reason

Hamann, Herder, Kant, Hegel, Kierkegaard, Nietzsche, Heidegger, Foucault, Rorty, and many other philosophers have contributed to a debate about what reason means, or ought to mean. Some, like Kierkegaard, Nietzsche, and Rorty, are skeptical about subject-centred, universal, or instrumental reason, and even skeptical toward reason as a whole. Others, including Hegel, believe that it has obscured the importance of intersubjectivity, or "spirit" in human life, and attempt to reconstruct a model of what reason should be.

Some thinkers, e.g. Foucault, believe there are other forms of reason, neglected but essential to modern life, and to our understanding of what it means to live a life according to reason.

In the last several decades, a number of proposals have been made to "re-orient" this critique of reason, or to recognize the "other voices" or "new departments" of reason:

For example, in opposition to subject-centred reason, Habermas has proposed a model of communicative reason that sees it as an essentially cooperative activity, based on the fact of linguistic intersubjectivity.

Nikolas Kompridis has proposed a widely encompassing view of reason as "that ensemble of practices that contributes to the opening and preserving of openness" in human affairs, and a focus on reason's possibilities for social change.

The philosopher Charles Taylor, influenced by the 20th century German philosopher Martin Heidegger, has proposed that reason ought to include the faculty of disclosure, which is tied to the way we make sense of things in everyday life, as a new "department" of reason.

In the essay "What is Enlightenment?", Michel Foucault proposed a concept of critique based on Kant's distinction between "private" and "public" uses of reason. This distinction, as suggested, has two dimensions:

  • Private reason is the reason that is used when an individual is "a cog in a machine" or when one "has a role to play in society and jobs to do: to be a soldier, to have taxes to pay, to be in charge of a parish, to be a civil servant".
  • Public reason is the reason used "when one is reasoning as a reasonable being (and not as a cog in a machine), when one is reasoning as a member of reasonable humanity". In these circumstances, "the use of reason must be free and public."

Reason compared to related concepts

Compared to logic

The terms logic or logical are sometimes used as if they were identical with the term reason or with the concept of being rational, or sometimes logic is seen as the most pure or the defining form of reason: "Logic is about reasoning—about going from premises to a conclusion. ... When you do logic, you try to clarify reasoning and separate good from bad reasoning." In modern economics, rational choice is assumed to equate to logically consistent choice.

Reason and logic can however be thought of as distinct, although logic is one important aspect of reason. Author Douglas Hofstadter, in Gödel, Escher, Bach, characterizes the distinction in this way: Logic is done inside a system while reason is done outside the system by such methods as skipping steps, working backward, drawing diagrams, looking at examples, or seeing what happens if you change the rules of the system. Psychologists Mark H. Bickard and Robert L. Campbell argued that "rationality cannot be simply assimilated to logicality"; they noted that "human knowledge of logic and logical systems has developed" over time through reasoning, and logical systems "can't construct new logical systems more powerful than themselves", so reasoning and rationality must involve more than a system of logic. Psychologist David Moshman, citing Bickhard and Campbell, argued for a "metacognitive conception of rationality" in which a person's development of reason "involves increasing consciousness and control of logical and other inferences".

Reason is a type of thought, and logic involves the attempt to describe a system of formal rules or norms of appropriate reasoning. The oldest surviving writing to explicitly consider the rules by which reason operates are the works of the Greek philosopher Aristotle, especially Prior Analysis and Posterior Analysis. Although the Ancient Greeks had no separate word for logic as distinct from language and reason, Aristotle's newly coined word "syllogism" (syllogismos) identified logic clearly for the first time as a distinct field of study. When Aristotle referred to "the logical" (hē logikē), he was referring more broadly to rational thought.

Reason compared to cause-and-effect thinking, and symbolic thinking

As pointed out by philosophers such as Hobbes, Locke and Hume, some animals are also clearly capable of a type of "associative thinking", even to the extent of associating causes and effects. A dog once kicked, can learn how to recognize the warning signs and avoid being kicked in the future, but this does not mean the dog has reason in any strict sense of the word. It also does not mean that humans acting on the basis of experience or habit are using their reason.

Human reason requires more than being able to associate two ideas, even if those two ideas might be described by a reasoning human as a cause and an effect, perceptions of smoke, for example, and memories of fire. For reason to be involved, the association of smoke and the fire would have to be thought through in a way which can be explained, for example as cause and effect. In the explanation of Locke, for example, reason requires the mental use of a third idea in order to make this comparison by use of syllogism.

More generally, reason in the strict sense requires the ability to create and manipulate a system of symbols, as well as indices and icons, according to Charles Sanders Peirce, the symbols having only a nominal, though habitual, connection to either smoke or fire. One example of such a system of artificial symbols and signs is language.

The connection of reason to symbolic thinking has been expressed in different ways by philosophers. Thomas Hobbes described the creation of "Markes, or Notes of remembrance" (Leviathan Ch. 4) as speech. He used the word speech as an English version of the Greek word logos so that speech did not need to be communicated. When communicated, such speech becomes language, and the marks or notes or remembrance are called "Signes" by Hobbes. Going further back, although Aristotle is a source of the idea that only humans have reason (logos), he does mention that animals with imagination, for whom sense perceptions can persist, come closest to having something like reasoning and nous, and even uses the word "logos" in one place to describe the distinctions which animals can perceive in such cases.

Reason, imagination, mimesis, and memory

Reason and imagination rely on similar mental processes. Imagination is not only found in humans. Aristotle, for example, stated that phantasia (imagination: that which can hold images or phantasmata) and phronein (a type of thinking that can judge and understand in some sense) also exist in some animals. According to him, both are related to the primary perceptive ability of animals, which gathers the perceptions of different senses and defines the order of the things that are perceived without distinguishing universals, and without deliberation or logos. But this is not yet reason, because human imagination is different.

The recent modern writings of Terrence Deacon and Merlin Donald, writing about the origin of language, also connect reason connected to not only language, but also mimesis. More specifically they describe the ability to create language as part of an internal modeling of reality specific to humankind. Other results are consciousness, and imagination or fantasy. In contrast, modern proponents of a genetic predisposition to language itself include Noam Chomsky and Steven Pinker, to whom Donald and Deacon can be contrasted.

As reason is symbolic thinking, and peculiarly human, then this implies that humans have a special ability to maintain a clear consciousness of the distinctness of "icons" or images and the real things they represent. Starting with a modern author, Merlin Donald writes

A dog might perceive the "meaning" of a fight that was realistically play-acted by humans, but it could not reconstruct the message or distinguish the representation from its referent (a real fight). [...] Trained apes are able to make this distinction; young children make this distinction early – hence, their effortless distinction between play-acting an event and the event itself

In classical descriptions, an equivalent description of this mental faculty is eikasia, in the philosophy of Plato. This is the ability to perceive whether a perception is an image of something else, related somehow but not the same, and therefore allows humans to perceive that a dream or memory or a reflection in a mirror is not reality as such. What Klein refers to as dianoetic eikasia is the eikasia concerned specifically with thinking and mental images, such as those mental symbols, icons, signes, and marks discussed above as definitive of reason. Explaining reason from this direction: human thinking is special in the way that we often understand visible things as if they were themselves images of our intelligible "objects of thought" as "foundations" (hypothēses in Ancient Greek). This thinking (dianoia) is "...an activity which consists in making the vast and diffuse jungle of the visible world depend on a plurality of more 'precise' noēta".

Both Merlin Donald and the Socratic authors such as Plato and Aristotle emphasize the importance of mimesis, often translated as imitation or representation. Donald writes

Imitation is found especially in monkeys and apes [... but ...] Mimesis is fundamentally different from imitation and mimicry in that it involves the invention of intentional representations. [...] Mimesis is not absolutely tied to external communication.

Mimēsis is a concept, now popular again in academic discussion, that was particularly prevalent in Plato's works, and within Aristotle, it is discussed mainly in the Poetics. In Michael Davis's account of the theory of man in this work.

It is the distinctive feature of human action, that whenever we choose what we do, we imagine an action for ourselves as though we were inspecting it from the outside. Intentions are nothing more than imagined actions, internalizings of the external. All action is therefore imitation of action; it is poetic...

Donald like Plato (and Aristotle, especially in On Memory and Recollection), emphasizes the peculiarity in humans of voluntary initiation of a search through one's mental world. The ancient Greek anamnēsis, normally translated as "recollection" was opposed to mneme or memory. Memory, shared with some animals, requires a consciousness not only of what happened in the past, but also that something happened in the past, which is in other words a kind of eikasia "...but nothing except man is able to recollect." Recollection is a deliberate effort to search for and recapture something once known. Klein writes that, "To become aware of our having forgotten something means to begin recollecting." Donald calls the same thing autocueing, which he explains as follows: "Mimetic acts are reproducible on the basis of internal, self-generated cues. This permits voluntary recall of mimetic representations, without the aid of external cues – probably the earliest form of representational thinking."

In a celebrated paper in modern times, the fantasy author and philologist J.R.R. Tolkien wrote in his essay "On Fairy Stories" that the terms "fantasy" and "enchantment" are connected to not only "....the satisfaction of certain primordial human desires...." but also "...the origin of language and of the mind".

Logical reasoning methods and argumentation

A subdivision of philosophy is logic. Logic is the study of reasoning. Looking at logical categorizations of different types of reasoning, the traditional main division made in philosophy is between deductive reasoning and inductive reasoning. Formal logic has been described as the science of deduction. The study of inductive reasoning is generally carried out within the field known as informal logic or critical thinking.

Deductive reasoning

Deduction is a form of reasoning in which a conclusion follows necessarily from the stated premises. A deduction is also the conclusion reached by a deductive reasoning process. One classic example of deductive reasoning is that found in syllogisms like the following:

  • Premise 1: All humans are mortal.
  • Premise 2: Socrates is a human.
  • Conclusion: Socrates is mortal.

The reasoning in this argument is deductively valid because there is no way in which the premises, 1 and 2, could be true and the conclusion, 3, be false.

Inductive reasoning

Induction is a form of inference producing propositions about unobserved objects or types, either specifically or generally, based on previous observation. It is used to ascribe properties or relations to objects or types based on previous observations or experiences, or to formulate general statements or laws based on limited observations of recurring phenomenal patterns.

Inductive reasoning contrasts strongly with deductive reasoning in that, even in the best, or strongest, cases of inductive reasoning, the truth of the premises does not guarantee the truth of the conclusion. Instead, the conclusion of an inductive argument follows with some degree of probability. Relatedly, the conclusion of an inductive argument contains more information than is already contained in the premises. Thus, this method of reasoning is ampliative.

A classic example of inductive reasoning comes from the empiricist David Hume:

  • Premise: The sun has risen in the east every morning up until now.
  • Conclusion: The sun will also rise in the east tomorrow.

Analogical reasoning

Analogical reasoning is a form of inductive reasoning from a particular to a particular. It is often used in case-based reasoning, especially legal reasoning. An example follows:

  • Premise 1: Socrates is human and mortal.
  • Premise 2: Plato is human.
  • Conclusion: Plato is mortal.

Analogical reasoning is a weaker form of inductive reasoning from a single example, because inductive reasoning typically uses a large number of examples to reason from the particular to the general. Analogical reasoning often leads to wrong conclusions. For example:

  • Premise 1: Socrates is human and male.
  • Premise 2: Ada Lovelace is human.
  • Conclusion: Therefore Ada Lovelace is male.

Abductive reasoning

Abductive reasoning, or argument to the best explanation, is a form of reasoning that doesn't fit in deductive or inductive, since it starts with incomplete set of observations and proceeds with likely possible explanations so the conclusion in an abductive argument does not follow with certainty from its premises and concerns something unobserved. What distinguishes abduction from the other forms of reasoning is an attempt to favour one conclusion above others, by subjective judgement or attempting to falsify alternative explanations or by demonstrating the likelihood of the favoured conclusion, given a set of more or less disputable assumptions. For example, when a patient displays certain symptoms, there might be various possible causes, but one of these is preferred above others as being more probable.

Fallacious reasoning

Flawed reasoning in arguments is known as fallacious reasoning. Bad reasoning within arguments can be because it commits either a formal fallacy or an informal fallacy.

Formal fallacies occur when there is a problem with the form, or structure, of the argument. The word "formal" refers to this link to the form of the argument. An argument that contains a formal fallacy will always be invalid.

An informal fallacy is an error in reasoning that occurs due to a problem with the content, rather than mere structure, of the argument.

Traditional problems raised concerning reason

Philosophy is sometimes described as a life of reason, with normal human reason pursued in a more consistent and dedicated way than usual. Two categories of problem concerning reason have long been discussed by philosophers concerning reason, essentially being reasonings about reasoning itself as a human aim, or philosophizing about philosophizing. The first question is concerning whether we can be confident that reason can achieve knowledge of truth better than other ways of trying to achieve such knowledge. The other question is whether a life of reason, a life that aims to be guided by reason, can be expected to achieve a happy life more so than other ways of life (whether such a life of reason results in knowledge or not).

Reason versus truth, and "first principles"

Since classical times a question has remained constant in philosophical debate (which is sometimes seen as a conflict between movements called Platonism and Aristotelianism) concerning the role of reason in confirming truth. People use logic, deduction, and induction, to reach conclusions they think are true. Conclusions reached in this way are considered, according to Aristotle, more certain than sense perceptions on their own. On the other hand, if such reasoned conclusions are only built originally upon a foundation of sense perceptions, then, our most logical conclusions can never be said to be certain because they are built upon the very same fallible perceptions they seek to better.

This leads to the question of what types of first principles, or starting points of reasoning, are available for someone seeking to come to true conclusions. In Greek, "first principles" are archai, "starting points", and the faculty used to perceive them is sometimes referred to in Aristotle and Plato as nous which was close in meaning to awareness or consciousness.

Empiricism (sometimes associated with Aristotle but more correctly associated with British philosophers such as John Locke and David Hume, as well as their ancient equivalents such as Democritus) asserts that sensory impressions are the only available starting points for reasoning and attempting to attain truth. This approach always leads to the controversial conclusion that absolute knowledge is not attainable. Idealism, (associated with Plato and his school), claims that there is a "higher" reality, from which certain people can directly arrive at truth without needing to rely only upon the senses, and that this higher reality is therefore the primary source of truth.

Philosophers such as Plato, Aristotle, Al-Farabi, Avicenna, Averroes, Maimonides, Aquinas and Hegel are sometimes said to have argued that reason must be fixed and discoverable—perhaps by dialectic, analysis, or study. In the vision of these thinkers, reason is divine or at least has divine attributes. Such an approach allowed religious philosophers such as Thomas Aquinas and Étienne Gilson to try to show that reason and revelation are compatible. According to Hegel, "...the only thought which Philosophy brings with it to the contemplation of History, is the simple conception of reason; that reason is the Sovereign of the World; that the history of the world, therefore, presents us with a rational process."

Since the 17th century rationalists, reason has often been taken to be a subjective faculty, or rather the unaided ability (pure reason) to form concepts. For Descartes, Spinoza and Leibniz, this was associated with mathematics. Kant attempted to show that pure reason could form concepts (time and space) that are the conditions of experience. Kant made his argument in opposition to Hume, who denied that reason had any role to play in experience.

Reason versus emotion or passion

After Plato and Aristotle, western literature often treated reason as being the faculty that trained the passions and appetites. Stoic philosophy, by contrast, claimed most emotions were merely false judgements. According to the stoics the only good is virtue, and the only evil is vice, therefore emotions that judged things other than vice to be bad (such as fear or distress), or things other than virtue to be good (such as greed) were simply false judgements and should be discarded (though positive emotions based on true judgements, such as kindness, were acceptable). After the critiques of reason in the early Enlightenment the appetites were rarely discussed or conflated with the passions. Some Enlightenment camps took after the Stoics to say Reason should oppose Passion rather than order it, while others like the Romantics believed that Passion displaces Reason, as in the maxim "follow your heart".

Reason has been seen as cold, an "enemy of mystery and ambiguity", a slave, or judge, of the passions, notably in the work of David Hume, and more recently of Freud. Reasoning which claims that the object of a desire is demanded by logic alone is called rationalization.

Rousseau first proposed, in his second Discourse, that reason and political life is not natural and possibly harmful to mankind. He asked what really can be said about what is natural to mankind. What, other than reason and civil society, "best suits his constitution"? Rousseau saw "two principles prior to reason" in human nature. First we hold an intense interest in our own well-being. Secondly we object to the suffering or death of any sentient being, especially one like ourselves. These two passions lead us to desire more than we could achieve. We become dependent upon each other, and on relationships of authority and obedience. This effectively puts the human race into slavery. Rousseau says that he almost dares to assert that nature does not destine men to be healthy. According to Richard Velkley, "Rousseau outlines certain programs of rational self-correction, most notably the political legislation of the Contrat Social and the moral education in Émile. All the same, Rousseau understands such corrections to be only ameliorations of an essentially unsatisfactory condition, that of socially and intellectually corrupted humanity."

This quandary presented by Rousseau led to Kant's new way of justifying reason as freedom to create good and evil. These therefore are not to be blamed on nature or God. In various ways, German Idealism after Kant, and major later figures such Nietzsche, Bergson, Husserl, Scheler, and Heidegger, remain preoccupied with problems coming from the metaphysical demands or urges of reason. The influence of Rousseau and these later writers is also large upon art and politics. Many writers (such as Nikos Kazantzakis) extol passion and disparage reason. In politics modern nationalism comes from Rousseau's argument that rationalist cosmopolitanism brings man ever further from his natural state.

Another view on reason and emotion was proposed in the 1994 book titled Descartes' Error by Antonio Damasio. In it, Damasio presents the "Somatic Marker Hypothesis" which states that emotions guide behavior and decision-making. Damasio argues that these somatic markers (known collectively as "gut feelings") are "intuitive signals" that direct our decision making processes in a certain way that cannot be solved with rationality alone. Damasio further argues that rationality requires emotional input in order to function.

Reason versus faith or tradition

There are many religious traditions, some of which are explicitly fideist and others of which claim varying degrees of rationalism. Secular critics sometimes accuse all religious adherents of irrationality, since they claim such adherents are guilty of ignoring, suppressing, or forbidding some kinds of reasoning concerning some subjects (such as religious dogmas, moral taboos, etc.). Though the theologies and religions such as classical monotheism typically do not admit to be irrational, there is often a perceived conflict or tension between faith and tradition on the one hand, and reason on the other, as potentially competing sources of wisdom, law and truth.

Religious adherents sometimes respond by arguing that faith and reason can be reconciled, or have different non-overlapping domains, or that critics engage in a similar kind of irrationalism:

  • Reconciliation: Philosopher Alvin Plantinga argues that there is no real conflict between reason and classical theism because classical theism explains (among other things) why the universe is intelligible and why reason can successfully grasp it.
  • Non-overlapping magisteria: Evolutionary biologist Stephen Jay Gould argues that there need not be conflict between reason and religious belief because they are each authoritative in their own domain (or "magisterium"). If so, reason can work on those problems over which it has authority while other sources of knowledge or opinion can have authority on the big questions.
  • Tu quoque: Philosophers Alasdair MacIntyre and Charles Taylor argue that those critics of traditional religion who are adherents of secular liberalism are also sometimes guilty of ignoring, suppressing, and forbidding some kinds of reasoning about subjects. Similarly, philosophers of science such as Paul Feyarabend argue that scientists sometimes ignore or suppress evidence contrary to the dominant paradigm.
  • Unification: Theologian Joseph Ratzinger, later Benedict XVI, asserted that "Christianity has understood itself as the religion of the Logos, as the religion according to reason," referring to John 1:Ἐν ἀρχῇ ἦν ὁ λόγος, usually translated as "In the beginning was the Word (Logos)." Thus, he said that the Christian faith is "open to all that is truly rational", and that the rationality of Western Enlightenment "is of Christian origin".

Some commentators have claimed that Western civilization can be almost defined by its serious testing of the limits of tension between "unaided" reason and faith in "revealed" truths—figuratively summarized as Athens and Jerusalem, respectively. Leo Strauss spoke of a "Greater West" that included all areas under the influence of the tension between Greek rationalism and Abrahamic revelation, including the Muslim lands. He was particularly influenced by the great Muslim philosopher Al-Farabi. To consider to what extent Eastern philosophy might have partaken of these important tensions, Strauss thought it best to consider whether dharma or tao may be equivalent to Nature (by which we mean physis in Greek). According to Strauss the beginning of philosophy involved the "discovery or invention of nature" and the "pre-philosophical equivalent of nature" was supplied by "such notions as 'custom' or 'ways'", which appear to be really universal in all times and places. The philosophical concept of nature or natures as a way of understanding archai (first principles of knowledge) brought about a peculiar tension between reasoning on the one hand, and tradition or faith on the other.

Although there is this special history of debate concerning reason and faith in the Islamic, Christian and Jewish traditions, the pursuit of reason is sometimes argued to be compatible with the other practice of other religions of a different nature, such as Hinduism, because they do not define their tenets in such an absolute way.

Reason in particular fields of study

Psychology and cognitive science

Scientific research into reasoning is carried out within the fields of psychology and cognitive science. Psychologists attempt to determine whether or not people are capable of rational thought in a number of different circumstances.

Assessing how well someone engages in reasoning is the project of determining the extent to which the person is rational or acts rationally. It is a key research question in the psychology of reasoning and cognitive science of reasoning. Rationality is often divided into its respective theoretical and practical counterparts.

Behavioral experiments on human reasoning

Experimental cognitive psychologists carry out research on reasoning behaviour. Such research may focus, for example, on how people perform on tests of reasoning such as intelligence or IQ tests, or on how well people's reasoning matches ideals set by logic (see, for example, the Wason test). Experiments examine how people make inferences from conditionals e.g., If A then B and how they make inferences about alternatives, e.g., A or else B. They test whether people can make valid deductions about spatial and temporal relations, e.g., A is to the left of B, or A happens after B, and about quantified assertions, e.g., All the A are B. Experiments investigate how people make inferences about factual situations, hypothetical possibilities, probabilities, and counterfactual situations.

Developmental studies of children's reasoning

Developmental psychologists investigate the development of reasoning from birth to adulthood. Piaget's theory of cognitive development was the first complete theory of reasoning development. Subsequently, several alternative theories were proposed, including the neo-Piagetian theories of cognitive development.

Neuroscience of reasoning

The biological functioning of the brain is studied by neurophysiologists, cognitive neuroscientists and neuropsychologists. Research in this area includes research into the structure and function of normally functioning brains, and of damaged or otherwise unusual brains. In addition to carrying out research into reasoning, some psychologists, for example, clinical psychologists and psychotherapists work to alter people's reasoning habits when they are unhelpful.

Computer science

Automated reasoning

In artificial intelligence and computer science, scientists study and use automated reasoning for diverse applications including automated theorem proving the formal semantics of programming languages, and formal specification in software engineering.

Meta-reasoning

Meta-reasoning is reasoning about reasoning. In computer science, a system performs meta-reasoning when it is reasoning about its own operation. This requires a programming language capable of reflection, the ability to observe and modify its own structure and behaviour.

Evolution of reason

Dan Sperber believes that reasoning in groups is more effective and promotes their evolutionary fitness.

A species could benefit greatly from better abilities to reason about, predict and understand the world. French social and cognitive scientists Dan Sperber and Hugo Mercier argue that there could have been other forces driving the evolution of reason. They point out that reasoning is very difficult for humans to do effectively, and that it is hard for individuals to doubt their own beliefs (confirmation bias). Reasoning is most effective when it is done as a collective – as demonstrated by the success of projects like science. They suggest that there are not just individual, but group selection pressures at play. Any group that managed to find ways of reasoning effectively would reap benefits for all its members, increasing their fitness. This could also help explain why humans, according to Sperber, are not optimized to reason effectively alone. Their argumentative theory of reasoning claims that reason may have more to do with winning arguments than with the search for the truth.

Reason in political philosophy and ethics

Aristotle famously described reason (with language) as a part of human nature, which means that it is best for humans to live "politically" meaning in communities of about the size and type of a small city state (polis in Greek). For example...

It is clear, then, that a human being is more of a political [politikon = of the polis] animal [zōion] than is any bee or than any of those animals that live in herds. For nature, as we say, makes nothing in vain, and humans are the only animals who possess reasoned speech [logos]. Voice, of course, serves to indicate what is painful and pleasant; that is why it is also found in other animals, because their nature has reached the point where they can perceive what is painful and pleasant and express these to each other. But speech [logos] serves to make plain what is advantageous and harmful and so also what is just and unjust. For it is a peculiarity of humans, in contrast to the other animals, to have perception of good and bad, just and unjust, and the like; and the community in these things makes a household or city [polis]. [...] By nature, then, the drive for such a community exists in everyone, but the first to set one up is responsible for things of very great goodness. For as humans are the best of all animals when perfected, so they are the worst when divorced from law and right. The reason is that injustice is most difficult to deal with when furnished with weapons, and the weapons a human being has are meant by nature to go along with prudence and virtue, but it is only too possible to turn them to contrary uses. Consequently, if a human being lacks virtue, he is the most unholy and savage thing, and when it comes to sex and food, the worst. But justice is something political [to do with the polis], for right is the arrangement of the political community, and right is discrimination of what is just. (Aristotle's Politics 1253a 1.2. Peter Simpson's translation, with Greek terms inserted in square brackets.)

The concept of human nature being fixed in this way, implied, in other words, that we can define what type of community is always best for people. This argument has remained a central argument in all political, ethical and moral thinking since then, and has become especially controversial since firstly Rousseau's Second Discourse, and secondly, the Theory of Evolution. Already in Aristotle there was an awareness that the polis had not always existed and had needed to be invented or developed by humans themselves. The household came first, and the first villages and cities were just extensions of that, with the first cities being run as if they were still families with Kings acting like fathers.

Friendship [philia] seems to prevail [in] man and woman according to nature [kata phusin]; for people are by nature [tēi phusei] pairing [sunduastikon] more than political [politikon = of the polis], in as much as the household [oikos] is prior [proteron = earlier] and more necessary than the polis and making children is more common [koinoteron] with the animals. In the other animals, community [koinōnia] goes no further than this, but people live together [sumoikousin] not only for the sake of making children, but also for the things for life; for from the start the functions [erga] are divided, and are different [for] man and woman. Thus they supply each other, putting their own into the common [eis to koinon]. It is for these [reasons] that both utility [chrēsimon] and pleasure [hēdu] seem to be found in this kind of friendship. (Nicomachean Ethics, VIII.12.1162a. Rough literal translation with Greek terms shown in square brackets.)

Rousseau in his Second Discourse finally took the shocking step of claiming that this traditional account has things in reverse: with reason, language and rationally organized communities all having developed over a long period of time merely as a result of the fact that some habits of cooperation were found to solve certain types of problems, and that once such cooperation became more important, it forced people to develop increasingly complex cooperation—often only to defend themselves from each other.

In other words, according to Rousseau, reason, language and rational community did not arise because of any conscious decision or plan by humans or gods, nor because of any pre-existing human nature. As a result, he claimed, living together in rationally organized communities like modern humans is a development with many negative aspects compared to the original state of man as an ape. If anything is specifically human in this theory, it is the flexibility and adaptability of humans. This view of the animal origins of distinctive human characteristics later received support from Charles Darwin's Theory of Evolution.

The two competing theories concerning the origins of reason are relevant to political and ethical thought because, according to the Aristotelian theory, a best way of living together exists independently of historical circumstances. According to Rousseau, we should even doubt that reason, language and politics are a good thing, as opposed to being simply the best option given the particular course of events that lead to today. Rousseau's theory, that human nature is malleable rather than fixed, is often taken to imply, for example by Karl Marx, a wider range of possible ways of living together than traditionally known.

However, while Rousseau's initial impact encouraged bloody revolutions against traditional politics, including both the French Revolution and the Russian Revolution, his own conclusions about the best forms of community seem to have been remarkably classical, in favor of city-states such as Geneva, and rural living.

Ocean temperature

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Ocean_temperature Graph showing ocean tempe...