Search This Blog

Thursday, December 18, 2025

Unconscious mind

From Wikipedia, the free encyclopedia

In psychoanalysis and other psychological theories, the unconscious mind (or the unconscious) is the part of the psyche that is not available to introspection. Although these processes exist beneath the surface of conscious awareness, they are thought to exert an effect on conscious thought processes and behavior. The term was coined by the 18th-century German Romantic philosopher Friedrich Schelling and later introduced into English by the poet and essayist Samuel Taylor Coleridge.

The emergence of the concept of the unconscious in psychology and general culture was mainly due to the work of Austrian neurologist and psychoanalyst Sigmund Freud. In psychoanalytic theory, the unconscious mind consists of ideas and drives that have been subject to the mechanism of repression: anxiety-producing impulses in childhood are barred from consciousness, but do not cease to exist, and exert a constant pressure in the direction of consciousness. However, the content of the unconscious is only knowable to consciousness through its representation in a disguised or distorted form, by way of dreams and neurotic symptoms, as well as in slips of the tongue and jokes. The psychoanalyst seeks to interpret these conscious manifestations in order to understand the nature of the repressed.

The unconscious mind can be seen as the source of dreams and automatic thoughts (those that appear without any apparent cause), the repository of forgotten memories (that may still be accessible to consciousness at some later time), and the locus of implicit knowledge (the things that we have learned so well that we do them without thinking). Phenomena related to semi-consciousness include awakening, implicit memory, subliminal messages, trances, hypnagogia and hypnosis. While sleep, sleepwalking, dreaming, delirium and comas may signal the presence of unconscious processes, these processes are seen as symptoms rather than the unconscious mind itself.

Some critics have doubted the existence of the unconscious altogether.

Historical overview

German

The term "unconscious" (German: unbewusst) was coined by the 18th-century German Romantic philosopher Friedrich Schelling (in his System of Transcendental Idealism, ch. 6, § 3) and later introduced into English by the poet and essayist Samuel Taylor Coleridge (in his Biographia Literaria). Some rare earlier instances of the term "unconsciousness" (Unbewußtseyn) can be found in the work of the 18th-century German physician and philosopher Ernst Platner.

Vedas

Influences on thinking that originate from outside an individual's consciousness were reflected in the ancient ideas of temptation, divine inspiration, and the predominant role of the gods in affecting motives and actions. The idea of internalised unconscious processes in the mind was present in antiquity, and has been explored across a wide variety of cultures. Unconscious aspects of mentality were referred to between 2,500 and 600 BC in the Hindu texts known as the Vedas, found today in Ayurvedic medicine.

Paracelsus

Paracelsus is credited as the first to make mention of an unconscious aspect of cognition in his work Von den Krankheiten (translates as "About illnesses", 1567), and his clinical methodology created a cogent system that is regarded by some as the beginning of modern scientific psychology.

Shakespeare

William Shakespeare explored the role of the unconscious in many of his plays, without naming it as such.

Philosophy

In his work Anthropology, philosopher Immanuel Kant was one of the first to discuss the subject of unconscious ideas.

Western philosophers such as Arthur SchopenhauerBaruch Spinoza, Gottfried Wilhelm LeibnizJohann Gottlieb Fichte, Georg Wilhelm Friedrich Hegel, Karl Robert Eduard von Hartmann, Carl Gustav Carus, Søren Aabye Kierkegaard, Friedrich Wilhelm Nietzsche and Thomas Carlyle used the word unconscious.

In 1880 at the Sorbonne, Edmond Colsenet defended a philosophy thesis (PhD) on the unconscious. Elie Rabier and Alfred Fouillee performed syntheses of the unconscious "at a time when Freud was not interested in the concept".

Psychology

Nineteenth century

According to historian of psychology Mark Altschule, "It is difficult—or perhaps impossible—to find a nineteenth-century psychologist or psychiatrist who did not recognize unconscious cerebration as not only real but of the highest importance." In 1890, when psychoanalysis was still unheard of, William James, in his monumental treatise on psychology (The Principles of Psychology), examined the way Schopenhauer, von Hartmann, Janet, Binet and others had used the term 'unconscious' and 'subconscious.'" German psychologists, Gustav Fechner and Wilhelm Wundt, had begun to use the term in their experimental psychology, in the context of manifold, jumbled sense data that the mind organizes at an unconscious level before revealing it as a cogent totality in conscious form." Eduard von Hartmann published a book dedicated to the topic, Philosophy of the Unconscious, in 1869.

Freud

The iceberg metaphor proposed by G. T. Fechner is often used to provide a visual representation of Freud's theory that most of the human mind operates unconsciously.

Sigmund Freud and his followers developed an account of the unconscious mind. He worked with the unconscious mind to develop an explanation for mental illness.

For Freud, the unconscious is not merely that which is not conscious. He refers to that as the descriptive unconscious and it is only the starting postulate for real investigation into the psyche. He further distinguishes the unconscious from the pre-conscious: the pre-conscious is merely latent – thoughts, memories, etc. that are not present to consciousness but are capable of becoming so; the unconscious consists of psychic material that is made completely inaccessible to consciousness by the act of repression. The distinctions and inter-relationships between these three regions of the psyche—the conscious, the pre-conscious, and the unconscious—form what Freud calls the topographical model of the psyche. He later sought to respond to the perceived ambiguity of the term "unconscious" by developing what he called the structural model of the psyche, in which unconscious processes were described in terms of the id and the superego in their relation to the ego.

In the psychoanalytic view, unconscious mental processes can only be recognized through analysis of their effects in consciousness. Unconscious thoughts are not directly accessible to ordinary introspection, but they are capable of partially evading the censorship mechanism of repression in a disguised form, manifesting, for example, as dream elements or neurotic symptoms. Such symptoms are supposed to be capable of being "interpreted" during psychoanalysis, with the help of methods such as free association, dream analysis, and analysis of verbal slips and other unintentional manifestations in conscious life.

Jung

Carl Gustav Jung agreed with Freud that the unconscious is a determinant of personality, but he proposed that the unconscious be divided into two layers: the personal unconscious and the collective unconscious. The personal unconscious is a reservoir of material that was once conscious but has been forgotten or suppressed, much like Freud's notion. The collective unconscious, however, is the deepest level of the psyche, containing the accumulation of inherited psychic structures and archetypal experiences. Archetypes are not memories but energy centers or psychological functions that are apparent in the culture's use of symbols. The collective unconscious is therefore said to be inherited and contain material of an entire species rather than of an individual. The collective unconscious is, according to Jung, "[the] whole spiritual heritage of mankind's evolution, born anew in the brain structure of every individual".

In addition to the structure of the unconscious, Jung differed from Freud in that he did not believe that sexuality was at the base of all unconscious thoughts.

Dreams

Freud

The purpose of dreams, according to Freud, is to fulfill repressed wishes while simultaneously allowing the dreamer to remain asleep. The dream is a disguised fulfillment of the wish because the unconscious desire in its raw form would disturb the sleeper and can only avoid censorship by associating itself with elements that are not subject to repression. Thus Freud distinguished between the manifest content and latent content of the dream. The manifest content consists of the plot and elements of a dream as they appear to consciousness, particularly upon waking, as the dream is recalled. The latent content refers to the hidden or disguised meaning of the events and elements of the dream. It represents the unconscious psychic realities of the dreamer's current issues and childhood conflicts, the nature of which the analyst is seeking to understand through interpretation of the manifest content.

In Freud's theory, dreams are instigated by the events and thoughts of everyday life. In what he called the "dream-work", these events and thoughts, governed by the rules of language and the reality principle, become subject to the "primary process" of unconscious thought, which is governed by the pleasure principle, wish gratification and the repressed sexual scenarios of childhood. The dream-work involves a process of disguising these unconscious desires in order to preserve sleep. This process occurs primarily by means of what Freud called condensation and displacement. Condensation is the focusing of the energy of several ideas into one, and displacement is the surrender of one idea's energy to another more trivial representative. The manifest content is thus thought to be a highly significant simplification of the latent content, capable of being deciphered in the analytic process, potentially allowing conscious insight into unconscious mental activity.

Neurobiological theory of dreams

Allan Hobson and colleagues developed what they called the activation-synthesis hypothesis which proposes that dreams are simply the side effects of the neural activity in the brain that produces beta brain waves during REM sleep that are associated with wakefulness. According to this hypothesis, neurons fire periodically during sleep in the lower brain levels and thus send random signals to the cortex. The cortex then synthesizes a dream in reaction to these signals in order to try to make sense of why the brain is sending them. However, the hypothesis does not state that dreams are meaningless, it just downplays the role that emotional factors play in determining dreams.

Contemporary cognitive psychology

Research

There is an extensive body of research in contemporary cognitive psychology devoted to mental activity that is not mediated by conscious awareness. Most of this research on unconscious processes has been done in the academic tradition of the information processing paradigm. The cognitive tradition of research into unconscious processes does not rely on the clinical observations and theoretical bases of the psychoanalytic tradition; instead it is mostly data driven. Cognitive research reveals that individuals automatically register and acquire more information than they are consciously aware of or can consciously remember and report.

Much research has focused on the differences between conscious and unconscious perception. There is evidence that whether something is consciously perceived depends both on the incoming stimulus (bottom up strength) and on top-down mechanisms like attention. Recent research indicates that some unconsciously perceived information can become consciously accessible if there is cumulative evidence. Similarly, content that would normally be conscious can become unconscious through inattention (e.g. in the attentional blink) or through distracting stimuli like visual masking.

Unconscious processing of information about frequency

An extensive line of research conducted by Hasher and Zacks has demonstrated that individuals register information about the frequency of events automatically (outside conscious awareness and without engaging conscious information processing resources). Moreover, perceivers do this unintentionally, truly "automatically", regardless of the instructions they receive, and regardless of the information processing goals they have. The ability to unconsciously and relatively accurately tally the frequency of events appears to have little or no relation to the individual's age, education, intelligence, or personality. Thus it may represent one of the fundamental building blocks of human orientation in the environment and possibly the acquisition of procedural knowledge and experience, in general.

Criticism of the Freudian concept

The notion that the unconscious mind exists at all has been disputed.

Franz Brentano rejected the concept of the unconscious in his 1874 book Psychology from an Empirical Standpoint, although his rejection followed largely from his definitions of consciousness and unconsciousness.

Jean-Paul Sartre offers a critique of Freud's theory of the unconscious in Being and Nothingness, based on the claim that consciousness is essentially self-conscious. Sartre also argues that Freud's theory of repression is internally flawed. Philosopher Thomas Baldwin argues that Sartre's argument is based on a misunderstanding of Freud.

Erich Fromm contends that "The term 'the unconscious' is actually a mystification (even though one might use it for reasons of convenience, as I am guilty of doing in these pages). There is no such thing as the unconscious; there are only experiences of which we are aware, and others of which we are not aware, that is, of which we are unconscious. If I hate a man because I am afraid of him, and if I am aware of my hate but not of my fear, we may say that my hate is conscious and that my fear is unconscious; still my fear does not lie in that mysterious place: 'the' unconscious."

John Searle has offered a critique of the Freudian unconscious. He argues that the Freudian cases of shallow, consciously held mental states would be best characterized as 'repressed consciousness,' while the idea of more deeply unconscious mental states is more problematic. He contends that the very notion of a collection of "thoughts" that exist in a privileged region of the mind such that they are in principle never accessible to conscious awareness, is incoherent. This is not to imply that there are not "nonconscious" processes that form the basis of much of conscious life. Rather, Searle simply claims that to posit the existence of something that is like a "thought" in every way except for the fact that no one can ever be aware of it (can never, indeed, "think" it) is an incoherent concept. To speak of "something" as a "thought" either implies that it is being thought by a thinker or that it could be thought by a thinker. Processes that are not causally related to the phenomenon called thinking are more appropriately called the nonconscious processes of the brain.

Other critics of the Freudian unconscious include David StannardRichard WebsterEthan WattersRichard Ofshe, and Eric Thomas Weber.

Some scientific researchers proposed the existence of unconscious mechanisms that are very different from the Freudian ones. They speak of a "cognitive unconscious" (John Kihlstrom), an "adaptive unconscious" (Timothy Wilson), or a "dumb unconscious" (Loftus and Klinger), which executes automatic processes but lacks the complex mechanisms of repression and symbolic return of the repressed, and the "deep unconscious system" of Robert Langs.

In modern cognitive psychology, many researchers have sought to strip the notion of the unconscious from its Freudian heritage, and alternative terms such as "implicit" or "automatic" have been used. These traditions emphasize the degree to which cognitive processing happens outside the scope of cognitive awareness, and show that things we are unaware of can nonetheless influence other cognitive processes as well as behavior. Active research traditions related to the unconscious include implicit memory (for example, priming), and Pawel Lewicki's nonconscious acquisition of knowledge.

Lucid dream

From Wikipedia, the free encyclopedia

In the psychology subfield of oneirology, a lucid dream is a type of dream wherein the dreamer realizes that they are dreaming during their dream. The capacity to have and sustain lucid dreams is a trainable cognitive skill. During a lucid dream, the dreamer may gain some amount of volitional control over the dream characters, narrative, or environment, although this control of dream content is not the salient feature of lucid dreaming. An important distinction is that lucid dreaming is a distinct type of dream from other types of dreams such as prelucid dreams and vivid dreams, although prelucid dreams are a precursor to lucid dreams, and lucid dreams are often accompanied with enhanced dream vividness. Lucid dreams are also a distinct state from other lucid boundary sleep states such as lucid hypnagogia or lucid hypnopompia.

In formal psychology, lucid dreaming has been studied and reported for many years. Prominent figures from ancient to modern times have been fascinated by lucid dreams and have sought ways to better understand their causes and purpose. Many different theories have emerged as a result of scientific research on the subject. Further developments in psychological research have pointed to ways in which this form of dreaming may be utilized as a therapeutic technique.

The term lucid dream was coined by Dutch author and psychiatrist Frederik van Eeden in his 1913 article A Study of Dreams, though descriptions of dreamers being aware that they are dreaming predate the article. Psychologist Stephen LaBerge is widely considered the progenitor and leading pioneer of modern lucid dreaming research. He is the founder of the Lucidity Institute at Stanford University.

Definition

Paul Tholey laid the epistemological basis for the research of lucid dreams, proposing seven different conditions of clarity that a dream must fulfill to be defined as a lucid dream:

  1. Awareness of the dream state (orientation)
  2. Awareness of the capacity to make decisions
  3. Awareness of memory functions
  4. Awareness of self
  5. Awareness of the dream environment
  6. Awareness of the meaning of the dream
  7. Awareness of concentration and focus (the subjective clarity of that state)

Later, in 1992, a study by Deirdre Barrett examined whether lucid dreams contained four "corollaries" of lucidity:

  1. The dreamer is aware that they are dreaming
  2. They are aware that actions will not carry over after waking
  3. Physical laws need not apply in the dream
  4. The dreamer has a clear memory of the waking world

Barrett found that less than a quarter of lucidity accounts exhibited all four.

Subsequently, Stephen LaBerge studied the prevalence among lucid dreams of the ability to control the dream scenario, and found that while dream control and dream awareness are correlated, neither requires the other. LaBerge found dreams that exhibit one clearly without the capacity for the other. He also found dreams where, although the dreamer is lucid and aware they could exercise control, they choose simply to observe.

History

Eastern

The practice of lucid dreaming is central to both the ancient Indian Hindu practice of Yoga nidra and the Tibetan Buddhist practice of dream Yoga. The cultivation of such awareness was a common practice among early Buddhists.

Western

Early references to the phenomenon are also found in ancient Greek writing. For example, the philosopher Aristotle wrote: "often when one is asleep, there is something in consciousness which declares that what then presents itself is but a dream." Meanwhile, the physician Galen of Pergamon used lucid dreams as a form of therapy. In addition, a letter written by Saint Augustine of Hippo in AD 415 tells the story of a dreamer, Doctor Gennadius, and refers to lucid dreaming.

Philosopher and physician Sir Thomas Browne (1605–1682) was fascinated by dreams and described his own ability to lucid dream in his Religio Medici, stating: "...yet in one dream I can compose a whole Comedy, behold the action, apprehend the jests and laugh my self awake at the conceits thereof."

Samuel Pepys, in his diary entry for 15 August 1665, records a dream, stating: "I had my Lady Castlemayne in my arms and was admitted to use all the dalliance I desired with her, and then dreamt that this could not be awake, but that it was only a dream."

In 1867, the French sinologist Marie-Jean-Léon, Marquis d'Hervey de Saint Denys anonymously published Les Rêves et Les Moyens de Les Diriger; Observations Pratiques ("Dreams and the ways to direct them; practical observations"), in which he describes his own experiences of lucid dreaming, and proposes that it is possible for anyone to learn to dream consciously.

Frederik van Eeden (left) and Marquis d'Hervey de Saint Denys (right), early researchers of lucid dreaming

In 1913, Dutch psychiatrist and writer Frederik (Willem) van Eeden (1860–1932) coined the term "lucid dream" in an article entitled "A Study of Dreams".

Some have suggested that the term is a misnomer because Van Eeden was referring to a phenomenon more specific than a lucid dream. Van Eeden intended the term lucid to denote "having insight", as in the phrase a lucid interval applied to someone in temporary remission from a psychosis, rather than as a reference to the perceptual quality of the experience, which may or may not be clear and vivid.

Skill mastery

Clinical psychologist, Kristen LaMarca outlined four stages towards mastering the skill of using lucid dreaming:

Lucid Dreaming Skill Levels
Stage Title Description Rarity
1
Beginner
The practitioner may have no recollection of ever having a lucid dream, and perhaps has at most experienced only brief moments of lucidity.
Common
2
Experienced
An experienced lucid dreaming practitioner wields an increased ability of dream control and capacity to execute pre-intended actions. However, there are still aspects of lucid dream practice about employing lucidity productively that are yet to be honed. One's understanding of accessing and maintaining dream lucidity deepen as one has more lucid dreams.
Less common
3
Proficient
A proficient lucid dreamer is marked by a deliberate ability to accomplish intended actions in lucid dreams, along with knowledge of the best actions for given dream scenarios. The proficient lucid dream practitioner's practice is well-planned, drawing upon a broad skill set facilitating flexible oneironautic exploration, which can include contemplative practices or athletic motor skill training. This level of skill adequacy is not necessary to develop a fulfilling lucid dream practice.
Uncommon
4
Expert
Expertise in lucid dream skill is accompanied by normalization of greater intensity of lucidity during lucid dreams. LaMarca writes that the expert's practice is "characterized by at least a decade of intense dedication, long training hours, and mentorship by other more advanced experts." Spiritual figures, such Tibetan Buddhist masters, tend to display the highest order of mastery.
Extremely rare

Progression along the skill levels is akin to a maturity in the development of the practitioner's discipline, methodology and application.

Cognitive science

In 1968, Celia Green analyzed the main characteristics of such dreams, reviewing previously published literature on the subject and incorporating new data from participants of her own. She concluded that lucid dreams were a category of experience quite distinct from ordinary dreams and said they were associated with rapid eye movement sleep (REM sleep). Green was also the first to link lucid dreams to the phenomenon of false awakenings, which has since been corroborated by more recent studies.

In 1973, the National Institute of Mental Health reported that researchers at the University of California, San Francisco, were able to train sleeping subjects to recognize they were in REM dreaming and indicate this by pressing micro switches on their thumbs. Using tones and mild shocks as cues, the experiments showed that the subjects were able to signal knowledge of their various sleep stages, including dreaming.

In 1975, Dr. Keith Hearne had the idea to exploit the nature of rapid eye movements (REM) to allow a dreamer to send a message directly from dreams to the waking world. Working with an experienced lucid dreamer (Alan Worsley), he eventually succeeded in recording (via the use of an electrooculogram or EOG) a pre-defined set of eye movements signaled from within Worsley's lucid dream. This occurred at around 8 am on the morning of April 12, 1975. Hearne's EOG experiment was formally recognized through publication in the journal for The Society for Psychical Research. Lucid dreaming was subsequently researched by asking dreamers to perform pre-determined physical responses while experiencing a dream, including eye movement signals.

In 1980, Stephen LaBerge at Stanford University developed such techniques as part of his doctoral dissertation. In 1985, LaBerge performed a pilot study that showed that time perception while counting during a lucid dream is about the same as during waking life. Lucid dreamers counted out ten seconds while dreaming, signaling the start and the end of the count with a pre-arranged eye signal measured with electrooculogram recording. LaBerge's results were confirmed by German researchers D. Erlacher and M. Schredl in 2004.

In a further study by Stephen LaBerge, four subjects were compared, either singing or counting while dreaming. LaBerge found that the right hemisphere was more active during singing and the left hemisphere was more active during counting.

Neuroscientist J. Allan Hobson has hypothesized what might be occurring in the brain while lucid. The first step to lucid dreaming is recognizing that one is dreaming. This recognition might occur in the dorsolateral prefrontal cortex, which is one of the few areas deactivated during REM sleep and where working memory occurs. Once this area is activated and the recognition of dreaming occurs, the dreamer must be cautious to let the dream continue, but be conscious enough to remember that it is a dream. While maintaining this balance, the amygdala and parahippocampal cortex might be less intensely activated. To continue the intensity of the dream hallucinations, it is expected the pons and the parieto-occipital junction stay active.

Using electroencephalography (EEG) and other polysomnographical measurements, LaBerge and others have shown that lucid dreams begin in the rapid eye movement (REM) stage of sleep. LaBerge also proposes that there are higher amounts of beta-1 frequency band (13–19 Hz) brain wave activity experienced by lucid dreamers, hence there is an increased amount of activity in the parietal lobes making lucid dreaming a conscious process.

Paul Tholey, a German Gestalt psychologist and a professor of psychology and sports science, originally studied dreams in order to resolve the question of whether one dreams in colour or black and white. In his phenomenological research, he outlined an epistemological frame using critical realism. Tholey instructed his subjects to continuously suspect waking life to be a dream, in order that such a habit would manifest itself during dreams. He called this technique for inducing lucid dreams the Reflexionstechnik (reflection technique). Subjects learned to have such lucid dreams; they observed their dream content and reported it soon after awakening. Tholey could examine the cognitive abilities of dream figures. Nine trained lucid dreamers were directed to set other dream figures arithmetic and verbal tasks during lucid dreaming. Dream figures who agreed to perform the tasks proved more successful in verbal than in arithmetic tasks. Tholey discussed his scientific results with Stephen LaBerge, who has a similar approach.

A study was conducted by Stephen LaBerge and other scientists to see if it were possible to attain the ability to lucid dream through a drug. In 2018, galantamine was given to 121 patients in a double-blind, placebo-controlled trial, the only one of its kind. Some participants found as much as a 42 percent increase in their ability to lucid dream, compared to self-reports from the past six months, and ten people experienced a lucid dream for the first time. It is theorized that galantamine allows acetylcholine to build up, leading to greater recollection and awareness during dreaming.

Two-way communication

Graphical abstract of "Real-time dialogue between experimenters and dreamers during REM sleep"

Teams of cognitive scientists have established real-time two-way communication with people undergoing a lucid dream. During dreaming, they were able to consciously communicate with experimenters via eye movements or facial muscle signals, were able to comprehend complex questions and use working memory. Such interactive lucid dreaming could be a new approach for the scientific exploration of the dream state and could have applications for learning and creativity.

Alternative theories

Other researchers suggest that lucid dreaming is not a state of sleep, but of brief wakefulness, or "micro-awakening". Experiments by Stephen LaBerge used "perception of the outside world" as a criterion for wakefulness while studying lucid dreamers, and their sleep state was corroborated with physiological measurements. LaBerge's subjects experienced their lucid dream while in a state of REM, which critics felt may mean that the subjects are fully awake. J. Allen Hobson responded that lucid dreaming must be a state of both waking and dreaming.

Philosopher Norman Malcolm was a proponent of dream skepticism. He has argued against the possibility of checking the accuracy of dream reports, pointing out that "the only criterion of the truth of a statement that someone has had a certain dream is, essentially, his saying so." Yet dream reports are not the only evidence that some inner drama is being played out during REM sleep. Electromyography on speech and body muscles has demonstrated the sleeping body covertly walking, gesturing and talking while in REM.

Prevalence and frequency

In 2016, a meta-analytic study by David Saunders and colleagues on 34 lucid dreaming studies, taken from a period of 50 years, demonstrated that 55% of a pooled sample of 24,282 people claimed to have experienced lucid dreams at least once or more in their lifetime. Furthermore, for those that stated they did experience lucid dreams, approximately 23% reported to experience them on a regular basis, as often as once a month or more. In a 2004 study on lucid dream frequency and personality, a moderate correlation between nightmare frequency and frequency of lucid dreaming was demonstrated. Some lucid dreamers also reported that nightmares are a trigger for dream lucidity. Previous studies have reported that lucid dreaming is more common among adolescents than adults.

A 2015 study by Julian Mutz and Amir-Homayoun Javadi showed that people who had practiced meditation for a long time tended to have more lucid dreams. The authors claimed that "Lucid dreaming is a hybrid state of consciousness with features of both waking and dreaming" in a review they published in Neuroscience of Consciousness in 2017.

Mutz and Javadi found that during lucid dreaming, there is an increase in activity of the dorsolateral prefrontal cortex, the bilateral frontopolar prefrontal cortex, the precuneus, the inferior parietal lobules, and the supramarginal gyrus. All are brain functions related to higher cognitive functions, including working memory, planning, and self-consciousness. The researchers also found that during a lucid dream, "levels of self-determination" were similar to those that people experienced during states of wakefulness. They also found that lucid dreamers can only control limited aspects of their dream at once.

Mutz and Javadi also have stated that by studying lucid dreaming further, scientists could learn more about various types of consciousness, which happen to be less easy to separate and research at other times.

Suggested applications

Treating nightmares

It has been suggested that those who suffer from nightmares could benefit from the ability to be aware they are indeed dreaming. A pilot study performed in 2006 showed that lucid dreaming therapy treatment was successful in reducing nightmare frequency. This treatment consisted of exposure to the idea, mastery of the technique, and lucidity exercises. It was not clear what aspects of the treatment were responsible for the success of overcoming nightmares, though the treatment as a whole was said to be successful.

Australian psychologist Milan Colic has explored the application of principles from narrative therapy to clients' lucid dreams, to reduce the impact not only of nightmares during sleep but also depression, self-mutilation, and other problems in waking life. Colic found that therapeutic conversations could reduce the distressing content of dreams, while understandings about life—and even characters—from lucid dreams could be applied to their lives with marked therapeutic benefits.

Psychotherapists have applied lucid dreaming as a part of therapy. Studies have shown that, by inducing a lucid dream, recurrent nightmares can be alleviated. It is unclear whether this alleviation is due to lucidity or the ability to alter the dream itself. A 2006 study performed by Victor Spoormaker and Van den Bout evaluated the validity of lucid dreaming treatment (LDT) in chronic nightmare sufferers. LDT is composed of exposure, mastery and lucidity exercises. Results of lucid dreaming treatment revealed that the nightmare frequency of the treatment groups had decreased. In another study, Spoormaker, Van den Bout, and Meijer (2003) investigated lucid dreaming treatment for nightmares by testing eight subjects who received a one-hour individual session, which consisted of lucid dreaming exercises. The results of the study revealed that the nightmare frequency had decreased and the sleep quality had slightly increased.

Holzinger, Klösch, and Saletu managed a psychotherapy study under the working name of 'Cognition during dreaming—a therapeutic intervention in nightmares', which included 40 subjects, men and women, 18–50 years old, whose life quality was significantly altered by nightmares. The test subjects were administered Gestalt group therapy, and 24 of them were also taught to enter the state of lucid dreaming by Holzinger. This was purposefully taught in order to change the course of their nightmares. The subjects then reported the diminishment of their nightmare prevalence from 2–3 times a week to 2–3 times per month.

Creativity

In her book The Committee of Sleep, Deirdre Barrett describes how some experienced lucid dreamers have learned to remember specific practical goals such as artists looking for inspiration seeking a show of their own work once they become lucid or computer programmers looking for a screen with their desired code. However, most of these dreamers had many experiences of failing to recall waking objectives before gaining this level of control.

Exploring the World of Lucid Dreaming by Stephen LaBerge and Howard Rheingold (1990) discusses creativity within dreams and lucid dreams, including testimonials from a number of people who claim they have used the practice of lucid dreaming to help them solve a number of creative issues, from an aspiring parent thinking of potential baby names to a surgeon practicing surgical techniques. The authors discuss how creativity in dreams could stem from "conscious access to the contents of our unconscious minds"; access to "tacit knowledge"—the things we know but can't explain, or things we know but are unaware that we know.

The Dreams Behind the Music book by Craig Webb (2016) details lucid dreams of a number of musical artists, including how they are able not just to hear, but also compose, mix, arrange, practice, and perform music while conscious within their dreams.

Risks

Though lucid dreaming can be beneficial to a number of aspects of life, some risks have been suggested. Those struggling with certain mental illnesses could find it hard to tell the difference between reality and the lucid dream (psychosis).

A very small percentage of people may experience sleep paralysis, which can sometimes be confused with lucid dreaming. Although from the outside, both seem to be quite similar, there are a few distinct differences that can help differentiate them. A person usually experiences sleep paralysis when they partially wake up in REM atonia, a state in which said person is partially paralyzed and cannot move their limbs. When in sleep paralysis, people may also experience hallucinations. Although said hallucinations cannot cause physical damage, they may still be frightening. There are three common types of hallucinations: an intruder in the same room, a crushing feeling on one's chest or back, and a feeling of flying or levitating. About 7.6% of the general population have experienced sleep paralysis at least once. Exiting sleep paralysis to a waking state can be achieved by intently focusing on a part of the body, such as a finger, and wiggling it, then continuing the action of moving to the hand, the arm, and so on, until the person is fully awake.

Long-term risks with lucid dreaming have not been extensively studied, although many people have reported lucid dreaming for many years without any adverse effects. In 2018, researchers at the Wisconsin Institute for Sleep and Consciousness conducted a study that concluded individuals who lucid dream more frequently have a more active and well-connected prefrontal cortex.

Self-reference

From Wikipedia, the free encyclopedia
The ancient symbol Ouroboros, a dragon that continually consumes itself, denotes self-reference.

Self-reference is a concept that involves referring to oneself or one's own attributes, characteristics, or actions. It can occur in language, logic, mathematics, philosophy, and other fields.

In natural or formal languages, self-reference occurs when a sentence, idea or formula refers to itself. The reference may be expressed either directly—through some intermediate sentence or formula—or by means of some encoding.

In philosophy, self-reference also refers to the ability of a subject to speak of or refer to itself, that is, to have the kind of thought expressed by the first person nominative singular pronoun "I" in English.

Self-reference is studied and has applications in mathematics, philosophy, computer programming, second-order cybernetics, and linguistics, as well as in humor. Self-referential statements are sometimes paradoxical, and can also be considered recursive.

In logic, mathematics and computing

In classical philosophy, paradoxes were created by self-referential concepts such as the omnipotence paradox of asking if it was possible for a being to exist so powerful that it could create a stone that it could not lift. The Epimenides paradox, 'All Cretans are liars' when uttered by an ancient Greek Cretan was one of the first recorded versions. Contemporary philosophy sometimes employs the same technique to demonstrate that a supposed concept is meaningless or ill-defined.

In mathematics and computability theory, self-reference (also known as impredicativity) is the key concept in proving limitations of many systems. Gödel's theorem uses it to show that no formal consistent system of mathematics can ever contain all possible mathematical truths, because it cannot prove some truths about its own structure. The halting problem equivalent, in computation theory, shows that there is always some task that a computer cannot perform, namely reasoning about itself. These proofs relate to a long tradition of mathematical paradoxes such as Russell's paradox and Berry's paradox, and ultimately to classical philosophical paradoxes.

In game theory, undefined behaviors can occur where two players must model each other's mental states and behaviors, leading to infinite regress.

In computer programming, self-reference occurs in reflection, where a program can read or modify its own instructions like any other data. Numerous programming languages support reflection to some extent with varying degrees of expressiveness. Additionally, self-reference is seen in recursion (related to the mathematical recurrence relation) in functional programming, where a code structure refers back to itself during computation. 'Taming' self-reference from potentially paradoxical concepts into well-behaved recursions has been one of the great successes of computer science, and is now used routinely in, for example, writing compilers using the 'meta-language' ML. Using a compiler to compile itself is known as bootstrapping. Self-modifying code is possible to write (programs which operate on themselves), both with assembler and with functional languages such as Lisp, but is generally discouraged in real-world programming. Computing hardware makes fundamental use of self-reference in flip-flops, the basic units of digital memory, which convert potentially paradoxical logical self-relations into memory by expanding their terms over time. Thinking in terms of self-reference is a pervasive part of programmer culture, with many programs and acronyms named self-referentially as a form of humor, such as GNU ('GNU's not Unix') and PINE ('Pine is not Elm'). The GNU Hurd is named for a pair of mutually self-referential acronyms.

Tupper's self-referential formula is a mathematical curiosity which plots an image of its own formula.

In art

Drawloom, with drawboy above to control the harnesses, woven as a repeating pattern in an early-1800s piece of Japanese silk. The silk illustrates the means by which it was produced.
graffiti art on a wall stating "SORRY ABOUT YOUR WALL"
A self-referencing work of graffiti apologizing for its own existence
Self-referential graffiti. The painter drawn on a wall erases his own graffiti, and may be erased himself by the next facade cleaner.

Self-reference occurs in literature and film when an author refers to his or her own work in the context of the work itself. Examples include Miguel de Cervantes' Don Quixote, Shakespeare's A Midsummer Night's Dream, The Tempest and Twelfth Night, Denis Diderot's Jacques le fataliste et son maître, Italo Calvino's If on a winter's night a traveler, many stories by Nikolai Gogol, Lost in the Funhouse by John Barth, Luigi Pirandello's Six Characters in Search of an Author, Federico Fellini's and Bryan Forbes's The L-Shaped Room. Speculative fiction writer Samuel R. Delany makes use of this in his novels Nova and Dhalgren. In the former, Katin (a space-faring novelist) is wary of a long-standing curse wherein a novelist dies before completing any given work. Nova ends mid-sentence, thus lending credence to the curse and the realization that the novelist is the author of the story; likewise, throughout Dhalgren, Delany has a protagonist simply named The Kid (or Kidd, in some sections), whose life and work are mirror images of themselves and of the novel itself. In the sci-fi spoof film Spaceballs, Director Mel Brooks includes a scene wherein the evil characters are viewing a VHS copy of their own story, which shows them watching themselves "watching themselves", ad infinitum. Perhaps the earliest example is in Homer's Iliad, where Helen of Troy laments: "for generations still unborn/we will live in song" (appearing in the song itself).

Self-reference in art is closely related to the concepts of breaking the fourth wall and meta-reference, which often involve self-reference. The short stories of Jorge Luis Borges play with self-reference and related paradoxes in many ways. Samuel Beckett's Krapp's Last Tape consists entirely of the protagonist listening to and making recordings of himself, mostly about other recordings. During the 1990s and 2000s filmic self-reference was a popular part of the rubber reality movement, notably in Charlie Kaufman's films Being John Malkovich and Adaptation, the latter pushing the concept arguably to its breaking point as it attempts to portray its own creation, in a dramatized version of the Droste effect.

Various creation myths invoke self-reference to solve the problem of what created the creator. For example, the Egyptian creation myth has a god swallowing his own semen to create himself. The Ouroboros is a mythical dragon which eats itself.

The Quran includes numerous instances of self-referentiality.

The surrealist painter René Magritte is famous for his self-referential works. His painting The Treachery of Images, includes the words "this is not a pipe", the truth of which depends entirely on whether the word ceci (in English, "this") refers to the pipe depicted—or to the painting or the word or sentence itself. M.C. Escher's art also contains many self-referential concepts such as hands drawing themselves.

In language

A word that describes itself is called an autological word (or autonym). This generally applies to adjectives, for example sesquipedalian (i.e. "sesquipedalian" is a sesquipedalian word), but can also apply to other parts of speech, such as TLA, as a three-letter abbreviation for "three-letter abbreviation".

A sentence which inventories its own letters and punctuation marks is called an autogram.

There is a special case of meta-sentence in which the content of the sentence in the metalanguage and the content of the sentence in the object language are the same. Such a sentence is referring to itself. However some meta-sentences of this type can lead to paradoxes. "This is a sentence." can be considered to be a self-referential meta-sentence which is obviously true. However "This sentence is false" is a meta-sentence which leads to a self-referential paradox. Such sentences can lead to problems, for example, in law, where statements bringing laws into existence can contradict one another or themselves. Kurt Gödel claimed to have found such a loophole in the United States Constitution at his citizenship ceremony.

Self-reference occasionally occurs in the media when it is required to write about itself, for example the BBC reporting on job cuts at the BBC. Notable encyclopedias may be required to feature articles about themselves, such as Wikipedia's article on Wikipedia.

Fumblerules are a list of rules of good grammar and writing, demonstrated through sentences that violate those very rules, such as "Avoid cliches like the plague" and "Don't use no double negatives". The term was coined in a published list of such rules by William Safire.

Circular definition is a type of self-reference in which the definition of a term or concept includes the term or concept itself, either explicitly or implicitly. Circular definitions are considered fallacious because they only define a term in terms of itself. This type of self-reference may be useful in argumentation, but can result in a lack of clarity in communication.

The adverb "hereby" is used in a self-referential way, for example in the statement "I hereby declare you husband and wife."

In law

Several constitutions contain self-referential clauses defining how the constitution itself may be amended. An example is Article Five of the United States Constitution.

Inverse problem

From Wikipedia, the free encyclopedia

An inverse problem in science is the process of calculating from a set of observations the causal factors that produced them: for example, calculating an image in X-ray computed tomography, source reconstruction in acoustics, or calculating the density of the Earth from measurements of its gravity field. It is called an inverse problem because it starts with the effects and then calculates the causes. It is the inverse of a forward problem, which starts with the causes and then calculates the effects.

Inverse problems are some of the most important mathematical problems in science and mathematics because they tell us about parameters that we cannot directly observe. They can be found in system identification, optics, radar, acoustics, communication theory, signal processing, medical imaging, computer visiongeophysics, oceanography, meteorology, astronomy, remote sensing, natural language processing, machine learningnondestructive testing, slope stability analysis and many other fields.

History

Starting with the effects to discover the causes has concerned physicists for centuries. A historical example is the calculations of Adams and Le Verrier which led to the discovery of Neptune from the perturbed trajectory of Uranus. However, a formal study of inverse problems was not initiated until the 20th century.

One of the earliest examples of a solution to an inverse problem was discovered by Hermann Weyl and published in 1911, describing the asymptotic behavior of eigenvalues of the Laplace–Beltrami operator. Today known as Weyl's law, it is perhaps most easily understood as an answer to the question of whether it is possible to hear the shape of a drum. Weyl conjectured that the eigenfrequencies of a drum would be related to the area and perimeter of the drum by a particular equation, a result improved upon by later mathematicians.

The field of inverse problems was later touched on by Soviet-Armenian physicist, Viktor Ambartsumian.

While still a student, Ambartsumian thoroughly studied the theory of atomic structure, the formation of energy levels, and the Schrödinger equation and its properties, and when he mastered the theory of eigenvalues of differential equations, he pointed out the apparent analogy between discrete energy levels and the eigenvalues of differential equations. He then asked: given a family of eigenvalues, is it possible to find the form of the equations whose eigenvalues they are? Essentially Ambartsumian was examining the inverse Sturm–Liouville problem, which dealt with determining the equations of a vibrating string. This paper was published in 1929 in the German physics journal Zeitschrift für Physik and remained in obscurity for a rather long time. Describing this situation after many decades, Ambartsumian said, "If an astronomer publishes an article with a mathematical content in a physics journal, then the most likely thing that will happen to it is oblivion."

Nonetheless, toward the end of the Second World War, this article, written by the 20-year-old Ambartsumian, was found by Swedish mathematicians and formed the starting point for a whole area of research on inverse problems, becoming the foundation of an entire discipline.

Then important efforts have been devoted to a "direct solution" of the inverse scattering problem especially by Gelfand and Levitan in the Soviet Union. They proposed an analytic constructive method for determining the solution. When computers became available, some authors have investigated the possibility of applying their approach to similar problems such as the inverse problem in the 1D wave equation. But it rapidly turned out that the inversion is an unstable process: noise and errors can be tremendously amplified making a direct solution hardly practicable. Then, around the seventies, the least-squares and probabilistic approaches came in and turned out to be very helpful for the determination of parameters involved in various physical systems. This approach met a lot of success. Nowadays inverse problems are also investigated in fields outside physics, such as chemistry, economics, and computer science. Eventually, as numerical models become prevalent in many parts of society, we may expect an inverse problem associated with each of these numerical models.

Conceptual understanding

Since Newton, scientists have extensively attempted to model the world. In particular, when a mathematical model is available (for instance, Newton's gravitational law or Coulomb's equation for electrostatics), we can foresee, given some parameters that describe a physical system (such as a distribution of mass or a distribution of electric charges), the behavior of the system. This approach is known as mathematical modeling and the above-mentioned physical parameters are called the model parameters or simply the model. To be precise, we introduce the notion of state of the physical system: it is the solution of the mathematical model's equation. In optimal control theory, these equations are referred to as the state equations. In many situations we are not truly interested in knowing the physical state but just its effects on some objects (for instance, the effects the gravitational field has on a specific planet). Hence we have to introduce another operator, called the observation operator, which converts the state of the physical system (here the predicted gravitational field) into what we want to observe (here the movements of the considered planet). We can now introduce the so-called forward problem, which consists of two steps:

  • determination of the state of the system from the physical parameters that describe it
  • application of the observation operator to the estimated state of the system so as to predict the behavior of what we want to observe.

This leads to introduce another operator (F stands for "forward") which maps model parameters into , the data that model predicts that is the result of this two-step procedure. Operator is called forward operator or forward map. In this approach we basically attempt at predicting the effects knowing the causes.

The table below shows, the Earth being considered as the physical system and for different physical phenomena, the model parameters that describe the system, the physical quantity that describes the state of the physical system and observations commonly made on the state of the system.

Governing equations Model parameters (input of the model) State of the physical system Common observations on the system
Newton's law of gravity Distribution of mass Gravitational field Measurement made by gravimeters at different surface locations
Maxwell's equations Distribution of magnetic susceptibility Magnetic field Magnetic field measured at different surface locations by magnetometers (case of a steady state)
Wave equation Distribution of wave-speeds and densities Wave-field caused by artificial or natural seismic sources Particle velocity measured by seismometers placed at different surface locations
Diffusion equation Distribution of Diffusion coefficient Diffusing material concentration as a function of space and time Monitoring of this concentration measured at different locations

In the inverse problem approach we, roughly speaking, try to know the causes given the effects.

General statement of the inverse problem

The inverse problem is the "inverse" of the forward problem: instead of determining the data produced by particular model parameters, we want to determine the model parameters that produce the data that is the observation we have recorded (the subscript obs stands for observed). Our goal, in other words, is to determine the model parameters such that (at least approximately) where is the forward map. We denote by the (possibly infinite) number of model parameters, and by the number of recorded data. We introduce some useful concepts and the associated notations that will be used below:

  • The space of models denoted by : the vector space spanned by model parameters; it has dimensions;
  • The space of data denoted by : if we organize the measured samples in a vector with components (if our measurements consist of functions, is a vector space with infinite dimensions);
  • : the response of model ; it consists of the data predicted by model ;
  • : the image of by the forward map, it is a subset of (but not a subspace unless is linear) made of responses of all models;
  • : the data misfits (or residuals) associated with model : they can be arranged as a vector, an element of .

The concept of residuals is very important: in the scope of finding a model that matches the data, their analysis reveals if the considered model can be considered as realistic or not. Systematic unrealistic discrepancies between the data and the model responses also reveals that the forward map is inadequate and may give insights about an improved forward map.

When operator is linear, the inverse problem is linear. Otherwise, that is most often, the inverse problem is nonlinear. Also, models cannot always be described by a finite number of parameters. It is the case when we look for distributed parameters (a distribution of wave-speeds for instance): in such cases the goal of the inverse problem is to retrieve one or several functions. Such inverse problems are inverse problems with infinite dimension.

Linear inverse problems

In the case of a linear forward map and when we deal with a finite number of model parameters, the forward map can be written as a linear system where is the matrix that characterizes the forward map. The linear system can be systematically solved by means of both regularization and Bayesian methods.

An elementary example: Earth's gravitational field

Only a few physical systems are actually linear with respect to the model parameters. One such system from geophysics is that of the Earth's gravitational field. The Earth's gravitational field is determined by the density distribution of the Earth in the subsurface. Because the lithology of the Earth changes quite significantly, we are able to observe minute differences in the Earth's gravitational field on the surface of the Earth. From our understanding of gravity (Newton's Law of Gravitation), we know that the mathematical expression for gravity is: here is a measure of the local gravitational acceleration, is the universal gravitational constant, is the local mass (which is related to density) of the rock in the subsurface and is the distance from the mass to the observation point.

By discretizing the above expression, we are able to relate the discrete data observations on the surface of the Earth to the discrete model parameters (density) in the subsurface that we wish to know more about. For example, consider the case where we have measurements carried out at 5 locations on the surface of the Earth. In this case, our data vector, is a column vector of dimension (5×1): its -th component is associated with the -th observation location. We also know that we only have five unknown masses in the subsurface (unrealistic but used to demonstrate the concept) with known location: we denote by the distance between the -th observation location and the -th mass. Thus, we can construct the linear system relating the five unknown masses to the five data points as follows:

To solve for the model parameters that fit our data, we might be able to invert the matrix to directly convert the measurements into our model parameters. For example: A system with five equations and five unknowns is a very specific situation: our example was designed to end up with this specificity. In general, the numbers of data and unknowns are different so that matrix is not square.

However, even a square matrix can have no inverse: matrix can be rank deficient (i.e. has zero eigenvalues) and the solution of the system is not unique. Then the solution of the inverse problem will be undetermined. This is a first difficulty. Over-determined systems (more equations than unknowns) have other issues. Also noise may corrupt our observations making possibly outside the space of possible responses to model parameters so that solution of the system may not exist. This is another difficulty.

Tools to overcome the first difficulty

The first difficulty reflects a crucial problem: Our observations do not contain enough information and additional data are required. Additional data can come from physical prior information on the parameter values, on their spatial distribution or, more generally, on their mutual dependence. It can also come from other experiments: For instance, we may think of integrating data recorded by gravimeters and seismographs for a better estimation of densities. The integration of this additional information is basically a problem of statistics. This discipline is the one that can answer the question: How to mix quantities of different nature? We will be more precise in the section "Bayesian approach" below.

Concerning distributed parameters, prior information about their spatial distribution often consists of information about some derivatives of these distributed parameters. Also, it is common practice, although somewhat artificial, to look for the "simplest" model that reasonably matches the data. This is usually achieved by penalizing the norm of the gradient (or the total variation) of the parameters (this approach is also referred to as the maximization of the entropy). One can also make the model simple through a parametrization that introduces degrees of freedom only when necessary.

Additional information may also be integrated through inequality constraints on the model parameters or some functions of them. Such constraints are important to avoid unrealistic values for the parameters (negative values for instance). In this case, the space spanned by model parameters will no longer be a vector space but a subset of admissible models denoted by in the sequel.

Tools to overcome the second difficulty

As mentioned above, noise may be such that our measurements are not the image of any model, so that we cannot look for a model that produces the data but rather look for the best (or optimal) model: that is, the one that best matches the data. This leads us to minimize an objective function, namely a functional that quantifies how big the residuals are or how far the predicted data are from the observed data. Of course, when we have perfect data (i.e. no noise) then the recovered model should fit the observed data perfectly. A standard objective function, , is of the form: where is the Euclidean norm (it will be the norm when the measurements are functions instead of samples) of the residuals. This approach amounts to making use of ordinary least squares, an approach widely used in statistics. However, the Euclidean norm is known to be very sensitive to outliers: to avoid this difficulty we may think of using other distances, for instance the norm, in replacement of the norm.

Bayesian approach

Very similar to the least-squares approach is the probabilistic approach: If we know the statistics of the noise that contaminates the data, we can think of seeking the most likely model m, which is the model that matches the maximum likelihood criterion. If the noise is Gaussian, the maximum likelihood criterion appears as a least-squares criterion, the Euclidean scalar product in data space being replaced by a scalar product involving the co-variance of the noise. Also, should prior information on model parameters be available, we could think of using Bayesian inference to formulate the solution of the inverse problem. This approach is described in detail in Tarantola's book.

Numerical solution of our elementary example

Here we make use of the Euclidean norm to quantify the data misfits. As we deal with a linear inverse problem, the objective function is quadratic. For its minimization, it is classical to compute its gradient using the same rationale (as we would to minimize a function of only one variable). At the optimal model , this gradient vanishes which can be written as: where FT denotes the matrix transpose of F. This equation simplifies to:

This expression is known as the normal equation and gives us a possible solution to the inverse problem. In our example matrix turns out to be generally full rank so that the equation above makes sense and determines uniquely the model parameters: we do not need integrating additional information for ending up with a unique solution.

Mathematical and computational aspects

Inverse problems are typically ill-posed, as opposed to the well-posed problems usually met in mathematical modeling. Of the three conditions for a well-posed problem suggested by Jacques Hadamard (existence, uniqueness, and stability of the solution or solutions) the condition of stability is most often violated. In the sense of functional analysis, the inverse problem is represented by a mapping between metric spaces. While inverse problems are often formulated in infinite dimensional spaces, limitations to a finite number of measurements, and the practical consideration of recovering only a finite number of unknown parameters, may lead to the problems being recast in discrete form. In this case the inverse problem will typically be ill-conditioned. In these cases, regularization may be used to introduce mild assumptions on the solution and prevent overfitting. Many instances of regularized inverse problems can be interpreted as special cases of Bayesian inference.

Numerical solution of the optimization problem

Some inverse problems have a very simple solution, for instance, when one has a set of unisolvent functions, meaning a set of functions such that evaluating them at distinct points yields a set of linearly independent vectors. This means that given a linear combination of these functions, the coefficients can be computed by arranging the vectors as the columns of a matrix and then inverting this matrix. The simplest example of unisolvent functions is polynomials constructed, using the unisolvence theorem, so as to be unisolvent. Concretely, this is done by inverting the Vandermonde matrix. But this a very specific situation.

In general, the solution of an inverse problem requires sophisticated optimization algorithms. When the model is described by a large number of parameters (the number of unknowns involved in some diffraction tomography applications can reach one billion), solving the linear system associated with the normal equations can be cumbersome. The numerical method to be used for solving the optimization problem depends in particular on the cost required for computing the solution of the forward problem. Once chosen the appropriate algorithm for solving the forward problem (a straightforward matrix-vector multiplication may be not adequate when matrix is huge), the appropriate algorithm for carrying out the minimization can be found in textbooks dealing with numerical methods for the solution of linear systems and for the minimization of quadratic functions (see for instance Ciarlet or Nocedal).

Also, the user may wish to add physical constraints to the models: In this case, they have to be familiar with constrained optimization methods, a subject in itself. In all cases, computing the gradient of the objective function often is a key element for the solution of the optimization problem. As mentioned above, information about the spatial distribution of a distributed parameter can be introduced through the parametrization. One can also think of adapting this parametrization during the optimization.

Should the objective function be based on a norm other than the Euclidean norm, we have to leave the area of quadratic optimization. As a result, the optimization problem becomes more difficult. In particular, when the norm is used for quantifying the data misfit the objective function is no longer differentiable: its gradient does not make sense any longer. Dedicated methods (see for instance Lemaréchal) from non differentiable optimization come in.

Once the optimal model is computed we have to address the question: "Can we trust this model?" The question can be formulated as follows: How large is the set of models that match the data "nearly as well" as this model? In the case of quadratic objective functions, this set is contained in a hyper-ellipsoid, a subset of ( is the number of unknowns), whose size depends on what we mean with "nearly as well", that is on the noise level. The direction of the largest axis of this ellipsoid (eigenvector associated with the smallest eigenvalue of matrix ) is the direction of poorly determined components: if we follow this direction, we can bring a strong perturbation to the model without changing significantly the value of the objective function and thus end up with a significantly different quasi-optimal model. We clearly see that the answer to the question "can we trust this model" is governed by the noise level and by the eigenvalues of the Hessian of the objective function or equivalently, in the case where no regularization has been integrated, by the singular values of matrix . Of course, the use of regularization (or other kinds of prior information) reduces the size of the set of almost optimal solutions and, in turn, increases the confidence we can put in the computed solution.

Stability, regularization and model discretization in infinite dimension

We focus here on the recovery of a distributed parameter. When looking for distributed parameters we have to discretize these unknown functions. Doing so, we reduce the dimension of the problem to something finite. But now, the question is: is there any link between the solution we compute and that of the initial problem? Then another question: what do we mean with the solution of the initial problem? Since a finite number of data does not allow the determination of an infinity of unknowns, the original data misfit functional has to be regularized to ensure the uniqueness of the solution. Many times, reducing the unknowns to a finite-dimensional space will provide an adequate regularization: the computed solution will look like a discrete version of the solution we were looking for. For example, a naïve discretization will often work for solving the deconvolution problem: it will work as long as we do not allow missing frequencies to show up in the numerical solution. But many times, regularization has to be integrated explicitly in the objective function.

In order to understand what may happen, we have to keep in mind that solving such a linear inverse problem amount to solving a Fredholm integral equation of the first kind:

where is the kernel, and are vectors of , and is a domain in . This holds for a 2D application. For a 3D application, we consider . Note that here the model parameters consist of a function and that the response of a model also consists of a function denoted by . This equation is an extension to infinite dimension of the matrix equation given in the case of discrete problems.

For sufficiently smooth the operator defined above is compact on reasonable Banach spaces such as the . F. Riesz theory states that the set of singular values of such an operator contains zero (hence the existence of a null-space), is finite or at most countable, and, in the latter case, they constitute a sequence that goes to zero. In the case of a symmetric kernel, we have an infinity of eigenvalues and the associated eigenvectors constitute a hilbertian basis of . Thus any solution of this equation is determined up to an additive function in the null-space and, in the case of infinity of singular values, the solution (which involves the reciprocal of arbitrary small eigenvalues) is unstable: two ingredients that make the solution of this integral equation a typical ill-posed problem! However, we can define a solution through the pseudo-inverse of the forward map (again up to an arbitrary additive function). When the forward map is compact, the classical Tikhonov regularization will work if we use it for integrating prior information stating that the norm of the solution should be as small as possible: this will make the inverse problem well-posed. Yet, as in the finite dimension case, we have to question the confidence we can put in the computed solution. Again, basically, the information lies in the eigenvalues of the Hessian operator. Should subspaces containing eigenvectors associated with small eigenvalues be explored for computing the solution, then the solution can hardly be trusted: some of its components will be poorly determined. The smallest eigenvalue is equal to the weight introduced in Tikhonov regularization.

Irregular kernels may yield a forward map which is not compact and even unbounded if we naively equip the space of models with the norm. In such cases, the Hessian is not a bounded operator and the notion of eigenvalue does not make sense any longer. A mathematical analysis is required to make it a bounded operator and design a well-posed problem: an illustration can be found in. Again, we have to question the confidence we can put in the computed solution and we have to generalize the notion of eigenvalue to get the answer.

Analysis of the spectrum of the Hessian operator is thus a key element to determine how reliable the computed solution is. However, such an analysis is usually a very heavy task. This has led several authors to investigate alternative approaches in the case where we are not interested in all the components of the unknown function but only in sub-unknowns that are the images of the unknown function by a linear operator. These approaches are referred to as the " Backus and Gilbert method", Lions's sentinels approach, and the SOLA method: these approaches turned out to be strongly related with one another as explained in Chavent Finally, the concept of limited resolution, often invoked by physicists, is nothing but a specific view of the fact that some poorly determined components may corrupt the solution. But, generally speaking, these poorly determined components of the model are not necessarily associated with high frequencies.

Some classical linear inverse problems for the recovery of distributed parameters

The problems mentioned below correspond to different versions of the Fredholm integral: each of these is associated with a specific kernel .

Deconvolution

The goal of deconvolution is to reconstruct the original image or signal which appears as noisy and blurred on the data . From a mathematical point of view, the kernel here only depends on the difference between and .

Tomographic methods

In these methods we attempt at recovering a distributed parameter, the observation consisting in the measurement of the integrals of this parameter carried out along a family of lines. We denote by the line in this family associated with measurement point . The observation at can thus be written as: where is the arc-length along and a known weighting function. Comparing this equation with the Fredholm integral above, we notice that the kernel is kind of a delta function that peaks on line . With such a kernel, the forward map is not compact.

Computed tomography

In X-ray computed tomography the lines on which the parameter is integrated are straight lines: the tomographic reconstruction of the parameter distribution is based on the inversion of the Radon transform. Although from a theoretical point of view many linear inverse problems are well understood, problems involving the Radon transform and its generalisations still present many practical challenges, such as with of sufficiency of data. Such problems include incomplete data for the x-ray transform in three dimensions and problems involving the generalisation of the x-ray transform to tensor fields. Solutions explored include Algebraic Reconstruction Technique, filtered backprojection, and as computing power has increased, iterative reconstruction methods such as iterative Sparse Asymptotic Minimum Variance.

Diffraction tomography

Diffraction tomography is a classical linear inverse problem in exploration seismology: the amplitude recorded at one time for a given source-receiver pair is the sum of contributions arising from points such that the sum of the distances, measured in traveltimes, from the source and the receiver, respectively, is equal to the corresponding recording time. In 3D the parameter is not integrated along lines but over surfaces. Should the propagation velocity be constant, such points are distributed on an ellipsoid. The inverse problems consists in retrieving the distribution of diffracting points from the seismograms recorded along the survey, the velocity distribution being known. A direct solution has been originally proposed by Beylkin and Lambaré et al.: these works were the starting points of approaches known as amplitude preserved migration (see Beylkin and Bleistein). Should geometrical optics techniques (i.e. rays) be used for the solving the wave equation, these methods turn out to be closely related to the so-called least-squares migration methods derived from the least-squares approach (see Lailly, Tarantola).

Doppler tomography (astrophysics)

If we consider a rotating stellar object, the spectral lines we can observe on a spectral profile will be shifted due to Doppler effect. Doppler tomography aims at converting the information contained in spectral monitoring of the object into a 2D image of the emission (as a function of the radial velocity and of the phase in the periodic rotation movement) of the stellar atmosphere. As explained by Tom Marsh this linear inverse problem is tomography like: we have to recover a distributed parameter which has been integrated along lines to produce its effects in the recordings.

Inverse heat conduction

Early publications on inverse heat conduction arose from determining surface heat flux during atmospheric re-entry from buried temperature sensors. Other applications where surface heat flux is needed but surface sensors are not practical include: inside reciprocating engines, inside rocket engines; and, testing of nuclear reactor components. A variety of numerical techniques have been developed to address the ill-posedness and sensitivity to measurement error caused by damping and lagging in the temperature signal.

Non-linear inverse problems

Non-linear inverse problems constitute an inherently more difficult family of inverse problems. Here the forward map is a non-linear operator. Modeling of physical phenomena often relies on the solution of a partial differential equation (see table above except for gravity law): although these partial differential equations are often linear, the physical parameters that appear in these equations depend in a non-linear way of the state of the system and therefore on the observations we make on it.

Some classical non-linear inverse problems

Inverse scattering problems

Whereas linear inverse problems were completely solved from the theoretical point of view at the end of the nineteenth century , only one class of nonlinear inverse problems was so before 1970, that of inverse spectral and (one space dimension) inverse scattering problems, after the seminal work of the Russian mathematical school (Krein, Gelfand, Levitan, Marchenko). A large review of the results has been given by Chadan and Sabatier in their book "Inverse Problems of Quantum Scattering Theory" (two editions in English, one in Russian).

In this kind of problem, data are properties of the spectrum of a linear operator which describe the scattering. The spectrum is made of eigenvalues and eigenfunctions, forming together the "discrete spectrum", and generalizations, called the continuous spectrum. The very remarkable physical point is that scattering experiments give information only on the continuous spectrum, and that knowing its full spectrum is both necessary and sufficient in recovering the scattering operator. Hence we have invisible parameters, much more interesting than the null space which has a similar property in linear inverse problems. In addition, there are physical motions in which the spectrum of such an operator is conserved as a consequence of such motion. This phenomenon is governed by special nonlinear partial differential evolution equations, for example the Korteweg–de Vries equation. If the spectrum of the operator is reduced to one single eigenvalue, its corresponding motion is that of a single bump that propagates at constant velocity and without deformation, a solitary wave called a "soliton".

A perfect signal and its generalizations for the Korteweg–de Vries equation or other integrable nonlinear partial differential equations are of great interest, with many possible applications. This area has been studied as a branch of mathematical physics since the 1970s. Nonlinear inverse problems are also currently studied in many fields of applied science (acoustics, mechanics, quantum mechanics, electromagnetic scattering - in particular radar soundings, seismic soundings, and nearly all imaging modalities).

A final example related to the Riemann hypothesis was given by Wu and Sprung, the idea is that in the semiclassical old quantum theory the inverse of the potential inside the Hamiltonian is proportional to the half-derivative of the eigenvalues (energies) counting function n(x).

Permeability matching in oil and gas reservoirs

The goal is to recover the diffusion coefficient in the parabolic partial differential equation that models single phase fluid flows in porous media. This problem has been the object of many studies since a pioneering work carried out in the early seventies. Concerning two-phase flows an important problem is to estimate the relative permeabilities and the capillary pressures.

Inverse problems in the wave equations

The goal is to recover the wave-speeds (P and S waves) and the density distributions from seismograms. Such inverse problems are of prime interest in seismology and exploration geophysics. We can basically consider two mathematical models:

These basic hyperbolic equations can be upgraded by incorporating attenuation, anisotropy, ...

The solution of the inverse problem in the 1D wave equation has been the object of many studies. It is one of the very few non-linear inverse problems for which we can prove the uniqueness of the solution. The analysis of the stability of the solution was another challenge. Practical applications, using the least-squares approach, were developed. Extension to 2D or 3D problems and to the elastodynamics equations was attempted since the 80's but turned out to be very difficult ! This problem often referred to as Full Waveform Inversion (FWI), is not yet completely solved: among the main difficulties are the existence of non-Gaussian noise into the seismograms, cycle-skipping issues (also known as phase ambiguity), and the chaotic behavior of the data misfit function. Some authors have investigated the possibility of reformulating the inverse problem so as to make the objective function less chaotic than the data misfit function.

Travel-time tomography

Realizing how difficult is the inverse problem in the wave equation, seismologists investigated a simplified approach making use of geometrical optics. In particular they aimed at inverting for the propagation velocity distribution, knowing the arrival times of wave-fronts observed on seismograms. These wave-fronts can be associated with direct arrivals or with reflections associated with reflectors whose geometry is to be determined, jointly with the velocity distribution.

The arrival time distribution ( is a point in physical space) of a wave-front issued from a point source, satisfies the Eikonal equation: where denotes the slowness (reciprocal of the velocity) distribution. The presence of makes this equation nonlinear. It is classically solved by shooting rays (trajectories about which the arrival time is stationary) from the point source.

This problem is tomography like: the measured arrival times are the integral along the ray-path of the slowness. But this tomography like problem is nonlinear, mainly because the unknown ray-path geometry depends upon the velocity (or slowness) distribution. In spite of its nonlinear character, travel-time tomography turned out to be very effective for determining the propagation velocity in the Earth or in the subsurface, the latter aspect being a key element for seismic imaging, in particular using methods mentioned in Section "Diffraction tomography".

Mathematical aspects: Hadamard's questions

The questions concern well-posedness: Does the least-squares problem have a unique solution which depends continuously on the data (stability problem)? It is the first question, but it is also a difficult one because of the non-linearity of . In order to see where the difficulties arise from, Chavent proposed to conceptually split the minimization of the data misfit function into two consecutive steps ( is the subset of admissible models):

  • projection step: given find a projection on (nearest point on according to the distance involved in the definition of the objective function)
  • given this projection find one pre-image that is a model whose image by operator is this projection.

Difficulties can - and usually will - arise in both steps:

  1. operator is not likely to be one-to-one, therefore there can be more than one pre-image,
  2. even when is one-to-one, its inverse may not be continuous over ,
  3. the projection on may not exist, should this set be not closed,
  4. the projection on can be non-unique and not continuous as this can be non-convex due to the non-linearity of .

We refer to Chavent for a mathematical analysis of these points.

Computational aspects

A non-convex data misfit function

The forward map being nonlinear, the data misfit function is likely to be non-convex, making local minimization techniques inefficient. Several approaches have been investigated to overcome this difficulty:

  • use of global optimization techniques such as sampling of the posterior density function and Metropolis algorithm in the inverse problem probabilistic framework, genetic algorithms (alone or in combination with Metropolis algorithm: see for an application to the determination of permeabilities that match the existing permeability data), neural networks, regularization techniques including multi scale analysis;
  • reformulation of the least-squares objective function so as to make it smoother (see for the inverse problem in the wave equations.)

Computation of the gradient of the objective function

Inverse problems, especially in infinite dimension, may be large size, thus requiring important computing time. When the forward map is nonlinear, the computational difficulties increase and minimizing the objective function can be difficult. Contrary to the linear situation, an explicit use of the Hessian matrix for solving the normal equations does not make sense here: the Hessian matrix varies with models. Much more effective is the evaluation of the gradient of the objective function for some models. Important computational effort can be saved when we can avoid the very heavy computation of the Jacobian (often called "Fréchet derivatives"): the adjoint state method, proposed by Chavent and Lions, is aimed to avoid this very heavy computation. It is now very widely used.

An inversion algorithm (published Under a Creative Commons license, CC BY-NC-ND by Elsevier)

Applications

Inverse problem theory is used extensively in weather predictions, oceanography, hydrology, neuroscience, and petroleum engineering. Another application is inversion of elastic waves for non-destructive characterization of engineering structures.

Inverse problems are also found in the field of heat transfer, where a surface heat flux is estimated outgoing from temperature data measured inside a rigid body; and, in understanding the controls on plant-matter decay. The linear inverse problem is also the fundamental of spectral estimation and direction-of-arrival (DOA) estimation in signal processing.

Inverse lithography is used in photomask design for semiconductor device fabrication.

Academic journals

Four main academic journals cover inverse problems in general:

  • Inverse Problems
  • Journal of Inverse and Ill-posed Problems
  • Inverse Problems in Science and Engineering
  • Inverse Problems and Imaging

Many journals on medical imaging, geophysics, non-destructive testing, etc. are dominated by inverse problems in those areas.

Relationship between science and religion

From Wikipedia, the free encyclopedia "Science and Religion" redirects here. For the 1991 book by John Hedley Brooke, see  Science...