Search This Blog

Wednesday, July 6, 2022

Association rule learning

From Wikipedia, the free encyclopedia
 

Association rule learning is a rule-based machine learning method for discovering interesting relations between variables in large databases. It is intended to identify strong rules discovered in databases using some measures of interestingness. In any given transaction with a variety of items, association rules are meant to discover the rules that determine how or why certain items are connected.

Based on the concept of strong rules, Rakesh Agrawal, Tomasz Imieliński and Arun Swami introduced association rules for discovering regularities between products in large-scale transaction data recorded by point-of-sale (POS) systems in supermarkets. For example, the rule found in the sales data of a supermarket would indicate that if a customer buys onions and potatoes together, they are likely to also buy hamburger meat. Such information can be used as the basis for decisions about marketing activities such as, e.g., promotional pricing or product placements.

In addition to the above example from market basket analysis association rules are employed today in many application areas including Web usage mining, intrusion detection, continuous production, and bioinformatics. In contrast with sequence mining, association rule learning typically does not consider the order of items either within a transaction or across transactions.

The association rule algorithm itself consists of various parameters that can make it difficult for those without some expertise in data mining to execute, with many rules that are arduous to understand.

Despite this, association rule learning is a great system for predicting the behavior in data interconnections. This makes it a noteworthy technique for classification, or discovering patterns in data, when implementing machine learning methods.

Definition

A Venn Diagram to show the associations between itemsets X and Y of a dataset. All transactions that contain item X are located in the white, left portion of the circle, while those containing Y are colored red and on the right. Any transaction containing both X and Y are located in the middle and are colored pink. Multiple concepts can be used to depict information from this graph. For example, if one takes all of the transactions in the pink section and divided them by the total amount of transactions (transactions containing X (white) + transactions containing Y(red)), the output would be known as the support. An instance of getting the result of a method known as the confidence, one can take all of the transactions in the middle (pink) and divide them by all transactions that contain Y (red and pink). In this case, Y is the antecedent and X is the consequent.

Following the original definition by Agrawal, Imieliński, Swami the problem of association rule mining is defined as:

Let be a set of binary attributes called items.

Let be a set of transactions called the database.

Each transaction in has a unique transaction ID and contains a subset of the items in .

A rule is defined as an implication of the form:

, where .

In Agrawal, Imieliński, Swami a rule is defined only between a set and a single item, for .

Every rule is composed by two different sets of items, also known as itemsets, and , where is called antecedent or left-hand-side (LHS) and consequent or right-hand-side (RHS). The antecedent is that item that can be found in the data while the consequent is the item found when combined with the antecedent. The statement is often read as if then , where the antecedent ( ) is the if and the consequent () is the then. This simply implies that, in theory, whenever occurs in a dataset, then will as well.

Process

Association rules are made by searching data for frequent if-then patterns and by using a certain criterion under Support and Confidence to define what the most important relationships are. Support is the evidence of how frequent an item appears in the data given, as Confidence is defined by how many times the if-then statements are found true. However, there is a third criteria that can be used, it is called Lift and it can be used to compare the expected Confidence and the actual Confidence. Lift will show how many times the if-then statement is expected to be found to be true.

Association rules are made to calculate from itemsets, which are created by two or more items. If the rules were built from the analyzing from all the possible itemsets from the data then there would be so many rules that they wouldn’t have any meaning. That is why Association rules are typically made from rules that are well represented by the data.

There are many different data mining techniques you could use to find certain analytics and results, for example, there is Classification analysis, Clustering analysis, and Regression analysis. Depending on what you are looking for with your data depends on what technique you should use. Association rules are primarily used to find analytics and a prediction of customer behavior. For Classification analysis, it would most likely be used to question, make decisions, and predict behavior. Clustering analysis is primarily used when there are no assumptions made about the likely relationships within the data. Regression analysis Is used when you want to predict the value of a continuous dependent from a number of independent variables.

Benefits

There are many benefits of using Association rules like finding the pattern that helps understand the correlations and co-occurrences between data sets. A very good real-world example that uses Association rules would be medicine. Medicine uses Association rules to help diagnose patients. When diagnosing patients there are many variables to consider as many diseases will share similar symptoms. With the use of the Association rules, doctors can determine the conditional probability of an illness by comparing symptom relationships from past cases.

Downfalls

However, Association rules also lead to many different downfalls such as finding the appropriate parameter and threshold settings for the mining algorithm. But there is also the downfall of having a large number of discovered rules. The reason is that this does not guarantee that the rules will be found relevant, but it could also cause the algorithm to have low performance. Sometimes the implemented algorithms will contain too many variables and parameters. For someone that doesn’t have a good concept of data mining, this might cause them to have trouble understanding it.

Thresholds

Frequent itemset lattice, where the color of the box indicates how many transactions contain the combination of items. Note that lower levels of the lattice can contain at most the minimum number of their parents' items; e.g. {ac} can have only at most items. This is called the downward-closure property.

When using Association rules, you are most likely to only use Support and Confidence. However, this means you have to satisfy a user-specified minimum support and a user-specified minimum confidence at the same time. Usually, the Association rule generation is split into two different steps that needs to be applied:

  1. A minimum Support threshold to find all the frequent itemsets that are in the database.
  2. A minimum Confidence threshold to the frequent itemsets found to create rules.
Table 1. Example of Threshold for Support and Confidence.
Items Support Confidence
Items Support Confidence
Item A 30% 50% Item C 45% 55%
Item B 15% 25% Item A 30% 50%
Item C 45% 55% Item D 35% 40%
Item D 35% 40% Item B 15% 25%

The Support Threshold is 30%, Confidence Threshold is 50%

The Table on the left is the original unorganized data and the table on the right is organized by the thresholds. In this case Item C is better than the thresholds for both Support and Confidence which is why it is first. Item A is second because its threshold values are spot on. Item D has met the threshold for Support but not Confidence. Item B has not met the threshold for either Support or Confidence and that is why it is last.

To find all the frequent itemsets in a database is not an easy task since it involves going through all the data to find all possible item combinations from all possible itemsets. The set of possible itemsets is the power set over I and has size , of course this means to exclude the empty set which is not considered to be a valid itemset. However, the size of the power set will grow exponentially in the number of item n that is within the power set I. An efficient search is possible by using the downward-closure property of support (also called anti-monotonicity). This would guarantee that a frequent itemset and all its subsets are also frequent and thus will have no infrequent itemsets as a subset of a frequent itemset. Exploiting this property, efficient algorithms (e.g., Apriori and Eclat) can find all frequent itemsets.

Useful Concepts

Table 2. Example database with 5 transactions and 5 items
transaction ID milk bread butter beer diapers eggs fruit
1 1 1 0 0 0 0 1
2 0 0 1 0 0 1 1
3 0 0 0 1 1 0 0
4 1 1 1 0 0 1 1
5 0 1 0 0 0 0 0

To illustrate the concepts, we use a small example from the supermarket domain. Table 2 shows a small database containing the items where, in each entry, the value 1 means the presence of the item in the corresponding transaction, and the value 0 represents the absence of an item in that transaction. The set of items is .

An example rule for the supermarket could be meaning that if butter and bread are bought, customers also buy milk.

In order to select interesting rules from the set of all possible rules, constraints on various measures of significance and interest are used. The best-known constraints are minimum thresholds on support and confidence.

Let be itemsets, an association rule and T a set of transactions of a given database.

Note: this example is extremely small. In practical applications, a rule needs a support of several hundred transactions before it can be considered statistically significant, and datasets often contain thousands or millions of transactions.

Support

Support is an indication of how frequently the itemset appears in the dataset.

In our example, it can be easier to explain support by writing where A and B are separate transactions that were made within the total set of transactions recorded.

Using Table 2 as an example, the itemset has a support of since it occurs in 20% of all transactions (1 out of 5 transactions). The argument of support of X is a set of preconditions, and thus becomes more restrictive as it grows (instead of more inclusive).

Furthermore, the itemset has a support of as it appears in 20% of all transactions as well.

When using antecedents and consequents, it allows a data miner to determine the support of multiple items being bought together in comparison to the whole data set. For example, Table 2 shows that if milk is bought, then bread is bought has a support of 0.4 or 40%. This because in 2 out 5 of the transactions, milk as well as bread are bought. In smaller data sets like this example, it is harder to see a strong correlation when there are few samples, but when the data set grows larger, support can be used to find correlation between two or more products in the supermarket example.

Minimum support thresholds are useful for determining which itemsets are preferred or interesting.

If we set the support threshold to ≥0.4 in Table 3, then the would be removed since it did not meet the minimum threshold of 0.4. Minimum threshold is used to remove samples where there is not a strong enough support or confidence to deem the sample as important or interesting in the dataset.

Another way of finding interesting samples is to find the value of (support)X(confidence); this allows a data miner to see the samples where support and confidence are high enough to be highlighted in the dataset and prompt a closer look at the sample to find more information on the connection between the items.

Support can be beneficial for finding the connection between products in comparison to the whole dataset, whereas confidence looks at the connection between one or more items and another item. Below is a table that shows the comparison and contrast between support and support x confidence, using the information from Table 4 to derive the confidence values.

Table 3. Example of Support, and support X confidence
if Antecedent then Consequent support support X confidence
if buy milk, then buy bread 2/5= 0.4 0.4X1.0= 0.4
if buy milk, then buy eggs 1/5= 0.2 0.2X0.5= 0.1
if buy bread, then buy fruit 2/5= 0.4 0.4X0.66= 0.264
if buy fruit, then buy eggs 2/5= 0.4 0.4X0.66= 0.264
if buy milk and bread, then buy fruit 2/5= 0.4 0.4X1.0= 0.4

The support of X with respect to T is defined as the proportion of transactions in the dataset which contains the itemset X. Denoting a transaction by where i is the unique identifier of the transaction and t is its itemset, the support may be written as:

This notation can be used when defining more complicated datasets where the items and itemsets may not be as easy as our supermarket example above. Other examples of where support can be used is in finding groups of genetic mutations that work collectively to cause a disease, investigating the number of subscribers that respond to upgrade offers, and discovering which products in a drug store are never bought together.

Confidence

Confidence is the percentage of all transactions satisfying X that also satisfy Y.

With respect to T, the confidence value of an association rule, often denoted as , is the ratio of transactions containing both X and Y to the total amount of X values present, where X is the antecedent and Y is the consequent.

Confidence can also be interpreted as an estimate of the conditional probability , the probability of finding the RHS of the rule in transactions under the condition that these transactions also contain the LHS.

It is commonly depicted as:

The equation illustrates that confidence can be computed by calculating the co-occurrence of transactions X and Y within the dataset in ratio to transactions containing only X. This means that the number of transactions in both X and Y is divided by those just in X .

For example, Table 2 shows the rule which has a confidence of in the dataset, which denotes that every time a customer buys butter and bread, they also buy milk. This particular example demonstrates the rule being correct 100% of the time for transactions containing both butter and bread. The rule , however, has a confidence of . This suggests that eggs are bought 67% of the times that fruit is brought. Within this particular dataset, fruit is purchased a total of 3 times, with two of those times consisting of egg purchases.

For larger datasets, a minimum threshold, or a percentage cutoff, for the confidence can be useful for determining item relationships. When applying this method to some of the data in Table 2, information that does not meet the requirements are removed. Table 4 shows association rule examples where the minimum threshold for confidence is 0.5 (50%). Any data that does not have a confidence of at least 0.5 is omitted. Generating thresholds allow for the association between items to become stronger as the data is further researched by emphasizing those that co-occur the most. The table uses the confidence information from Table 3 to implement the Support x Confidence column, where the relationship between items via their both confidence and support, instead of just one concept, is highlighted. Ranking the rules by Support x Confidence multiples the confidence of a particular rule to its support and is often implemented for a more in-depth understanding of the relationship between the items.

Table 4. Example of Confidence and Support x Confidence
if Antecedent then Consequent Confidence Support x Confidence
if buy milk, then buy bread 2/2= 1.0 0.4X1.0= 0.4
if buy milk, then buy eggs 1/2= 0.5 0.2X0.5= 0.1
if buy bread, then buy fruit 2/3= 0.66 0.4X0.66= 0.264
if buy fruit, then buy eggs 2/3= 0.66 0.4X0.66= 0.264
if buy milk and bread, then buy fruit 2/2= 1.0 0.4X1.0= 0.4

Overall, using confidence in association rule mining is great way to bring awareness to data relations. Its greatest benefit is highlighting the relationship between particular items to one another within the set, as it compares co-occurrences of items to the total occurrence of the antecedent in the specific rule. However, confidence is not the optimal method for every concept in association rule mining. The disadvantage of using it is that it does not offer multiple difference outlooks on the associations. Unlike support, for instance, confidence does not provide the perspective of relationships between certain items in comparison to the entire dataset, so while milk and bread, for example, may occur 100% of the time for confidence, it only has a support of 0.4 (40%). This is why it is important to look at other viewpoints, such as Support x Confidence, instead of solely relying on one concept incessantly define the relationships.

Lift

The lift of a rule is defined as:

or the ratio of the observed support to that expected if X and Y were independent.

For example, the rule has a lift of .

If the rule had a lift of 1, it would imply that the probability of occurrence of the antecedent and that of the consequent are independent of each other. When two events are independent of each other, no rule can be drawn involving those two events.

If the lift is > 1, that lets us know the degree to which those two occurrences are dependent on one another, and makes those rules potentially useful for predicting the consequent in future data sets.

If the lift is < 1, that lets us know the items are substitute to each other. This means that presence of one item has negative effect on presence of other item and vice versa.

The value of lift is that it considers both the support of the rule and the overall data set.

Conviction

The conviction of a rule is defined as .

For example, the rule has a conviction of , and can be interpreted as the ratio of the expected frequency that X occurs without Y (that is to say, the frequency that the rule makes an incorrect prediction) if X and Y were independent divided by the observed frequency of incorrect predictions. In this example, the conviction value of 1.2 shows that the rule would be incorrect 20% more often (1.2 times as often) if the association between X and Y was purely random chance.

Alternative measures of interestingness

In addition to confidence, other measures of interestingness for rules have been proposed. Some popular measures are:

  • All-confidence
  • Collective strength
  • Leverage

Several more measures are presented and compared by Tan et al. and by Hahsler. Looking for techniques that can model what the user has known (and using these models as interestingness measures) is currently an active research trend under the name of "Subjective Interestingness."

History

The concept of association rules was popularized particularly due to the 1993 article of Agrawal et al., which has acquired more than 23,790 citations according to Google Scholar, as of April 2021, and is thus one of the most cited papers in the Data Mining field. However, what is now called "association rules" is introduced already in the 1966 paper on GUHA, a general data mining method developed by Petr Hájek et al.

An early (circa 1989) use of minimum support and confidence to find all association rules is the Feature Based Modeling framework, which found all rules with and greater than user defined constraints.

Statistically sound associations

One limitation of the standard approach to discovering associations is that by searching massive numbers of possible associations to look for collections of items that appear to be associated, there is a large risk of finding many spurious associations. These are collections of items that co-occur with unexpected frequency in the data, but only do so by chance. For example, suppose we are considering a collection of 10,000 items and looking for rules containing two items in the left-hand-side and 1 item in the right-hand-side. There are approximately 1,000,000,000,000 such rules. If we apply a statistical test for independence with a significance level of 0.05 it means there is only a 5% chance of accepting a rule if there is no association. If we assume there are no associations, we should nonetheless expect to find 50,000,000,000 rules. Statistically sound association discovery controls this risk, in most cases reducing the risk of finding any spurious associations to a user-specified significance level.

Algorithms

Many algorithms for generating association rules have been proposed.

Some well-known algorithms are Apriori, Eclat and FP-Growth, but they only do half the job, since they are algorithms for mining frequent itemsets. Another step needs to be done after to generate rules from frequent itemsets found in a database.

Apriori algorithm

Apriori is given by R. Agrawal and R. Srikant in 1994 for frequent item set mining and association rule learning. It proceeds by identifying the frequent individual items in the database and extending them to larger and larger item sets as long as those item sets appear sufficiently often. The name of the algorithm is Apriori because it uses prior knowledge of frequent itemset properties.

The control flow diagram for the Apriori algorithm

Overview: Apriori uses a "bottom up" approach, where frequent subsets are extended one item at a time (a step known as candidate generation), and groups of candidates are tested against the data. The algorithm terminates when no further successful extensions are found. Apriori uses breadth-first search and a Hash tree structure to count candidate item sets efficiently. It generates candidate item sets of length  from item sets of length . Then it prunes the candidates which have an infrequent sub pattern. According to the downward closure lemma, the candidate set contains all frequent -length item sets. After that, it scans the transaction database to determine frequent item sets among the candidates.

Example: Assume that each row is a cancer sample with a certain combination of mutations labeled by a character in the alphabet. For example a row could have {a, c} which means it is affected by mutation 'a' and mutation 'c'.

Input Set
{a, b} {c, d} {a, d} {a, e} {b, d} {a, b, d} {a, c, d} {a, b, c, d}

Now we will generate the frequent item set by counting the number of occurrences of each character. This is also known as finding the support values. Then we will prune the item set by picking a minimum support threshold. For this pass of the algorithm we will pick 3.

Support Values
a b c d
6 4 3 6

Since all support values are three or above there is no pruning. The frequent item set is {a}, {b}, {c}, and {d}. After this we will repeat the process by counting pairs of mutations in the input set.

Support Values
{a, b} {a, c} {a, d} {b, c} {b, d} {c, d}
3 2 4 1 3 4

Now we will make our minimum support value 4 so only {a, d} and {c, d} will remain after pruning. Now we will use the frequent item set to make combinations of triplets. We will then repeat the process by counting occurrences of triplets of mutations in the input set.

Support Values
{a, c, d}
2

Since we only have one item the next set of combinations of quadruplets is empty so the algorithm will stop.

Advantages and Limitations:

Apriori has some limitations. Candidate generation can result in large candidate sets. For example a 10^4 frequent 1-itemset will generate a 10^7 candidate 2-itemset. The algorithm also needs to frequently scan the database, to be specific n+1 scans where n is the length of the longest pattern. Apriori is slower than the Eclat algorithm. However, Apriori performs well compared to Eclat when the dataset is large. This is because in the Eclat algorithm if the dataset is too large the tid-lists become too large for memory. FP-growth outperforms the Apriori and Eclat. This is due to the FP-growth algorithm not having candidate generation or test, using a compact data structure, and only having one database scan.

Eclat algorithm

Eclat (alt. ECLAT, stands for Equivalence Class Transformation) is a depth-first search algorithm based on set intersection. It is suitable for both sequential as well as parallel execution with locality-enhancing properties.

FP-growth algorithm

FP stands for frequent pattern.

In the first pass, the algorithm counts the occurrences of items (attribute-value pairs) in the dataset of transactions, and stores these counts in a 'header table'. In the second pass, it builds the FP-tree structure by inserting transactions into a trie.

Items in each transaction have to be sorted by descending order of their frequency in the dataset before being inserted so that the tree can be processed quickly. Items in each transaction that do not meet the minimum support requirement are discarded. If many transactions share most frequent items, the FP-tree provides high compression close to tree root.

Recursive processing of this compressed version of the main dataset grows frequent item sets directly, instead of generating candidate items and testing them against the entire database (as in the apriori algorithm).

Growth begins from the bottom of the header table i.e. the item with the smallest support by finding all sorted transactions that end in that item. Call this item .

A new conditional tree is created which is the original FP-tree projected onto . The supports of all nodes in the projected tree are re-counted with each node getting the sum of its children counts. Nodes (and hence subtrees) that do not meet the minimum support are pruned. Recursive growth ends when no individual items conditional on meet the minimum support threshold. The resulting paths from root to will be frequent itemsets. After this step, processing continues with the next least-supported header item of the original FP-tree.

Once the recursive process has completed, all frequent item sets will have been found, and association rule creation begins.

Others

ASSOC

The ASSOC procedure is a GUHA method which mines for generalized association rules using fast bitstrings operations. The association rules mined by this method are more general than those output by apriori, for example "items" can be connected both with conjunction and disjunctions and the relation between antecedent and consequent of the rule is not restricted to setting minimum support and confidence as in apriori: an arbitrary combination of supported interest measures can be used.

OPUS search

OPUS is an efficient algorithm for rule discovery that, in contrast to most alternatives, does not require either monotone or anti-monotone constraints such as minimum support. Initially used to find rules for a fixed consequent it has subsequently been extended to find rules with any item as a consequent. OPUS search is the core technology in the popular Magnum Opus association discovery system.

Lore

A famous story about association rule mining is the "beer and diaper" story. A purported survey of behavior of supermarket shoppers discovered that customers (presumably young men) who buy diapers tend also to buy beer. This anecdote became popular as an example of how unexpected association rules might be found from everyday data. There are varying opinions as to how much of the story is true. Daniel Powers says:

In 1992, Thomas Blischok, manager of a retail consulting group at Teradata, and his staff prepared an analysis of 1.2 million market baskets from about 25 Osco Drug stores. Database queries were developed to identify affinities. The analysis "did discover that between 5:00 and 7:00 p.m. that consumers bought beer and diapers". Osco managers did NOT exploit the beer and diapers relationship by moving the products closer together on the shelves.

Other types of association rule mining

Multi-Relation Association Rules: Multi-Relation Association Rules (MRAR) are association rules where each item may have several relations. These relations indicate indirect relationship between the entities. Consider the following MRAR where the first item consists of three relations live in, nearby and humid: “Those who live in a place which is nearby a city with humid climate type and also are younger than 20 -> their health condition is good”. Such association rules are extractable from RDBMS data or semantic web data.

Contrast set learning is a form of associative learning. Contrast set learners use rules that differ meaningfully in their distribution across subsets.

Weighted class learning is another form of associative learning in which weight may be assigned to classes to give focus to a particular issue of concern for the consumer of the data mining results.

High-order pattern discovery facilitate the capture of high-order (polythetic) patterns or event associations that are intrinsic to complex real-world data. 

K-optimal pattern discovery provides an alternative to the standard approach to association rule learning that requires that each pattern appear frequently in the data.

Approximate Frequent Itemset mining is a relaxed version of Frequent Itemset mining that allows some of the items in some of the rows to be 0.

Generalized Association Rules hierarchical taxonomy (concept hierarchy)

Quantitative Association Rules categorical and quantitative data

Interval Data Association Rules e.g. partition the age into 5-year-increment ranged

Sequential pattern mining discovers subsequences that are common to more than minsup sequences in a sequence database, where minsup is set by the user. A sequence is an ordered list of transactions.

Subspace Clustering, a specific type of Clustering high-dimensional data, is in many variants also based on the downward-closure property for specific clustering models.

Warmr is shipped as part of the ACE data mining suite. It allows association rule learning for first order relational rules.

Magnetotellurics

From Wikipedia, the free encyclopedia
 
Magnetotelluric station

Magnetotellurics (MT) is an electromagnetic geophysical method for inferring the earth's subsurface electrical conductivity from measurements of natural geomagnetic and geoelectric field variation at the Earth's surface.

Investigation depth ranges from 300 m below ground by recording higher frequencies down to 10,000 m or deeper with long-period soundings. Proposed in Japan in the 1940s, and France and the USSR during the early 1950s, MT is now an international academic discipline and is used in exploration surveys around the world.

Commercial uses include hydrocarbon (oil and gas) exploration, geothermal exploration, carbon sequestration, mining exploration, as well as hydrocarbon and groundwater monitoring. Research applications include experimentation to further develop the MT technique, long-period deep crustal exploration, deep mantle probing, sub-glacial water flow mapping, and earthquake precursor research.

History

The magnetotelluric technique was introduced independently by Japanese scientists in 1948 (Hirayama, Rikitake), Russian geophysicist Andrey Nikolayevich Tikhonov in 1950 and the French geophysicist Louis Cagniard in 1953. With advances in instrumentation, processing and modelling, magnetotellurics has become one of the most important tools in deep Earth research.

Since first being created in the 1950s, magnetotelluric sensors, receivers and data processing techniques have followed the general trends in electronics, becoming less expensive and more capable with each generation. Major advances in MT instrumentation and technique include the shift from analog to digital hardware, the advent of remote referencing, GPS time-based synchronization, and 3D data acquisition and processing.

Commercial applications

Hydrocarbon exploration

For hydrocarbon exploration, MT is mainly used as a complement to the primary technique of reflection seismology exploration. While seismic imaging is able to image subsurface structure, it cannot detect the changes in resistivity associated with hydrocarbons and hydrocarbon-bearing formations. MT does detect resistivity variations in subsurface structures, which can differentiate between structures bearing hydrocarbons and those that do not.

At a basic level of interpretation, resistivity is correlated with different rock types. High-velocity layers are typically highly resistive, whereas sediments – porous and permeable – are typically much less resistive. While high-velocity layers are an acoustic barrier and make seismic ineffective, their electrical resistivity means the magnetic signal passes through almost unimpeded. This allows MT to see deep beneath these acoustic barrier layers, complementing the seismic data and assisting interpretation. 3-D MT survey results in Uzbekistan (32 x 32 grid of soundings) have guided further seismic mapping of a large known gas-bearing formation with complex subsurface geology.

China National Petroleum Corporation (CNPC) and Nord-West Ltd use onshore MT more than any other oil company in the world, conducting thousands of MT soundings for hydrocarbon exploration and mapping throughout the globe.

Mining exploration

MT is used for various base metals (e.g. nickel) and precious metals exploration, as well as for kimberlite mapping.

INCO's 1991 proof-of-concept study in Sudbury, Ontario, Canada sensed a 1750-meter-deep nickel deposit. Falconbridge followed with a feasibility study in 1996 that accurately located two Ni-Cu mineralized zones at about 800 m and 1350 m depth. Since then, both major and junior mining companies are increasingly using MT and audio-magnetotellurics (AMT) for both brownfields and greenfields exploration. Significant MT mapping work has been done on areas of the Canadian Shield.

Diamond exploration, by detecting kimberlites, is also a proven application.

Geothermal exploration

MT geothermal exploration measurements allow detection of resistivity anomalies associated with productive geothermal structures, including faults and the presence of a cap rock, and allow for estimation of geothermal reservoir temperatures at various depths. Dozens of MT geothermal exploration surveys have been completed in Japan and the Philippines since the early 1980s, helping to identify several hundred megawatts of renewable power at places such as the Hatchobaru plant on Kyushu and the Togonang plant on Leyte. Geothermal exploration with MT has also been done in the United States, Iceland, New Zealand, Hungary, China, Ethiopia, Indonesia, Peru, Australia, and India.

Other

MT is also used for groundwater exploration and mapping research, hydrocarbon reservoir monitoring, deep investigation (100 km) of the electrical properties of the bedrock for high-voltage direct current (HVDC) transmission systems, carbon dioxide sequestration, and other environmental engineering applications (e.g. nuclear blast site monitoring and nuclear waste disposal site monitoring).

Research applications

Crust and mantle

MT has been used to investigate the distribution of silicate melts in the Earth's mantle and crust; large investigations have focused on the continental US (National Science Foundation EarthScope MT Program), the East Pacific Rise, Australia (AusLAMP MT Program), Southern Africa (SAMTEX MT Project), China (Part of the Sinoprobe project) and the Tibetan Plateau. Research work relevant to MT method may include understanding the plate-tectonic processes in the highly complex geological regions such as the collisions of the African and European plates and in orogenic processes in Western Europe such as the Variscan Orogeny and the Caledonian Orogeny

Earthquake precursor prediction

Fluctuations in the MT signal may be able to predict the onset of seismic events. Stationary MT monitoring systems have been installed in Japan since April 1996, providing a continuous recording of MT signals at the Wakuya Station (previously at the Mizusawa Geodetic Observatory) and the Esashi Station of the Geographical Survey Institute of Japan (GSIJ). These stations measure fluctuations in the Earth's electromagnetic field that correspond with seismic activity. The raw geophysical time-series data from these monitoring stations is freely available to the scientific community, enabling further study of the interaction between electromagnetic events and earthquake activity.

Additional MT earthquake precursor monitoring stations in Japan are located in Kagoshima, in Sawauchi, and on Shikoku. Similar stations are also deployed in Taiwan on Penghu Island, as well as in the Fushan Reserve on the island of Taiwan proper.

POLARIS is a Canadian research program investigating the structure and dynamics of the Earth's lithosphere and the prediction of earthquake ground motion.

Theory and practice

Energy sources

Solar energy and lightning cause natural variations in the earth's magnetic field, inducing electric currents (known as telluric currents) under the Earth's surface.

Different rocks, sediments and geological structures have a wide range of different electrical conductivities. Measuring electrical resistivity allows different materials and structures to be distinguished from one another and can improve knowledge of tectonic processes and geologic structures.

The Earth's naturally varying electric and magnetic fields are measured over a wide range of magnetotelluric frequencies from 10,000 Hz to 0.0001 Hz (10,000 s). These fields are due to electric currents flowing in the Earth and the magnetic fields that induce these currents. The magnetic fields are produced mainly by the interaction between the solar wind and the magnetosphere. In addition, worldwide thunderstorm activity causes magnetic fields at frequencies above 1 Hz. Combined, these natural phenomena create strong MT source signals over the entire frequency spectrum.

The ratio of the electric field to magnetic field provides simple information about subsurface conductivity. Because the skin effect phenomenon affects the electromagnetic fields, the ratio at higher frequency ranges gives information on the shallow Earth, whereas deeper information is provided by the low-frequency range. The ratio is usually represented as both apparent resistivity as a function of frequency and phase as a function of frequency.

A subsurface resistivity model is then created using this tensor.

Depth and resolution

MT measurements can investigate depths from about 300 m down to hundreds of kilometers, though investigations in the range of 500 m to 10,000 m are typical. Greater depth requires measuring lower frequencies, which in turn requires longer recording times. Very deep, very long-period measurements (mid-crust through upper mantle depths), may require recordings of several days to weeks or more to obtain satisfactory data quality.

Horizontal resolution of MT mainly depends on the distance between sounding locations- closer sounding locations increase the horizontal resolution. Continuous profiling (known as Emap) has been used, with only meters between the edges of each telluric dipole.

Vertical resolution of MT mainly depends on the frequency being measured, as lower frequencies have greater depths of penetration. Accordingly, vertical resolution decreases as depth of investigation increases.

Signal strength and recording times

Magnetic fields in the frequency range of 1 Hz to approximately 20 kHz are part of the audio-magnetotelluric (AMT) range. These are parallel to the Earth surface and move towards the Earth's centre. This large frequency band allows for a range of depth penetration from several metres to several kilometres below the Earth's surface. Due to the nature of magnetotelluric source, the waves generally fluctuate in amplitude height. Long recording times are needed to ascertain usable reading due to the fluctuations and the low signal strength. Generally, the signal is weak between 1 and 5 kHz, which is a crucial range in detecting the top 100 m of geology. The magnetotelluric method is also used in marine environments for hydrocarbon exploration and lithospheric studies. Due to the screening effect of the electrically conductive sea water, a usable upper limit of the spectrum is around 1 Hz.

2D and 3D magnetotellurics

Two-dimensional surveys consist of a longitudinal profile of MT soundings over the area of interest, providing two-dimensional "slices" of subsurface resistivity.

Three-dimensional surveys consist of a loose grid pattern of MT soundings over the area of interest, providing a more sophisticated three-dimensional model of subsurface resistivity.

Variants

Audio-magnetotellurics

Audio-magnetotellurics (AMT) is a higher-frequency magnetotelluric technique for shallower investigations. While AMT has less depth penetration than MT, AMT measurements often take only about one hour to perform (but deep AMT measurements during low-signal strength periods may take up to 24 hours) and use smaller and lighter magnetic sensors. Transient AMT is an AMT variant that records only temporarily during periods of more intense natural signal (transient impulses), improving signal-to-noise-ratio at the expense of strong linear polarization.

Controlled source electromagnetics

CSEM controlled source electro-magnetic is a deep-water offshore variant of controlled source audio magnetotellurics; CSEM is the name used in the offshore oil and gas industry. and for onshore exploration mostly Lotem is used in Russia, China the USA and Europe

Onshore CSEM / CSAMT may be effective where electromagnetic cultural noise (e.g. power lines, electric fences) present interference problems for natural-source geophysical methods. An extensive grounded wire (2 km or more) has currents at a range of frequencies (0.1 Hz to 100 kHz) passed through it. The electric field parallel to the source and the magnetic field which is at right angles are measured. The resistivity is then calculated, and the lower the resistivity, the more likely there is a conductive target (graphite, nickel ore or iron ore). CSAMT is also known in the oil and gas industry as onshore controlled source electromagnetics (Onshore CSEM).

An offshore variant of MT, the marine magnetotelluric (MMT) method, uses instruments and sensors in pressure housings deployed by ship into shallow coastal areas where water is less than 300 m deep. A derivative of MMT is offshore single-channel measurement of the vertical magnetic field only (the Hz, or "tipper"), which eliminates the need for telluric measurements and horizontal magnetic measurements.

Exploration surveys

MT exploration surveys are done to acquire resistivity data which can be interpreted to create a model of the subsurface. Data is acquired at each sounding location for a period of time (overnight soundings are common), with physical spacing between soundings dependent on the target size and geometry, local terrain constraints and financial cost. Reconnaissance surveys can have spacings of several kilometres, while more detailed work can have 200 m spacings, or even adjacent soundings (dipole-to-dipole).

The HSE impact of MT exploration is relatively low because of light-weight equipment, natural signal sources, and reduced hazards compared to other types of exploration (e.g. no drills, no explosives, and no high currents).

Remote reference soundings

Remote Reference is an MT technique used to account for cultural electrical noise by acquiring simultaneous data at more than one MT station. This greatly improves data quality, and may allow acquisition in areas where the natural MT signal is difficult to detect because of man-made EM interference.

Equipment

A typical full suite of MT equipment (for a "five component" sounding) consists of a receiver instrument with five sensors: three magnetic sensors (typically induction coil sensors), and two telluric (electric) sensors. For exclusively long-period MT (frequencies below approximately 0.1 Hz), the three discrete broadband magnetic field sensors may be substituted by a single compact triaxial fluxgate magnetometer. In many situations, only the telluric sensors will be used, and magnetic data borrowed from other nearby soundings to reduce acquisition costs.

A complete five-component set of MT equipment can be backpack-carried by a small field team (2 to 4 persons) or carried by a light helicopter, allowing deployment in remote and rugged areas. Most MT equipment is capable of reliable operation over a wide range of environmental conditions, with ratings of typically −25 °C to +55 °C, from dry desert to high-humidity (condensing) and temporary full immersion.

Data processing and interpretation

Post-acquisition processing is required to transform raw time-series data into frequency-based inversions. The resulting output of the processing program is used as the input for subsequent interpretation. Processing may include the use of remote reference data or local data only.

Processed MT data is modelled using various techniques to create a subsurface resistivity map, with lower frequencies generally corresponding to greater depth below ground. Anomalies such as faults, hydrocarbons, and conductive mineralization appear as areas of higher or lower resistivity from surrounding structures. Various software packages are used for interpretation (inversion) of magnetotelluric data, where apparent resistivity is used to create a model of the subsurface.

Instrument and sensor manufacturers

Three companies supply most of the commercial-use world market: one in the United States (Zonge International, Inc.), one in Canada; (Phoenix Geophysics, Ltd.), and one in Germany (Metronix Messgeraete und Elektronik GmbH).)

Government agencies and smaller companies producing MT instrumentation for internal use include Vega Geophysics, Ltd. in Russia and the Russian Academy of Sciences (SPbF IZMIRAN); and the National Space Research Institute of Ukraine.

Tuesday, July 5, 2022

Visual memory

From Wikipedia, the free encyclopedia
 
Close up of the human eye, the main organ of visual sensation

Visual memory describes the relationship between perceptual processing and the encoding, storage and retrieval of the resulting neural representations. Visual memory occurs over a broad time range spanning from eye movements to years in order to visually navigate to a previously visited location. Visual memory is a form of memory which preserves some characteristics of our senses pertaining to visual experience. We are able to place in memory visual information which resembles objects, places, animals or people in a mental image. The experience of visual memory is also referred to as the mind's eye through which we can retrieve from our memory a mental image of original objects, places, animals or people. Visual memory is one of several cognitive systems, which are all interconnected parts that combine to form the human memory. Types of palinopsia, the persistence or recurrence of a visual image after the stimulus has been removed, is a dysfunction of visual memory.

Neuroanatomy

In humans, areas specialized for visual object recognition in the ventral stream have a more inferior location in the temporal cortex, whereas areas specialized for the visual-spatial location of objects in the dorsal stream have a more superior location in the parietal cortex. However, these two streams hypothesis, although useful, are a simplification of the visual system because the two streams maintain intercommunication along their entire rostral course.

Posterior parietal cortex

Posterior parietal cortex (light green) is shown at the posterior area of the parietal lobe

The posterior parietal cortex is a portion of the parietal lobe, which manipulates mental images, and integrates sensory and motor portions of the brain.

A majority of experiments highlights a role of human posterior parietal cortex in visual working memory and attention. We therefore have to establish a clear separation of visual memory and attention from processes related to the planning of goal-directed motor behaviors.

We can only hold in mind a minute fraction of the visual scene. These mental representations are stored in visual short-term memory. Activity in the posterior parietal cortex is tightly correlated with the limited amount of scene information that can be stored in visual short-term memory. These results suggest that the posterior parietal cortex is a key neural locus of our impoverished mental representation of the visual world.

The posterior cortex might act as a capacity-limited store for the representation of the visual scene, the frontal/prefrontal cortex might be necessary for the consolidation and/or maintenance of this store, especially during extended retention intervals.

Visual cortex

The dorsal stream (green) and ventral stream (purple) are both actively involved in visual memory. Both pathways originate in the visual cortex.

There is a visual cortex in each hemisphere of the brain, much of which is located in the Occipital lobe. The left hemisphere visual cortex receives signals mainly from the right visual field and the right visual cortex mainly from the left visual field, although each cortex receives a considerable amount of information from the ipsilateral visual field as well. The visual cortex also receives information from subcortical regions, such as the lateral geniculate body, located in the thalamus. However, ample evidence indicates that object identity and location are preferentially processed in ventral (occipito-temporal) and dorsal (occipito-parietal) cortical visual streams, respectively. Comparison of rCBF during performance of the two tasks again revealed differences between the ventral and dorsal pathways.

Dorsal stream pathway

The dorsal stream pathway is mainly involved in the visual-spatial location of objects in the external world, and it is also known colloquially as the 'where' pathway. The dorsal stream pathway is also involved in the guidance of movements (e.g., reaching for an object in space), and is therefore implicated in the analysis of the movement of objects in addition to their spatial locations.

The dorsal stream pathway begins with purely visual information in the occipital lobe, and then this information is transferred to the parietal lobe for spatial awareness functions. Specifically, the posterior parietal cortex is essential for "the perception and interpretation of spatial relationships, accurate body image, and the learning of tasks involving coordination of the body in space."

Ventral stream pathway

The ventral stream pathway is mainly involved in object recognition, and is known colloquially as the 'what' pathway. It has connections to the medial temporal lobe (which is involved in the storage of long-term memories), the limbic system (which regulates emotions), and the dorsal stream pathway (which is involved in the visual-spatial locations and motions of objects). Therefore, the ventral stream pathway not only deals with the recognition of objects in the external world, but also the emotional judgement and analysis of these objects.

The ventral stream pathway begins with purely visual information in the primary visual cortex (occipital lobe), and then this information is transferred to the temporal lobe.

Occipital lobes

Located at the back of the brain, the occipital lobes receive and process visual information. The occipital lobes also process colors and shapes. Whereas the right occipital lobe interprets images from the left visual space, the left occipital lobe interprets images from the right visual space. Damage to the occipital lobes can permanently damage visual perception.

Occipital lobe injury complications

Damage to the occipital lobe is characterized by loss of visual capability and the inability to identify colors both important processes in visual memory.

Short term visual memory

Visual short term memory is the capacity for holding a small amount of visual information in mind in an active, readily available state for a short period of time (usually no more than 30 seconds). Although visual short term memory is essential for the execution of a wide array of perceptual and cognitive functions, and is supported by an extensive network of brain regions, its storage capacity is severely limited.

Visual short-term memory storage is mediated by distinctive posterior brain mechanisms, such that capacity is determined both by a fixed number of objects and by object complexity.

Long term visual memory

Recall of the patterns from long term visual memory is associated with rCBF increases in different areas of the prefrontal cortex and the anterior cingulate cortex. The retrieval of long term visual memories is associated with activation of both anterior and posterior temporal cortices. Posterior temporal cortical regions are more associated with retrieval of category-specific aspects of visual memory, whereas anterior regions of the temporal cortex are more associated with category-independent visual memory.

Methods of study

Benton visual retention test

An example of the Benton Visual Retention Test. The target stimulus is presented at the top, after a delay participants are asked to recall the correct target stimulus from the list of design cards.

The Benton Visual Retention Test is an assessment of visual perception, and visual memory abilities. More than 50 years of proven clinical use is the staple of the Benton Visual Retention Test. This test has proven its sensitivity to reading disabilities, nonverbal learning disabilities, traumatic brain injury, attention-deficit disorder, alzheimer's, and other forms of dementia. During testing participants are presented with 10 cards for 10 seconds with unique designs on each. After the time has passed participants are asked to immediately reproduce the designs from each card using their visual memory. In the second stage participants are asked to copy each of the 10 card designs while the cards are in view. The participants results from each task are then assessed and placed into six categories; omissions, distortions, preservations, rotations, misplacements, and sizing errors. The further the participant's scores varies from the averages provided in the Benton Visual Retention Test manual the worse the participant is assessed to be on visual memory ability. The Benton Visual Retention Test has proved to be a generalizable test with the ability to be accurately administered to participants aged 8-adult, and no gender effect. Some studies have suggested a significant gender and education interaction indicating that an age-associated decline in visual memory performance may be more prominent for those individuals with a lower education level.

Neuroimaging tests

An example of a colored geometrical pattern a subject would encode, store, and recall while performing a visual memory neuroimaging test.

Neuroimaging studies focus on the neural networks involved in visual memory using methods designed to activate brain areas involved in encoding, storage, and recall. These studies involve the use of one or multiple types of brain imaging techniques designed to measure timing or activation within the brain. The data collected from neuroimaging studies gives researchers the ability to visualize which brain regions are activated in specific cognitive visual memory tasks. With the use of brain imaging devices researchers able to further investigate memory performance above and beyond standard tests based on exact response times, and activation.

Control condition

The subject's resting brain activation level is first determined in order to form a control or 'baseline' to measure from. Subjects are blindfolded and instructed to lay motionless while simultaneously eliminating any visual imagery present in their mind's eye. These instructions are intended to minimize the activation of brain regions involved in visual memory to form a true resting brain state. After the scan is complete a control has been formed which can be compared with activated regions of the brain while performing visual memory tasks.

Activation condition

During encoding, participants are typically exposed to 1–10 visual patterns while connected to a brain imaging device. As the subject encodes the visual patterns researchers are able to directly view the activation of areas involved in visual memory encoding. During recall subjects again need to have all visual stimuli removed by means of a dark room or blindfolding to avoid interfering activation of other visual areas in the brain. Subjects are asked to recall each image clearly in their mind's eye. While recalling the images researchers are able view the areas activated by the visual memory task. Comparing the control 'baseline' state to the activated areas during the visual memory task allows researchers to view which areas are used during visual memory.

Current theories

Visuo-spatial sketch pad

The visuo-spatial sketchpad is part of Baddeley and Hitch's model of working memory. It is responsible for temporarily storing visual and spatial information, which is currently being used or encoded. It is thought of as a three-dimensional cognitive map, which contains spatial features about where the person is and visual images of the area, or an object being concentrated on. It is used in tasks such as mental image manipulation where a person imagines how a real object would look if it were changed in some way (rotated, flipped, moved, change of colour, etc.). It is also responsible for representing how vivid an image is. A vivid image is one which you have a high potential for retrieving its sensory details. The visuo-spatial sketchpad is responsible for holding onto the visual and spatial qualities of a vivid image in your working memory, and the degree of vividness is directly affected by the limits of the sketchpad.

Eidetic and photographic memories

Eidetic memory is an ability to recall images, sounds, or objects in memory with high precision for a few minutes without using mnemonics. It occurs in a small number of children and generally is not found in adults.

The popular culture concept of photographic memory—where, for example, someone can briefly look at a page of text and then recite it perfectly from memory—is not the same as seeing eidetic images, and photographic memory has never been demonstrated to exist.

Iconic memory

Iconic memory is the visual part of the sensory memory system. Iconic memory is responsible for visual priming, because it works very quickly and unconsciously. Iconic memory decays very quickly, but contains a very vivid image of the surrounding stimuli.

Spatial memory

Spatial memory is a person's knowledge of the space around them, and their whereabouts in it. It also encompasses all memories of areas and places, and how to get to and from them. Spatial memory is distinct from object memory and involves different parts of the brain. Spatial memory involves the dorsal parts of the brain and more specifically the hippocampus. However many times both types of memory are used together, such as when trying to remember where you put a lost object. A classic test of spatial memory is the Corsi block-tapping task, where an instructor taps a series of blocks in a random order and the participant attempts to imitate them. The number of blocks they can tap before performance breaks down on average is called their Corsi span. Spatial memory is always being used whenever a person is moving any part of their body; therefore it is generally more vulnerable to decay than object memory is.

Object memory

Object memory involves processing features of an object or material such as texture, color, size, and orientation. It is processed mainly in the ventral regions of the brain. A few studies have shown that on average most people can recall up to four items each with a set of four different visual qualities. It is a separate system from spatial memory and is not affected by interference from spatial tasks.

Accuracy

Visual memory is not always accurate and can be misled by outside conditions. This can be seen in studies carried out by Elizabeth Loftus and Gary Wells. In one study by Wells, individuals were exposed to misleading information after witnessing an event; they were then tested on their ability to remember details from this event. Their findings included: when given misinformation that contradicts the witnessed event they were less able to recall those details; and whether misinformation was given before or after the witnessed event did not seem to matter. Furthermore, visual memory can be subjected to various memory errors which will affect accuracy.

Visual memory in education

"We do not store and retrieve words based on visual memory." "Our phonological filing system is the basis for word memory/word recognition." -Dr. Kilpatrick (Equipped for Reading Success).

Visual memory, in an academic environment, entails work with pictures, symbols, numbers, letters, and especially words. Students must be able to look at a word, form an image of that word in their minds and be able to recall the appearance of the word later. When teachers introduce a new vocabulary word, generally they write it on the chalkboard, have the children spell it, read it and then use it in a sentence. The word is then erased from the chalkboard. Students with good visual memory will recognize that same word later in their readers or other texts and will be able to recall the appearance of the word to spell it.

Children who have not developed their visual memory skills cannot readily reproduce a sequence of visual stimuli. They frequently experience difficulty in remembering the overall visual appearance of words or the letter sequence of words for reading and spelling.

Factors affecting visual memory

Sleep

Findings surrounding sleep and visual memory have been mixed. Studies have reported performance increases after a bout of sleep compared with the same period of waking. The implications of this are that there is a slow, offline process during sleep that strengthens and enhances the memory trace. Further studies have found that quiet rest has shown the same learning benefits as sleep. Replay has been found to occur during post-training quiet wakefulness as well as sleep. In a recent study where a visual search task was administered quiet rest or sleep is found to be necessary for increasing the amount of associations between configurations and target locations that can be learned within a day. Reactivation in sleep was only observed after extensive training of rodents on familiar tasks. It rapidly dissipates; it also makes up a small proportion of total recorded activity in sleep. It has also been found that there are gender differences between males and females in regards to visual memory and sleep. In a study done testing sleep and memory for pictures it was found that daytime sleep contributed to retention of source memory rather than item memory in females, females did not have recollection or familiarity influenced by daytime sleep, whereas males undergoing daytime sleep had a trend towards increased familiarity. The reasons for this may be linked to different memory traces resulting from different encoding strategies, as well as with different electrophysiological changes during daytime sleep.

Brain damage

Brain damage is another factor that has been found to have an effect on visual memory. Memory impairment affects both novel and familiar experiences. Poor memory after damage to the brain is usually considered to result from information being lost or rendered inaccessible. With such impairment it is assumed that it must be due to the incorrect interpretation of previously encountered information as being novel. In experiments testing rats’ object recognition memory it was found that memory impairment can be the opposite, that there was a tendency to treat novel experiences as familiar. A possible solution for this impairment could be the use of a visual-restriction procedure that reduces interference.

Age

Studies have shown that with aging, in terms of short-term visual memory, viewing time and task complexity affect performance. When there is a delay or when the task is complex recall declines. In a study conducted to measure whether visual memory in older adults with age-related visual decline was caused by memory performance or visual functioning, the following were examined: relationships among age, visual activity, and visual and verbal memory in 89 community dwelling volunteers aged 60–87 years. The findings were that the effect of vision was not specific to visual memory. Therefore, vision was found to be correlated with general memory function in older adults and is not modality specific.

As we age performance in regards to spatial configurations deteriorates. In a task to store and combine two different spatial configurations to form a novel one young people out-performed the elderly. Vision also has an effect on performance. Sighted participants outperformed the visually impaired regardless of testing modality. This suggests that vision tends to shape the general supramodal mechanisms of memory.

Alcohol

Studies have shown that there is an effect of alcohol on visual memory. In a recent study visual working memory and its neutral correlates was assessed in university students who partake in binge drinking, the intermittent consumption of large amounts of alcohol. The findings revealed that there may be binge-drinking related functional alteration in recognition working memory processes. This suggests that impaired prefrontal cortex function may occur at an early age in binge drinkers.

Another study conducted in 2004 examined the level of response to alcohol and brain response during visual working memory. This study looked at the neural correlated of the low level of response to alcohol using functional magnetic resonance imaging during a challenging visual memory task. The results were that young people who report having needed more alcohol to feel the effects showed higher levels of brain response during visual working memory, this suggests that the individual's capacity to adjust to cognitive processing decreases, they are less able to adjust cognitive processing to contextual demands.

Dysfunction of visual memory

Hallucinatory palinopsia, which is a dysfunction of visual memory, is caused by posterior visual pathway cortical lesions and seizures, most commonly in the non-dominant parietal lobe. Focal hyperactivity causes persistent activation of a visual cortex-hippocampal neuronal circuit which encodes an object or scene that is already in visual memory. "All of the hallucinatory palinopsia symptoms occur concomitantly in a patient with one lesion, which supports current evidence that objects, features, and scenes are all units of visual memory, perhaps at different levels of processing. This alludes to neuroanatomical integration in visual memory creation and storage." Studying the excitability alterations associated with palinopsia in migraineurs could provide insight on mechanisms of encoding visual memory.

One common group of people that have visual memory problems are children with reading disabilities. It was often thought that disabilities are caused by failure to perceive the letters of a written word in the right order. However, studies show it is more likely that it is caused by a failure to encode and process the correct order of letters within the word. This means that the child perceives the word just as anyone else would, however their brains do not appear to hold onto the visual characteristics of the word. Although initially it was found that children with reading disabilities had comparable visual memory to those without difficulty, a more specific part of the visual memory system has been found to cause reading disabilities.

These parts are the sustained and transient visual processing systems. The sustained system is responsible for fine detail such as word and letter recognition and is very important in encoding words in their correct order. The transient system is responsible for controlling eye movements, and processing the larger visual environment around us. When these two processes do not work in synchronization this can cause reading disabilities. This has been tested by having children with and without reading disabilities perform on tasks related to the transient systems, where the children with reading disabilities did very poorly. It has also been found in postmortem examinations of the brains of people with reading disabilities that they have fewer neurons and connections in the areas representing the transient visual systems. However, there is debate over whether this is the only reason for reading disabilities, scotopic sensitivity syndrome, deficits in verbal memory and orthographic knowledge are other proposed factors.

Deficits in visual memory can also be caused by disease and/or trauma to the brain. These can lead to the patient losing their spatial memory, and/or their visual memory for specific things. For example, a patient “L.E.” suffered brain damage and her ability to draw from memory was severely diminished, whilst her spatial memory remained normal. Other patients represent the opposite, where memory for colors and shapes is unaffected but spatial memory for previously known places is greatly impaired. These case studies show that these two types of visual memory are located in different parts of the brain and are somewhat unrelated in terms of functioning in daily life.

Introduction to entropy

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Introduct...