Search This Blog

Wednesday, June 18, 2025

Clinical trial

From Wikipedia, the free encyclopedia
A clinical trial participant receives an injection.

Clinical trials are prospective biomedical or behavioral research studies on human participants designed to answer specific questions about biomedical or behavioral interventions, including new treatments (such as novel vaccines, drugs, dietary choices, dietary supplements, and medical devices) and known interventions that warrant further study and comparison. Clinical trials generate data on dosage, safety and efficacy. They are conducted only after they have received health authority/ethics committee approval in the country where approval of the therapy is sought. These authorities are responsible for vetting the risk/benefit ratio of the trial—their approval does not mean the therapy is 'safe' or effective, only that the trial may be conducted.

Depending on product type and development stage, investigators initially enroll volunteers or patients into small pilot studies, and subsequently conduct progressively larger scale comparative studies. Clinical trials can vary in size and cost, and they can involve a single research center or multiple centers, in one country or in multiple countries. Clinical study design aims to ensure the scientific validity and reproducibility of the results.

Costs for clinical trials can range into the billions of dollars per approved drug, and the complete trial process to approval may require 7–15 years. The sponsor may be a governmental organization or a pharmaceutical, biotechnology or medical-device company. Certain functions necessary to the trial, such as monitoring and lab work, may be managed by an outsourced partner, such as a contract research organization or a central laboratory. Only 10 percent of all drugs started in human clinical trials become approved drugs.

Overview

Trials of drugs

Some clinical trials involve healthy subjects with no pre-existing medical conditions. Other clinical trials pertain to people with specific health conditions who are willing to try an experimental treatment. Pilot experiments are conducted to gain insights for design of the clinical trial to follow.

There are two goals to testing medical treatments: to learn whether they work well enough, called "efficacy", or "effectiveness"; and to learn whether they are safe enough, called "safety". Neither is an absolute criterion; both safety and efficacy are evaluated relative to how the treatment is intended to be used, what other treatments are available, and the severity of the disease or condition. The benefits must outweigh the risks. For example, many drugs to treat cancer have severe side effects that would not be acceptable for an over-the-counter pain medication, yet the cancer drugs have been approved since they are used under a physician's care and are used for a life-threatening condition.

In the US the elderly constitute 14% of the population, while they consume over one-third of drugs. People over 55 (or a similar cutoff age) are often excluded from trials because their greater health issues and drug use complicate data interpretation, and because they have different physiological capacity than younger people. Children and people with unrelated medical conditions are also frequently excluded. Pregnant women are often excluded due to potential risks to the fetus.

The sponsor designs the trial in coordination with a panel of expert clinical investigators, including what alternative or existing treatments to compare to the new drug and what type(s) of patients might benefit. If the sponsor cannot obtain enough test subjects at one location investigators at other locations are recruited to join the study.

During the trial, investigators recruit subjects with the predetermined characteristics, administer the treatment(s) and collect data on the subjects' health for a defined time period. Data include measurements such as vital signs, concentration of the study drug in the blood or tissues, changes to symptoms, and whether improvement or worsening of the condition targeted by the study drug occurs. The researchers send the data to the trial sponsor, who then analyzes the pooled data using statistical tests.

Examples of clinical trial goals include assessing the safety and relative effectiveness of a medication or device:

  • On a specific kind of patient
  • At varying dosages
  • For a new indication
  • Evaluation for improved efficacy in treating a condition as compared to the standard therapy for that condition
  • Evaluation of the study drug or device relative to two or more already approved/common interventions for that condition

While most clinical trials test one alternative to the novel intervention, some expand to three or four and may include a placebo.

Except for small, single-location trials, the design and objectives are specified in a document called a clinical trial protocol. The protocol is the trial's "operating manual" and ensures all researchers perform the trial in the same way on similar subjects and that the data is comparable across all subjects.

As a trial is designed to test hypotheses and rigorously monitor and assess outcomes, it can be seen as an application of the scientific method, specifically the experimental step.

The most common clinical trials evaluate new pharmaceutical products, medical devices, biologics, diagnostic assays, psychological therapies, or other interventions. Clinical trials may be required before a national regulatory authority approves marketing of the innovation.

Trials of devices

Similarly to drugs, manufacturers of medical devices in the United States are required to conduct clinical trials for premarket approval. Device trials may compare a new device to an established therapy, or may compare similar devices to each other. An example of the former in the field of vascular surgery is the Open versus Endovascular Repair (OVER trial) for the treatment of abdominal aortic aneurysm, which compared the older open aortic repair technique to the newer endovascular aneurysm repair device. An example of the latter are clinical trials on mechanical devices used in the management of adult female urinary incontinence.

Trials of procedures

Similarly to drugs, medical or surgical procedures may be subjected to clinical trials, such as comparing different surgical approaches in treatment of fibroids for subfertility. However, when clinical trials are unethical or logistically impossible in the surgical setting, case-controlled studies will be replaced.

Patient and public involvement

Besides being participants in a clinical trial, members of the public can be actively collaborate with researchers in designing and conducting clinical research. This is known as patient and public involvement (PPI). Public involvement involves a working partnership between patients, caregivers, people with lived experience, and researchers to shape and influence what is researcher and how. PPI can improve the quality of research and make it more relevant and accessible. People with current or past experience of illness can provide a different perspective than professionals and compliment their knowledge. Through their personal knowledge they can identify research topics that are relevant and important to those living with an illness or using a service. They can also help to make the research more grounded in the needs of the specific communities they are part of. Public contributors can also ensure that the research is presented in plain language that is clear to the wider society and the specific groups it is most relevant for.

History

Development

Edward Jenner vaccinating James Phipps, a boy of eight, on 14 May 1796. Jenner failed to use a control group.

Although early medical experimentation was performed often, the use of a control group to provide an accurate comparison for the demonstration of the intervention's efficacy was generally lacking. For instance, Lady Mary Wortley Montagu, who campaigned for the introduction of inoculation (then called variolation) to prevent smallpox, arranged for seven prisoners who had been sentenced to death to undergo variolation in exchange for their life. Although they survived and did not contract smallpox, there was no control group to assess whether this result was due to the inoculation or some other factor. Similar experiments performed by Edward Jenner over his smallpox vaccine were equally conceptually flawed.

The first proper clinical trial was conducted by the Scottish physician James Lind. The disease scurvy, now known to be caused by a Vitamin C deficiency, would often have terrible effects on the welfare of the crew of long-distance ocean voyages. In 1740, the catastrophic result of Anson's circumnavigation attracted much attention in Europe; out of 1900 men, 1400 had died, most of them allegedly from having contracted scurvy. John Woodall, an English military surgeon of the British East India Company, had recommended the consumption of citrus fruit from the 17th century, but their use did not become widespread.

Lind conducted the first systematic clinical trial in 1747. He included a dietary supplement of an acidic quality in the experiment after two months at sea, when the ship was already afflicted with scurvy. He divided twelve scorbutic sailors into six groups of two. They all received the same diet but, in addition, group one was given a quart of cider daily, group two twenty-five drops of elixir of vitriol (sulfuric acid), group three six spoonfuls of vinegar, group four half a pint of seawater, group five received two oranges and one lemon, and the last group a spicy paste plus a drink of barley water. The treatment of group five stopped after six days when they ran out of fruit, but by then one sailor was fit for duty while the other had almost recovered. Apart from that, only group one also showed some effect of its treatment. Each year, May 20 is celebrated as Clinical Trials Day in honor of Lind's research.

After 1750 the discipline began to take its modern shape. The English doctor John Haygarth demonstrated the importance of a control group for the correct identification of the placebo effect in his celebrated study of the ineffective remedy called Perkin's tractors. Further work in that direction was carried out by the eminent physician Sir William Gull, 1st Baronet in the 1860s.

Frederick Akbar Mahomed (d. 1884), who worked at Guy's Hospital in London, made substantial contributions to the process of clinical trials, where "he separated chronic nephritis with secondary hypertension from what we now term essential hypertension. He also founded the Collective Investigation Record for the British Medical Association; this organization collected data from physicians practicing outside the hospital setting and was the precursor of modern collaborative clinical trials."

Modern trials

Austin Bradford Hill was a pivotal figure in the modern development of clinical trials.

Ideas of Sir Ronald A. Fisher still play a role in clinical trials. While working for the Rothamsted experimental station in the field of agriculture, Fisher developed his Principles of experimental design in the 1920s as an accurate methodology for the proper design of experiments. Among his major ideas include the importance of randomization—the random assignment of individual elements (eg crops or patients) to different groups for the experiment; replication—to reduce uncertainty, measurements should be repeated and experiments replicated to identify sources of variation; blocking—to arrange experimental units into groups of units that are similar to each other, and thus reducing irrelevant sources of variation; use of factorial experiments—efficient at evaluating the effects and possible interactions of several independent factors. Of these, blocking and factorial design are seldom applied in clinical trials, because the experimental units are human subjects and there is typically only one independent intervention: the treatment.

The British Medical Research Council officially recognized the importance of clinical trials from the 1930s. The council established the Therapeutic Trials Committee to advise and assist in the arrangement of properly controlled clinical trials on new products that seem likely on experimental grounds to have value in the treatment of disease.

The first randomised curative trial was carried out at the MRC Tuberculosis Research Unit by Sir Geoffrey Marshall (1887–1982). The trial, carried out between 1946 and 1947, aimed to test the efficacy of the chemical streptomycin for curing pulmonary tuberculosis. The trial was both double-blind and placebo-controlled.

The methodology of clinical trials was further developed by Sir Austin Bradford Hill, who had been involved in the streptomycin trials. From the 1920s, Hill applied statistics to medicine, attending the lectures of renowned mathematician Karl Pearson, among others. He became famous for a landmark study carried out in collaboration with Richard Doll on the correlation between smoking and lung cancer. They carried out a case-control study in 1950, which compared lung cancer patients with matched control and also began a sustained long-term prospective study into the broader issue of smoking and health, which involved studying the smoking habits and health of more than 30,000 doctors over a period of several years. His certificate for election to the Royal Society called him "... the leader in the development in medicine of the precise experimental methods now used nationally and internationally in the evaluation of new therapeutic and prophylactic agents."

International clinical trials day is celebrated on 20 May.

The acronyms used in the titling of clinical trials are often contrived, and have been the subject of derision.

Types

Clinical trials are classified by the research objective created by the investigators.

  • In an observational study, the investigators observe the subjects and measure their outcomes. The researchers do not actively manage the study.
  • In an interventional study, the investigators give the research subjects an experimental drug, surgical procedure, use of a medical device, diagnostic or other intervention to compare the treated subjects with those receiving no treatment or the standard treatment. Then the researchers assess how the subjects' health changes.

Trials are classified by their purpose. After approval for human research is granted to the trial sponsor, the U.S. Food and Drug Administration (FDA) organizes and monitors the results of trials according to type:

  • Prevention trials look for ways to prevent disease in people who have never had the disease or to prevent a disease from returning. These approaches may include drugs, vitamins or other micronutrients, vaccines, or lifestyle changes.
  • Screening trials test for ways to identify certain diseases or health conditions.
  • Diagnostic trials are conducted to find better tests or procedures for diagnosing a particular disease or condition.
  • Treatment trials test experimental drugs, new combinations of drugs, or new approaches to surgery or radiation therapy.
  • Quality of life trials (supportive care trials) evaluate how to improve comfort and quality of care for people with a chronic illness.
  • Genetic trials are conducted to assess the prediction accuracy of genetic disorders making a person more or less likely to develop a disease.
  • Epidemiological trials have the goal of identifying the general causes, patterns or control of diseases in large numbers of people.
  • Compassionate use trials or expanded access trials provide partially tested, unapproved therapeutics to a small number of patients who have no other realistic options. Usually, this involves a disease for which no effective therapy has been approved, or a patient who has already failed all standard treatments and whose health is too compromised to qualify for participation in randomized clinical trials. Usually, case-by-case approval must be granted by both the FDA and the pharmaceutical company for such exceptions.
  • Fixed trials consider existing data only during the trial's design, do not modify the trial after it begins, and do not assess the results until the study is completed.
  • Adaptive clinical trials use existing data to design the trial, and then use interim results to modify the trial as it proceeds. Modifications include dosage, sample size, drug undergoing trial, patient selection criteria and "cocktail" mix. Adaptive trials often employ a Bayesian experimental design to assess the trial's progress. In some cases, trials have become an ongoing process that regularly adds and drops therapies and patient groups as more information is gained. The aim is to more quickly identify drugs that have a therapeutic effect and to zero in on patient populations for whom the drug is appropriate.

Clinical trials are conducted typically in four phases, with each phase using different numbers of subjects and having a different purpose to construct focus on identifying a specific effect.

Phases

Clinical trials involving new drugs are commonly classified into five phases. Each phase of the drug approval process is treated as a separate clinical trial. The drug development process will normally proceed through phases I–IV over many years, frequently involving a decade or longer. If the drug successfully passes through phases I, II, and III, it will usually be approved by the national regulatory authority for use in the general population. Phase IV trials are performed after the newly approved drug, diagnostic or device is marketed, providing assessment about risks, benefits, or best uses.

Phase Aim Notes
Phase 0 Pharmacodynamics and pharmacokinetics in humans Phase 0 trials are optional first-in-human trials. Single subtherapeutic doses of the study drug or treatment are given to a small number of subjects (typically 10 to 15) to gather preliminary data on the agent's pharmacodynamics (what the drug does to the body) and pharmacokinetics (what the body does to the drugs). For a test drug, the trial documents the absorption, distribution, metabolization, and clearance (excretion) of the drug, and the drug's interactions within the body, to confirm that these appear to be as expected.
Phase I Screening for safety Often are first-in-person trials. Testing within a small group of people (typically 20–80) to evaluate safety, determine safe dosage ranges, and identify side effects.
Phase II Establishing the preliminary efficacy of the drug in a "treatment group", usually against a placebo control group Phase II-a is specifically designed to assess dosing requirements (how much drug should be given), while a Phase II-b trial is designed to determine efficacy (100–300 people), assessing how well the drug works at the prescribed dose(s) to establish a therapeutic dose range and monitor for possible side effects.
Phase III Final confirmation of safety and efficacy Testing with large groups of people (typically 1,000–3,000) to confirm drug efficacy, evaluate its effectiveness, monitor side effects, compare it to commonly used treatments, and collect information that will allow it to be used safely.
Phase IV Safety studies during sales Postmarketing studies delineate risks, benefits, and optimal use. As such, they are ongoing during the drug's lifetime of active medical use.

Trial design

A fundamental distinction in evidence-based practice is between observational studies and randomized controlled trials. Types of observational studies in epidemiology, such as the cohort study and the case-control study, provide less compelling evidence than the randomized controlled trial. In observational studies, the investigators retrospectively assess associations between the treatments given to participants and their health status, with potential for considerable errors in design and interpretation.

A randomized controlled trial can provide compelling evidence that the study treatment causes an effect on human health.

Some Phase II and most Phase III drug trials are designed as randomized, double-blind, and placebo-controlled.

  • Randomized: Each study subject is randomly assigned to receive either the study treatment or a placebo.
  • Blind: The subjects involved in the study do not know which study treatment they receive. If the study is double-blind, the researchers also do not know which treatment a subject receives. This intent is to prevent researchers from treating the two groups differently. A form of double-blind study called a "double-dummy" design allows additional insurance against bias. In this kind of study, all patients are given both placebo and active doses in alternating periods.
  • Placebo-controlled: The use of a placebo (fake treatment) allows the researchers to isolate the effect of the study treatment from the placebo effect.

Clinical studies having small numbers of subjects may be "sponsored" by single researchers or a small group of researchers, and are designed to test simple questions or feasibility to expand the research for a more comprehensive randomized controlled trial.

Clinical studies can be "sponsored" (financed and organized) by academic institutions, pharmaceutical companies, government entities and even private groups. Trials are conducted for new drugs, biotechnology, diagnostic assays or medical devices to determine their safety and efficacy prior to being submitted for regulatory review that would determine market approval.

Active control studies

In cases where giving a placebo to a person suffering from a disease may be unethical, "active comparator" (also known as "active control") trials may be conducted instead. In trials with an active control group, subjects are given either the experimental treatment or a previously approved treatment with known effectiveness. In other cases, sponsors may conduct an active comparator trial to establish an efficacy claim relative to the active comparator instead of the placebo in labeling.

Master protocol

A master protocol includes multiple substudies, which may have different objectives and involve coordinated efforts to evaluate one or more medical products in one or more diseases or conditions within the overall study structure. Trials that could develop a master protocol include the umbrella trial (multiple medical products for a single disease), platform trial (multiple products for a single disease entering and leaving the platform), and basket trial (one medical product for multiple diseases or disease subtypes).

Genetic testing enables researchers to group patients according to their genetic profile, deliver drugs based on that profile to that group and compare the results. Multiple companies can participate, each bringing a different drug. The first such approach targets squamous cell cancer, which includes varying genetic disruptions from patient to patient. Amgen, AstraZeneca and Pfizer are involved, the first time they have worked together in a late-stage trial. Patients whose genomic profiles do not match any of the trial drugs receive a drug designed to stimulate the immune system to attack cancer.

Clinical trial protocol

A clinical trial protocol is a document used to define and manage the trial. It is prepared by a panel of experts. All study investigators are expected to strictly observe the protocol.

The protocol describes the scientific rationale, objective(s), design, methodology, statistical considerations and organization of the planned trial. Details of the trial are provided in documents referenced in the protocol, such as an investigator's brochure.

The protocol contains a precise study plan to assure safety and health of the trial subjects and to provide an exact template for trial conduct by investigators. This allows data to be combined across all investigators/sites. The protocol also informs the study administrators (often a contract research organization).

The format and content of clinical trial protocols sponsored by pharmaceutical, biotechnology or medical device companies in the United States, European Union, or Japan have been standardized to follow Good Clinical Practice guidance issued by the International Conference on Harmonisation (ICH). Regulatory authorities in Canada, China, South Korea, and the UK also follow ICH guidelines. Journals such as Trials, encourage investigators to publish their protocols.

Design features

Example of informed consent document from the PARAMOUNT trial

Clinical trials recruit study subjects to sign a document representing their "informed consent". The document includes details such as its purpose, duration, required procedures, risks, potential benefits, key contacts and institutional requirements. The participant then decides whether to sign the document. The document is not a contract, as the participant can withdraw at any time without penalty.

Informed consent is a legal process in which a recruit is instructed about key facts before deciding whether to participate. Researchers explain the details of the study in terms the subject can understand. The information is presented in the subject's native language. Generally, children cannot autonomously provide informed consent, but depending on their age and other factors, may be required to provide informed assent.

Statistical power

In any clinical trial, the number of subjects, also called the sample size, has a large impact on the ability to reliably detect and measure the effects of the intervention. This ability is described as its "power", which must be calculated before initiating a study to figure out if the study is worth its costs. In general, a larger sample size increases the statistical power, also the cost.

The statistical power estimates the ability of a trial to detect a difference of a particular size (or larger) between the treatment and control groups. For example, a trial of a lipid-lowering drug versus placebo with 100 patients in each group might have a power of 0.90 to detect a difference between placebo and trial groups receiving dosage of 10 mg/dL or more, but only 0.70 to detect a difference of 6 mg/dL.

Placebo groups

Merely giving a treatment can have nonspecific effects. These are controlled for by the inclusion of patients who receive only a placebo. Subjects are assigned randomly without informing them to which group they belonged. Many trials are doubled-blinded so that researchers do not know to which group a subject is assigned.

Assigning a subject to a placebo group can pose an ethical problem if it violates his or her right to receive the best available treatment. The Declaration of Helsinki provides guidelines on this issue.

Duration

Timeline of various approval tracks and research phases in the US

Clinical trials are only a small part of the research that goes into developing a new treatment. Potential drugs, for example, first have to be discovered, purified, characterized, and tested in labs (in cell and animal studies) before ever undergoing clinical trials. In all, about 1,000 potential drugs are tested before just one reaches the point of being tested in a clinical trial. For example, a new cancer drug has, on average, six years of research behind it before it even makes it to clinical trials. But the major holdup in making new cancer drugs available is the time it takes to complete clinical trials themselves. On average, about eight years pass from the time a cancer drug enters clinical trials until it receives approval from regulatory agencies for sale to the public. Drugs for other diseases have similar timelines.

Some reasons a clinical trial might last several years:

  • For chronic conditions such as cancer, it takes months, if not years, to see if a cancer treatment has an effect on a patient.
  • For drugs that are not expected to have a strong effect (meaning a large number of patients must be recruited to observe 'any' effect), recruiting enough patients to test the drug's effectiveness (i.e., getting statistical power) can take several years.
  • Only certain people who have the target disease condition are eligible to take part in each clinical trial. Researchers who treat these particular patients must participate in the trial. Then they must identify the desirable patients and obtain consent from them or their families to take part in the trial.

A clinical trial might also include an extended post-study follow-up period from months to years for people who have participated in the trial, a so-called "extension phase", which aims to identify long-term impact of the treatment.

The biggest barrier to completing studies is the shortage of people who take part. All drug and many device trials target a subset of the population, meaning not everyone can participate. Some drug trials require patients to have unusual combinations of disease characteristics. It is a challenge to find the appropriate patients and obtain their consent, especially when they may receive no direct benefit (because they are not paid, the study drug is not yet proven to work, or the patient may receive a placebo). In the case of cancer patients, fewer than 5% of adults with cancer will participate in drug trials. According to the Pharmaceutical Research and Manufacturers of America (PhRMA), about 400 cancer medicines were being tested in clinical trials in 2005. Not all of these will prove to be useful, but those that are may be delayed in getting approved because the number of participants is so low.

For clinical trials involving potential for seasonal influences (such as airborne allergies, seasonal affective disorder, influenza, and skin diseases), the study may be done during a limited part of the year (such as spring for pollen allergies), when the drug can be tested.

Clinical trials that do not involve a new drug usually have a much shorter duration. (Exceptions are epidemiological studies, such as the Nurses' Health Study).

Administration

Clinical trials designed by a local investigator, and (in the US) federally funded clinical trials, are almost always administered by the researcher who designed the study and applied for the grant. Small-scale device studies may be administered by the sponsoring company. Clinical trials of new drugs are usually administered by a contract research organization (CRO) hired by the sponsoring company. The sponsor provides the drug and medical oversight. A CRO is contracted to perform all the administrative work on a clinical trial. For Phases II–IV the CRO recruits participating researchers, trains them, provides them with supplies, coordinates study administration and data collection, sets up meetings, monitors the sites for compliance with the clinical protocol, and ensures the sponsor receives data from every site. Specialist site management organizations can also be hired to coordinate with the CRO to ensure rapid IRB/IEC approval and faster site initiation and patient recruitment. Phase I clinical trials of new medicines are often conducted in a specialist clinical trial clinic, with dedicated pharmacologists, where the subjects can be observed by full-time staff. These clinics are often run by a CRO which specialises in these studies.

At a participating site, one or more research assistants (often nurses) do most of the work in conducting the clinical trial. The research assistant's job can include some or all of the following: providing the local institutional review board (IRB) with the documentation necessary to obtain its permission to conduct the study, assisting with study start-up, identifying eligible patients, obtaining consent from them or their families, administering study treatment(s), collecting and statistically analyzing data, maintaining and updating data files during followup, and communicating with the IRB, as well as the sponsor and CRO.

Quality

In the context of a clinical trial, quality typically refers to the absence of errors which can impact decision making, both during the conduct of the trial and in use of the trial results.

Marketing

An Interactional Justice Model may be used to test the effects of willingness to talk with a doctor about clinical trial enrollment.[66] Results found that potential clinical trial candidates were less likely to enroll in clinical trials if the patient is more willing to talk with their doctor. The reasoning behind this discovery may be patients are happy with their current care. Another reason for the negative relationship between perceived fairness and clinical trial enrollment is the lack of independence from the care provider. Results found that there is a positive relationship between a lack of willingness to talk with their doctor and clinical trial enrollment. Lack of willingness to talk about clinical trials with current care providers may be due to patients' independence from the doctor. Patients who are less likely to talk about clinical trials are more willing to use other sources of information to gain a better insight of alternative treatments. Clinical trial enrollment should be motivated to utilize websites and television advertising to inform the public about clinical trial enrollment.

Information technology

The last decade has seen a proliferation of information technology use in the planning and conduct of clinical trials. Clinical trial management systems are often used by research sponsors or CROs to help plan and manage the operational aspects of a clinical trial, particularly with respect to investigational sites. Advanced analytics for identifying researchers and research sites with expertise in a given area utilize public and private information about ongoing research. Web-based electronic data capture (EDC) and clinical data management systems are used in a majority of clinical trials to collect case report data from sites, manage its quality and prepare it for analysis. Interactive voice response systems are used by sites to register the enrollment of patients using a phone and to allocate patients to a particular treatment arm (although phones are being increasingly replaced with web-based (IWRS) tools which are sometimes part of the EDC system). While patient-reported outcome were often paper based in the past, measurements are increasingly being collected using web portals or hand-held ePRO (or eDiary) devices, sometimes wireless. Statistical software is used to analyze the collected data and prepare them for regulatory submission. Access to many of these applications are increasingly aggregated in web-based clinical trial portals. In 2011, the FDA approved a Phase I trial that used telemonitoring, also known as remote patient monitoring, to collect biometric data in patients' homes and transmit it electronically to the trial database. This technology provides many more data points and is far more convenient for patients, because they have fewer visits to trial sites. As noted below, decentralized clinical trials are those that do not require patients' physical presence at a site, and instead rely largely on digital health data collection, digital informed consent processes, and so on.

Analysis

A clinical trial produces data that could reveal quantitative differences between two or more interventions; statistical analyses are used to determine whether such differences are true, result from chance, or are the same as no treatment (placebo). Data from a clinical trial accumulate gradually over the trial duration, extending from months to years. Accordingly, results for participants recruited early in the study become available for analysis while subjects are still being assigned to treatment groups in the trial. Early analysis may allow the emerging evidence to assist decisions about whether to stop the study, or to reassign participants to the more successful segment of the trial. Investigators may also want to stop a trial when data analysis shows no treatment effect.

Ethical aspects

Clinical trials are closely supervised by appropriate regulatory authorities. All studies involving a medical or therapeutic intervention on patients must be approved by a supervising ethics committee before permission is granted to run the trial. The local ethics committee has discretion on how it will supervise noninterventional studies (observational studies or those using already collected data). In the US, this body is called the Institutional Review Board (IRB); in the EU, they are called Ethics committees. Most IRBs are located at the local investigator's hospital or institution, but some sponsors allow the use of a central (independent/for profit) IRB for investigators who work at smaller institutions.

To be ethical, researchers must obtain the full and informed consent of participating human subjects. (One of the IRB's main functions is to ensure potential patients are adequately informed about the clinical trial.) If the patient is unable to consent for him/herself, researchers can seek consent from the patient's legally authorized representative. In addition, the clinical trial participants must be made aware that they can withdraw from the clinical trial at any time without any adverse action taken against them. In California, the state has prioritized the individuals who can serve as the legally authorized representative.

In some US locations, the local IRB must certify researchers and their staff before they can conduct clinical trials. They must understand the federal patient privacy (HIPAA) law and good clinical practice. The International Conference of Harmonisation Guidelines for Good Clinical Practice is a set of standards used internationally for the conduct of clinical trials. The guidelines aim to ensure the "rights, safety and well being of trial subjects are protected".

The notion of informed consent of participating human subjects exists in many countries but its precise definition may still vary.

Informed consent is clearly a 'necessary' condition for ethical conduct but does not 'ensure' ethical conduct. In compassionate use trials the latter becomes a particularly difficult problem. The final objective is to serve the community of patients or future patients in a best-possible and most responsible way. See also Expanded access. However, it may be hard to turn this objective into a well-defined, quantified, objective function. In some cases this can be done, however, for instance, for questions of when to stop sequential treatments (see Odds algorithm), and then quantified methods may play an important role.

Additional ethical concerns are present when conducting clinical trials on children (pediatrics), and in emergency or epidemic situations.

Ethically balancing the rights of multiple stakeholders may be difficult. For example, when drug trials fail, the sponsors may have a duty to tell current and potential investors immediately, which means both the research staff and the enrolled participants may first hear about the end of a trial through public business news.

Conflicts of interest and unfavorable studies

In response to specific cases in which unfavorable data from pharmaceutical company-sponsored research were not published, the Pharmaceutical Research and Manufacturers of America published new guidelines urging companies to report all findings and limit the financial involvement in drug companies by researchers. The US Congress signed into law a bill which requires Phase II and Phase III clinical trials to be registered by the sponsor on the clinicaltrials.gov website compiled by the National Institutes of Health.

Drug researchers not directly employed by pharmaceutical companies often seek grants from manufacturers, and manufacturers often look to academic researchers to conduct studies within networks of universities and their hospitals, e.g., for translational cancer research. Similarly, competition for tenured academic positions, government grants and prestige create conflicts of interest among academic scientists. According to one study, approximately 75% of articles retracted for misconduct-related reasons have no declared industry financial support. Seeding trials are particularly controversial.

In the United States, all clinical trials submitted to the FDA as part of a drug approval process are independently assessed by clinical experts within the Food and Drug Administration, including inspections of primary data collection at selected clinical trial sites.

In 2001, the editors of 12 major journals issued a joint editorial, published in each journal, on the control over clinical trials exerted by sponsors, particularly targeting the use of contracts which allow sponsors to review the studies prior to publication and withhold publication. They strengthened editorial restrictions to counter the effect. The editorial noted that contract research organizations had, by 2000, received 60% of the grants from pharmaceutical companies in the US. Researchers may be restricted from contributing to the trial design, accessing the raw data, and interpreting the results.

Despite explicit recommendations  by stakeholders of measures to improve the standards of industry-sponsored medical research, in 2013, Tohen warned of the persistence of a gap in the credibility of conclusions arising from industry-funded clinical trials, and called for ensuring strict adherence to ethical standards in industrial collaborations with academia, in order to avoid further erosion of the public's trust. Issues referred for attention in this respect include potential observation bias, duration of the observation time for maintenance studies, the selection of the patient populations, factors that affect placebo response, and funding sources.

During public health crisis

Conducting clinical trials of vaccines during epidemics and pandemics is subject to ethical concerns. For diseases with high mortality rates like Ebola, assigning individuals to a placebo or control group can be viewed as a death sentence. In response to ethical concerns regarding clinical research during epidemics, the National Academy of Medicine authored a report identifying seven ethical and scientific considerations. These considerations are:

  • Scientific value
  • Social value
  • Respect for persons
  • Community engagement
  • Concern for participant welfare and interests
  • A balance towards benefit over risks
  • Post-trial access to tested therapies that had been withheld during the trial

Pregnant women and children

Pregnant women and children are typically excluded from clinical trials as vulnerable populations, though the data to support excluding them is not robust. By excluding them from clinical trials, information about the safety and effectiveness of therapies for these populations is often lacking. During the early history of the HIV/AIDS epidemic, a scientist noted that by excluding these groups from potentially life-saving treatment, they were being "protected to death". Projects such as Research Ethics for Vaccines, Epidemics, and New Technologies (PREVENT) have advocated for the ethical inclusion of pregnant women in vaccine trials. Inclusion of children in clinical trials has additional moral considerations, as children lack decision-making autonomy. Trials in the past had been criticized for using hospitalized children or orphans; these ethical concerns effectively stopped future research. In efforts to maintain effective pediatric care, several European countries and the US have policies to entice or compel pharmaceutical companies to conduct pediatric trials. International guidance recommends ethical pediatric trials by limiting harm, considering varied risks, and taking into account the complexities of pediatric care.

Safety

Responsibility for the safety of the subjects in a clinical trial is shared between the sponsor, the local site investigators (if different from the sponsor), the various IRBs that supervise the study, and (in some cases, if the study involves a marketable drug or device), the regulatory agency for the country where the drug or device will be sold.

A systematic concurrent safety review is frequently employed to assure research participant safety. The conduct and on-going review is designed to be proportional to the risk of the trial. Typically this role is filled by a Data and Safety Committee, an externally appointed Medical Safety Monitor, an Independent Safety Officer, or for small or low-risk studies the principal investigator.

For safety reasons, many clinical trials of drugs are designed to exclude women of childbearing age, pregnant women, or women who become pregnant during the study. In some cases, the male partners of these women are also excluded or required to take birth control measures.

Throughout the clinical trial, the sponsor is responsible for accurately informing the local site investigators of the true historical safety record of the drug, device or other medical treatments to be tested, and of any potential interactions of the study treatment(s) with already approved treatments. This allows the local investigators to make an informed judgment on whether to participate in the study or not. The sponsor is also responsible for monitoring the results of the study as they come in from the various sites as the trial proceeds. In larger clinical trials, a sponsor will use the services of a data monitoring committee (DMC, known in the US as a data safety monitoring board). This independent group of clinicians and statisticians meets periodically to review the unblinded data the sponsor has received so far. The DMC has the power to recommend termination of the study based on their review, for example if the study treatment is causing more deaths than the standard treatment, or seems to be causing unexpected and study-related serious adverse events. The sponsor is responsible for collecting adverse event reports from all site investigators in the study, and for informing all the investigators of the sponsor's judgment as to whether these adverse events were related or not related to the study treatment.

The sponsor and the local site investigators are jointly responsible for writing a site-specific informed consent that accurately informs the potential subjects of the true risks and potential benefits of participating in the study, while at the same time presenting the material as briefly as possible and in ordinary language. FDA regulations state that participating in clinical trials is voluntary, with the subject having the right not to participate or to end participation at any time.

Local site investigators

The ethical principle of primum non-nocere ("first, do no harm") guides the trial, and if an investigator believes the study treatment may be harming subjects in the study, the investigator can stop participating at any time. On the other hand, investigators often have a financial interest in recruiting subjects, and could act unethically to obtain and maintain their participation.

The local investigators are responsible for conducting the study according to the study protocol, and supervising the study staff throughout the duration of the study. The local investigator or his/her study staff are also responsible for ensuring the potential subjects in the study understand the risks and potential benefits of participating in the study. In other words, they (or their legally authorized representatives) must give truly informed consent.

Local investigators are responsible for reviewing all adverse event reports sent by the sponsor. These adverse event reports contain the opinions of both the investigator (at the site where the adverse event occurred) and the sponsor, regarding the relationship of the adverse event to the study treatments. Local investigators also are responsible for making an independent judgment of these reports, and promptly informing the local IRB of all serious and study treatment-related adverse events.

When a local investigator is the sponsor, there may not be formal adverse event reports, but study staff at all locations are responsible for informing the coordinating investigator of anything unexpected. The local investigator is responsible for being truthful to the local IRB in all communications relating to the study.

Institutional review boards (IRBs)

Approval by an Institutional Review Board (IRB), or Independent Ethics Committee (IEC), is necessary before all but the most informal research can begin. In commercial clinical trials, the study protocol is not approved by an IRB before the sponsor recruits sites to conduct the trial. However, the study protocol and procedures have been tailored to fit generic IRB submission requirements. In this case, and where there is no independent sponsor, each local site investigator submits the study protocol, the consent(s), the data collection forms, and supporting documentation to the local IRB. Universities and most hospitals have in-house IRBs. Other researchers (such as in walk-in clinics) use independent IRBs.

The IRB scrutinizes the study both for medical safety and for protection of the patients involved in the study, before it allows the researcher to begin the study. It may require changes in study procedures or in the explanations given to the patient. A required yearly "continuing review" report from the investigator updates the IRB on the progress of the study and any new safety information related to the study.

Regulatory agencies

In the US, the FDA can audit the files of local site investigators after they have finished participating in a study, to see if they were correctly following study procedures. This audit may be random, or for cause (because the investigator is suspected of fraudulent data). Avoiding an audit is an incentive for investigators to follow study procedures. A 'covered clinical study' refers to a trial submitted to the FDA as part of a marketing application (for example, as part of an NDA or 510(k)), about which the FDA may require disclosure of financial interest of the clinical investigator in the outcome of the study. For example, the applicant must disclose whether an investigator owns equity in the sponsor, or owns proprietary interest in the product under investigation. The FDA defines a covered study as "... any study of a drug, biological product or device in humans submitted in a marketing application or reclassification petition that the applicant or FDA relies on to establish that the product is effective (including studies that show equivalence to an effective product) or any study in which a single investigator makes a significant contribution to the demonstration of safety."

Alternatively, many American pharmaceutical companies have moved some clinical trials overseas. Benefits of conducting trials abroad include lower costs (in some countries) and the ability to run larger trials in shorter timeframes, whereas a potential disadvantage exists in lower-quality trial management. Different countries have different regulatory requirements and enforcement abilities. An estimated 40% of all clinical trials now take place in Asia, Eastern Europe, and Central and South America. "There is no compulsory registration system for clinical trials in these countries and many do not follow European directives in their operations", says Jacob Sijtsma of the Netherlands-based WEMOS, an advocacy health organisation tracking clinical trials in developing countries.

Beginning in the 1980s, harmonization of clinical trial protocols was shown as feasible across countries of the European Union. At the same time, coordination between Europe, Japan and the United States led to a joint regulatory-industry initiative on international harmonization named after 1990 as the International Conference on Harmonisation of Technical Requirements for Registration of Pharmaceuticals for Human Use (ICH). Currently, most clinical trial programs follow ICH guidelines, aimed at "ensuring that good quality, safe and effective medicines are developed and registered in the most efficient and cost-effective manner. These activities are pursued in the interest of the consumer and public health, to prevent unnecessary duplication of clinical trials in humans and to minimize the use of animal testing without compromising the regulatory obligations of safety and effectiveness."

Aggregation of safety data during clinical development

Aggregating safety data across clinical trials during drug development is important because trials are generally designed to focus on determining how well the drug works. The safety data collected and aggregated across multiple trials as the drug is developed allows the sponsor, investigators and regulatory agencies to monitor the aggregate safety profile of experimental medicines as they are developed. The value of assessing aggregate safety data is: a) decisions based on aggregate safety assessment during development of the medicine can be made throughout the medicine's development and b) it sets up the sponsor and regulators well for assessing the medicine's safety after the drug is approved.

Economics

Clinical trial costs vary depending on trial phase, type of trial, and disease studied. A study of clinical trials conducted in the United States from 2004 to 2012 found the average cost of Phase I trials to be between $1.4 million and $6.6 million, depending on the type of disease. Phase II trials ranged from $7 million to $20 million, and Phase III trials from $11 million to $53 million.

The cost of a study depends on many factors, especially the number of sites conducting the study, the number of patients involved, and whether the study treatment is already approved for medical use.

The expenses incurred by a pharmaceutical company in administering a Phase III or IV clinical trial may include, among others:

  • production of the drug(s) or device(s) being evaluated
  • staff salaries for the designers and administrators of the trial
  • payments to the contract research organization, the site management organization (if used) and any outside consultants
  • payments to local researchers and their staff for their time and effort in recruiting test subjects and collecting data for the sponsor
  • the cost of study materials and the charges incurred to ship them
  • communication with the local researchers, including on-site monitoring by the CRO before and (in some cases) multiple times during the study
  • one or more investigator training meetings
  • expense incurred by the local researchers, such as pharmacy fees, IRB fees and postage
  • any payments to subjects enrolled in the trial
  • the expense of treating a test subject who develops a medical condition caused by the study drug

These expenses are incurred over several years.

In the US, sponsors may receive a 50 percent tax credit for clinical trials conducted on drugs being developed for the treatment of orphan diseases. National health agencies, such as the US National Institutes of Health, offer grants to investigators who design clinical trials that attempt to answer research questions of interest to the agency. In these cases, the investigator who writes the grant and administers the study acts as the sponsor, and coordinates data collection from any other sites. These other sites may or may not be paid for participating in the study, depending on the amount of the grant and the amount of effort expected from them. Using internet resources can, in some cases, reduce the economic burden.

Investigators

Investigators are often compensated for their work in clinical trials. These amounts can be small, just covering a partial salary for research assistants and the cost of any supplies (usually the case with national health agency studies), or be substantial and include "overhead" that allows the investigator to pay the research staff during times between clinical trials.

Subjects

Participants in Phase I drug trials do not gain any direct health benefit from taking part. They are generally paid a fee for their time, with payments regulated and not related to any risk involved. Motivations of healthy volunteers is not limited to financial reward and may include other motivations such as contributing to science and others. In later phase trials, subjects may not be paid to ensure their motivation for participating with potential for a health benefit or contributing to medical knowledge. Small payments may be made for study-related expenses such as travel or as compensation for their time in providing follow-up information about their health after the trial treatment ends.

Participant recruitment and participation

Newspaper advertisements seeking patients and healthy volunteers to participate in clinical trials

Phase 0 and Phase I drug trials seek healthy volunteers. Most other clinical trials seek patients who have a specific disease or medical condition. The diversity observed in society should be reflected in clinical trials through the appropriate inclusion of ethnic minority populations. Patient recruitment or participant recruitment plays a significant role in the activities and responsibilities of sites conducting clinical trials.

All volunteers being considered for a trial are required to undertake a medical screening. Requirements differ according to the trial needs, but typically volunteers would be screened in a medical laboratory for:

  • Measurement of the electrical activity of the heart (ECG)
  • Measurement of blood pressure, heart rate, and body temperature
  • Blood sampling
  • Urine sampling
  • Weight and height measurement
  • Drug abuse testing
  • Pregnancy testing

It has been observed that participants in clinical trials are disproportionately white. Often, minorities are not informed about clinical trials. One recent systematic review of the literature found that race/ethnicity as well as sex were not well-represented nor at times even tracked as participants in a large number of clinical trials of hearing loss management in adults. This may reduce the validity of findings in respect of non-white patients by not adequately representing the larger populations.

Locating trials

Depending on the kind of participants required, sponsors of clinical trials, or contract research organizations working on their behalf, try to find sites with qualified personnel as well as access to patients who could participate in the trial. Working with those sites, they may use various recruitment strategies, including patient databases, newspaper and radio advertisements, flyers, posters in places the patients might go (such as doctor's offices), and personal recruitment of patients by investigators.

Volunteers with specific conditions or diseases have additional online resources to help them locate clinical trials. For example, the Fox Trial Finder connects Parkinson's disease trials around the world to volunteers who have a specific set of criteria such as location, age, and symptoms. Other disease-specific services exist for volunteers to find trials related to their condition. Volunteers may search directly on ClinicalTrials.gov to locate trials using a registry run by the U.S. National Institutes of Health and National Library of Medicine. There also is software that allows clinicians to find trial options for an individual patient based on data such as genomic data.

Research

Eli Lilly booths
Eli Lilly pharmaceutical company recruiting participants at the Indiana State Fair

The risk information seeking and processing (RISP) model analyzes social implications that affect attitudes and decision making pertaining to clinical trials. People who hold a higher stake or interest in the treatment provided in a clinical trial showed a greater likelihood of seeking information about clinical trials. Cancer patients reported more optimistic attitudes towards clinical trials than the general population. Having a more optimistic outlook on clinical trials also leads to greater likelihood of enrolling.

Matching

Matching involves a systematic comparison of a patient's clinical and demographic information against the eligibility criteria of various trials. Methods include:

  • Manual: Healthcare providers or clinical trial coordinators manually review patient records and available trial criteria to identify potential matches. This might also include manually searching in clinical trial databases.
  • Electronic health records (EHR). Some systems integrate with EHRs to automatically flag patients that may be eligible for trials based on their medical data. These systems may leverage machine learning, artificial intelligence or precision medicine methods to more effectively match patients to trials. These methods are faced with the challenge of overcoming the limitations of EHR records such as omissions and logging errors.
  • Direct-to-patient services: Resources are specialized to support patients in finding clinical trials through online platforms, hotlines, and personalized support.

Decentralized trials

Although trials are commonly conducted at major medical centers, some participants are excluded due to the distance and expenses required for travel, leading to hardship, disadvantage, and inequity for participants, especially those in rural and underserved communities. Therefore, the concept of a "decentralized clinical trial" that minimizes or eliminates the need for patients to travel to sites, is now more widespread, a capability improved by telehealth and wearable technologies.

Factor analysis

From Wikipedia, the free encyclopedia
https://en.wikipedia.org/wiki/Factor_analysis 

Factor analysis
is a statistical method used to describe variability among observed, correlated variables in terms of a potentially lower number of unobserved variables called factors. For example, it is possible that variations in six observed variables mainly reflect the variations in two unobserved (underlying) variables. Factor analysis searches for such joint variations in response to unobserved latent variables. The observed variables are modelled as linear combinations of the potential factors plus "error" terms, hence factor analysis can be thought of as a special case of errors-in-variables models.

The correlation between a variable and a given factor, called the variable's factor loading, indicates the extent to which the two are related.

A common rationale behind factor analytic methods is that the information gained about the interdependencies between observed variables can be used later to reduce the set of variables in a dataset. Factor analysis is commonly used in psychometrics, personality psychology, biology, marketing, product management, operations research, finance, and machine learning. It may help to deal with data sets where there are large numbers of observed variables that are thought to reflect a smaller number of underlying/latent variables. It is one of the most commonly used inter-dependency techniques and is used when the relevant set of variables shows a systematic inter-dependence and the objective is to find out the latent factors that create a commonality.

Statistical model

Definition

The model attempts to explain a set of observations in each of individuals with a set of common factors () where there are fewer factors per unit than observations per unit (). Each individual has of their own common factors, and these are related to the observations via the factor loading matrix (), for a single observation, according to

where

  • is the value of the th observation of the th individual,
  • is the observation mean for the th observation,
  • is the loading for the th observation of the th factor,
  • is the value of the th factor of the th individual, and
  • is the th unobserved stochastic error term with mean zero and finite variance.

In matrix notation

where observation matrix , loading matrix , factor matrix , error term matrix and mean matrix whereby the th element is simply .

Also we will impose the following assumptions on :

  1. and are independent.
  2. ; where is Expectation
  3. where is the covariance matrix, to make sure that the factors are uncorrelated, and is the identity matrix.

Suppose . Then

and therefore, from conditions 1 and 2 imposed on above, and , giving

or, setting ,

For any orthogonal matrix , if we set and , the criteria for being factors and factor loadings still hold. Hence a set of factors and factor loadings is unique only up to an orthogonal transformation.

Example

Suppose a psychologist has the hypothesis that there are two kinds of intelligence, "verbal intelligence" and "mathematical intelligence", neither of which is directly observed. Evidence for the hypothesis is sought in the examination scores from each of 10 different academic fields of 1000 students. If each student is chosen randomly from a large population, then each student's 10 scores are random variables. The psychologist's hypothesis may say that for each of the 10 academic fields, the score averaged over the group of all students who share some common pair of values for verbal and mathematical "intelligences" is some constant times their level of verbal intelligence plus another constant times their level of mathematical intelligence, i.e., it is a linear combination of those two "factors". The numbers for a particular subject, by which the two kinds of intelligence are multiplied to obtain the expected score, are posited by the hypothesis to be the same for all intelligence level pairs, and are called "factor loading" for this subject.  For example, the hypothesis may hold that the predicted average student's aptitude in the field of astronomy is

{10 × the student's verbal intelligence} + {6 × the student's mathematical intelligence}.

The numbers 10 and 6 are the factor loadings associated with astronomy. Other academic subjects may have different factor loadings.

Two students assumed to have identical degrees of verbal and mathematical intelligence may have different measured aptitudes in astronomy because individual aptitudes differ from average aptitudes (predicted above) and because of measurement error itself. Such differences make up what is collectively called the "error" — a statistical term that means the amount by which an individual, as measured, differs from what is average for or predicted by his or her levels of intelligence (see errors and residuals in statistics).

The observable data that go into factor analysis would be 10 scores of each of the 1000 students, a total of 10,000 numbers. The factor loadings and levels of the two kinds of intelligence of each student must be inferred from the data.

Mathematical model of the same example

In the following, matrices will be indicated by indexed variables. "Academic Subject" indices will be indicated using letters , and , with values running from to which is equal to in the above example. "Factor" indices will be indicated using letters , and , with values running from to which is equal to in the above example. "Instance" or "sample" indices will be indicated using letters , and , with values running from to . In the example above, if a sample of students participated in the exams, the th student's score for the th exam is given by . The purpose of factor analysis is to characterize the correlations between the variables of which the are a particular instance, or set of observations. In order for the variables to be on equal footing, they are normalized into standard scores :

where the sample mean is:

and the sample variance is given by:

The factor analysis model for this particular sample is then:

or, more succinctly:

where

  • is the th student's "verbal intelligence",
  • is the th student's "mathematical intelligence",
  • are the factor loadings for the th subject, for .

In matrix notation, we have

Observe that by doubling the scale on which "verbal intelligence"—the first component in each column of —is measured, and simultaneously halving the factor loadings for verbal intelligence makes no difference to the model. Thus, no generality is lost by assuming that the standard deviation of the factors for verbal intelligence is . Likewise for mathematical intelligence. Moreover, for similar reasons, no generality is lost by assuming the two factors are uncorrelated with each other. In other words:

where is the Kronecker delta ( when and when ). The errors are assumed to be independent of the factors:

Since any rotation of a solution is also a solution, this makes interpreting the factors difficult. See disadvantages below. In this particular example, if we do not know beforehand that the two types of intelligence are uncorrelated, then we cannot interpret the two factors as the two different types of intelligence. Even if they are uncorrelated, we cannot tell which factor corresponds to verbal intelligence and which corresponds to mathematical intelligence without an outside argument.

The values of the loadings , the averages , and the variances of the "errors" must be estimated given the observed data and (the assumption about the levels of the factors is fixed for a given ). The "fundamental theorem" may be derived from the above conditions:

The term on the left is the -term of the correlation matrix (a matrix derived as the product of the matrix of standardized observations with its transpose) of the observed data, and its diagonal elements will be s. The second term on the right will be a diagonal matrix with terms less than unity. The first term on the right is the "reduced correlation matrix" and will be equal to the correlation matrix except for its diagonal values which will be less than unity. These diagonal elements of the reduced correlation matrix are called "communalities" (which represent the fraction of the variance in the observed variable that is accounted for by the factors):

The sample data will not exactly obey the fundamental equation given above due to sampling errors, inadequacy of the model, etc. The goal of any analysis of the above model is to find the factors and loadings which give a "best fit" to the data. In factor analysis, the best fit is defined as the minimum of the mean square error in the off-diagonal residuals of the correlation matrix:

This is equivalent to minimizing the off-diagonal components of the error covariance which, in the model equations have expected values of zero. This is to be contrasted with principal component analysis which seeks to minimize the mean square error of all residuals. Before the advent of high-speed computers, considerable effort was devoted to finding approximate solutions to the problem, particularly in estimating the communalities by other means, which then simplifies the problem considerably by yielding a known reduced correlation matrix. This was then used to estimate the factors and the loadings. With the advent of high-speed computers, the minimization problem can be solved iteratively with adequate speed, and the communalities are calculated in the process, rather than being needed beforehand. The MinRes algorithm is particularly suited to this problem, but is hardly the only iterative means of finding a solution.

If the solution factors are allowed to be correlated (as in 'oblimin' rotation, for example), then the corresponding mathematical model uses skew coordinates rather than orthogonal coordinates.

Geometric interpretation

Geometric interpretation of Factor Analysis parameters for 3 respondents to question "a". The "answer" is represented by the unit vector , which is projected onto a plane defined by two orthonormal vectors and . The projection vector is and the error is perpendicular to the plane, so that . The projection vector may be represented in terms of the factor vectors as . The square of the length of the projection vector is the communality: . If another data vector were plotted, the cosine of the angle between and would be  : the -entry in the correlation matrix. (Adapted from Harman Fig. 4.3)

The parameters and variables of factor analysis can be given a geometrical interpretation. The data (), the factors () and the errors () can be viewed as vectors in an -dimensional Euclidean space (sample space), represented as , and respectively. Since the data are standardized, the data vectors are of unit length (). The factor vectors define a -dimensional linear subspace (i.e. a hyperplane) in this space, upon which the data vectors are projected orthogonally. This follows from the model equation

and the independence of the factors and the errors: . In the above example, the hyperplane is just a 2-dimensional plane defined by the two factor vectors. The projection of the data vectors onto the hyperplane is given by

and the errors are vectors from that projected point to the data point and are perpendicular to the hyperplane. The goal of factor analysis is to find a hyperplane which is a "best fit" to the data in some sense, so it doesn't matter how the factor vectors which define this hyperplane are chosen, as long as they are independent and lie in the hyperplane. We are free to specify them as both orthogonal and normal () with no loss of generality. After a suitable set of factors are found, they may also be arbitrarily rotated within the hyperplane, so that any rotation of the factor vectors will define the same hyperplane, and also be a solution. As a result, in the above example, in which the fitting hyperplane is two dimensional, if we do not know beforehand that the two types of intelligence are uncorrelated, then we cannot interpret the two factors as the two different types of intelligence. Even if they are uncorrelated, we cannot tell which factor corresponds to verbal intelligence and which corresponds to mathematical intelligence, or whether the factors are linear combinations of both, without an outside argument.

The data vectors have unit length. The entries of the correlation matrix for the data are given by . The correlation matrix can be geometrically interpreted as the cosine of the angle between the two data vectors and . The diagonal elements will clearly be s and the off diagonal elements will have absolute values less than or equal to unity. The "reduced correlation matrix" is defined as

.

The goal of factor analysis is to choose the fitting hyperplane such that the reduced correlation matrix reproduces the correlation matrix as nearly as possible, except for the diagonal elements of the correlation matrix which are known to have unit value. In other words, the goal is to reproduce as accurately as possible the cross-correlations in the data. Specifically, for the fitting hyperplane, the mean square error in the off-diagonal components

is to be minimized, and this is accomplished by minimizing it with respect to a set of orthonormal factor vectors. It can be seen that

The term on the right is just the covariance of the errors. In the model, the error covariance is stated to be a diagonal matrix and so the above minimization problem will in fact yield a "best fit" to the model: It will yield a sample estimate of the error covariance which has its off-diagonal components minimized in the mean square sense. It can be seen that since the are orthogonal projections of the data vectors, their length will be less than or equal to the length of the projected data vector, which is unity. The square of these lengths are just the diagonal elements of the reduced correlation matrix. These diagonal elements of the reduced correlation matrix are known as "communalities":

Large values of the communalities will indicate that the fitting hyperplane is rather accurately reproducing the correlation matrix. The mean values of the factors must also be constrained to be zero, from which it follows that the mean values of the errors will also be zero.

Practical implementation

Types of factor analysis

Exploratory factor analysis

Exploratory factor analysis (EFA) is used to identify complex interrelationships among items and group items that are part of unified concepts. The researcher makes no a priori assumptions about relationships among factors.

Confirmatory factor analysis

Confirmatory factor analysis (CFA) is a more complex approach that tests the hypothesis that the items are associated with specific factors. CFA uses structural equation modeling to test a measurement model whereby loading on the factors allows for evaluation of relationships between observed variables and unobserved variables. Structural equation modeling approaches can accommodate measurement error and are less restrictive than least-squares estimation. Hypothesized models are tested against actual data, and the analysis would demonstrate loadings of observed variables on the latent variables (factors), as well as the correlation between the latent variables.

Types of factor extraction

Principal component analysis (PCA) is a widely used method for factor extraction, which is the first phase of EFA. Factor weights are computed to extract the maximum possible variance, with successive factoring continuing until there is no further meaningful variance left. The factor model must then be rotated for analysis.

Canonical factor analysis, also called Rao's canonical factoring, is a different method of computing the same model as PCA, which uses the principal axis method. Canonical factor analysis seeks factors that have the highest canonical correlation with the observed variables. Canonical factor analysis is unaffected by arbitrary rescaling of the data.

Common factor analysis, also called principal factor analysis (PFA) or principal axis factoring (PAF), seeks the fewest factors which can account for the common variance (correlation) of a set of variables.

Image factoring is based on the correlation matrix of predicted variables rather than actual variables, where each variable is predicted from the others using multiple regression.

Alpha factoring is based on maximizing the reliability of factors, assuming variables are randomly sampled from a universe of variables. All other methods assume cases to be sampled and variables fixed.

Factor regression model is a combinatorial model of factor model and regression model; or alternatively, it can be viewed as the hybrid factor model, whose factors are partially known.

Terminology

Factor loadings
Communality is the square of the standardized outer loading of an item. Analogous to Pearson's r-squared, the squared factor loading is the percent of variance in that indicator variable explained by the factor. To get the percent of variance in all the variables accounted for by each factor, add the sum of the squared factor loadings for that factor (column) and divide by the number of variables. (The number of variables equals the sum of their variances as the variance of a standardized variable is 1.) This is the same as dividing the factor's eigenvalue by the number of variables.
When interpreting, by one rule of thumb in confirmatory factor analysis, factor loadings should be .7 or higher to confirm that independent variables identified a priori are represented by a particular factor, on the rationale that the .7 level corresponds to about half of the variance in the indicator being explained by the factor. However, the .7 standard is a high one and real-life data may well not meet this criterion, which is why some researchers, particularly for exploratory purposes, will use a lower level such as .4 for the central factor and .25 for other factors. In any event, factor loadings must be interpreted in the light of theory, not by arbitrary cutoff levels.
In oblique rotation, one may examine both a pattern matrix and a structure matrix. The structure matrix is simply the factor loading matrix as in orthogonal rotation, representing the variance in a measured variable explained by a factor on both a unique and common contributions basis. The pattern matrix, in contrast, contains coefficients which just represent unique contributions. The more factors, the lower the pattern coefficients as a rule since there will be more common contributions to variance explained. For oblique rotation, the researcher looks at both the structure and pattern coefficients when attributing a label to a factor. Principles of oblique rotation can be derived from both cross entropy and its dual entropy.
Communality
The sum of the squared factor loadings for all factors for a given variable (row) is the variance in that variable accounted for by all the factors. The communality measures the percent of variance in a given variable explained by all the factors jointly and may be interpreted as the reliability of the indicator in the context of the factors being posited.
Spurious solutions
If the communality exceeds 1.0, there is a spurious solution, which may reflect too small a sample or the choice to extract too many or too few factors.
Uniqueness of a variable
The variability of a variable minus its communality.
Eigenvalues/characteristic roots
Eigenvalues measure the amount of variation in the total sample accounted for by each factor. The ratio of eigenvalues is the ratio of explanatory importance of the factors with respect to the variables. If a factor has a low eigenvalue, then it is contributing little to the explanation of variances in the variables and may be ignored as less important than the factors with higher eigenvalues.
Extraction sums of squared loadings
Initial eigenvalues and eigenvalues after extraction (listed by SPSS as "Extraction Sums of Squared Loadings") are the same for PCA extraction, but for other extraction methods, eigenvalues after extraction will be lower than their initial counterparts. SPSS also prints "Rotation Sums of Squared Loadings" and even for PCA, these eigenvalues will differ from initial and extraction eigenvalues, though their total will be the same.
Factor scores
Component scores (in PCA)

Explained from PCA perspective, not from Factor Analysis perspective.

The scores of each case (row) on each factor (column). To compute the factor score for a given case for a given factor, one takes the case's standardized score on each variable, multiplies by the corresponding loadings of the variable for the given factor, and sums these products. Computing factor scores allows one to look for factor outliers. Also, factor scores may be used as variables in subsequent modeling.

Criteria for determining the number of factors

Researchers wish to avoid such subjective or arbitrary criteria for factor retention as "it made sense to me". A number of objective methods have been developed to solve this problem, allowing users to determine an appropriate range of solutions to investigate. However these different methods often disagree with one another as to the number of factors that ought to be retained. For instance, the parallel analysis may suggest 5 factors while Velicer's MAP suggests 6, so the researcher may request both 5 and 6-factor solutions and discuss each in terms of their relation to external data and theory.

Modern criteria

Horn's parallel analysis (PA): A Monte-Carlo based simulation method that compares the observed eigenvalues with those obtained from uncorrelated normal variables. A factor or component is retained if the associated eigenvalue is bigger than the 95th percentile of the distribution of eigenvalues derived from the random data. PA is among the more commonly recommended rules for determining the number of components to retain, but many programs fail to include this option (a notable exception being R). However, Formann provided both theoretical and empirical evidence that its application might not be appropriate in many cases since its performance is considerably influenced by sample size, item discrimination, and type of correlation coefficient.

Velicer's (1976) MAP test as described by Courtney (2013) “involves a complete principal components analysis followed by the examination of a series of matrices of partial correlations” (p. 397 (though this quote does not occur in Velicer (1976) and the cited page number is outside the pages of the citation). The squared correlation for Step “0” (see Figure 4) is the average squared off-diagonal correlation for the unpartialed correlation matrix. On Step 1, the first principal component and its associated items are partialed out. Thereafter, the average squared off-diagonal correlation for the subsequent correlation matrix is then computed for Step 1. On Step 2, the first two principal components are partialed out and the resultant average squared off-diagonal correlation is again computed. The computations are carried out for k minus one step (k representing the total number of variables in the matrix). Thereafter, all of the average squared correlations for each step are lined up and the step number in the analyses that resulted in the lowest average squared partial correlation determines the number of components or factors to retain. By this method, components are maintained as long as the variance in the correlation matrix represents systematic variance, as opposed to residual or error variance. Although methodologically akin to principal components analysis, the MAP technique has been shown to perform quite well in determining the number of factors to retain in multiple simulation studies. This procedure is made available through SPSS's user interface, as well as the psych package for the R programming language.

Older methods

Kaiser criterion: The Kaiser rule is to drop all components with eigenvalues under 1.0 – this being the eigenvalue equal to the information accounted for by an average single item. The Kaiser criterion is the default in SPSS and most statistical software but is not recommended when used as the sole cut-off criterion for estimating the number of factors as it tends to over-extract factors. A variation of this method has been created where a researcher calculates confidence intervals for each eigenvalue and retains only factors which have the entire confidence interval greater than 1.0.

Scree plot: The Cattell scree test plots the components as the X-axis and the corresponding eigenvalues as the Y-axis. As one moves to the right, toward later components, the eigenvalues drop. When the drop ceases and the curve makes an elbow toward less steep decline, Cattell's scree test says to drop all further components after the one starting at the elbow. This rule is sometimes criticised for being amenable to researcher-controlled "fudging". That is, as picking the "elbow" can be subjective because the curve has multiple elbows or is a smooth curve, the researcher may be tempted to set the cut-off at the number of factors desired by their research agenda.

Variance explained criteria: Some researchers simply use the rule of keeping enough factors to account for 90% (sometimes 80%) of the variation. Where the researcher's goal emphasizes parsimony (explaining variance with as few factors as possible), the criterion could be as low as 50%.

Bayesian methods

By placing a prior distribution over the number of latent factors and then applying Bayes' theorem, Bayesian models can return a probability distribution over the number of latent factors. This has been modeled using the Indian buffet process, but can be modeled more simply by placing any discrete prior (e.g. a negative binomial distribution) on the number of components.

Rotation methods

The output of PCA maximizes the variance accounted for by the first factor first, then the second factor, etc. A disadvantage of this procedure is that most items load on the early factors, while very few items load on later variables. This makes interpreting the factors by reading through a list of questions and loadings difficult, as every question is strongly correlated with the first few components, while very few questions are strongly correlated with the last few components.

Rotation serves to make the output easier to interpret. By choosing a different basis for the same principal components – that is, choosing different factors to express the same correlation structure – it is possible to create variables that are more easily interpretable.

Rotations can be orthogonal or oblique; oblique rotations allow the factors to correlate. This increased flexibility means that more rotations are possible, some of which may be better at achieving a specified goal. However, this can also make the factors more difficult to interpret, as some information is "double-counted" and included multiple times in different components; some factors may even appear to be near-duplicates of each other.

Orthogonal methods

Two broad classes of orthogonal rotations exist: those that look for sparse rows (where each row is a case, i.e. subject), and those that look for sparse columns (where each column is a variable).

  • Simple factors: these rotations try to explain all factors by using only a few important variables. This effect can be achieved by using Varimax (the most common rotation).
  • Simple variables: these rotations try to explain all variables using only a few important factors. This effect can be achieved using either Quartimax or the unrotated components of PCA.
  • Both: these rotations try to compromise between both of the above goals, but in the process, may achieve a fit that is poor at both tasks; as such, they are unpopular compared to the above methods. Equamax is one such rotation.

Problems with factor rotation

It can be difficult to interpret a factor structure when each variable is loading on multiple factors. Small changes in the data can sometimes tip a balance in the factor rotation criterion so that a completely different factor rotation is produced. This can make it difficult to compare the results of different experiments. This problem is illustrated by a comparison of different studies of world-wide cultural differences. Each study has used different measures of cultural variables and produced a differently rotated factor analysis result. The authors of each study believed that they had discovered something new, and invented new names for the factors they found. A later comparison of the studies found that the results were rather similar when the unrotated results were compared. The common practice of factor rotation has obscured the similarity between the results of the different studies.

Higher order factor analysis

Higher-order factor analysis is a statistical method consisting of repeating steps factor analysis – oblique rotation – factor analysis of rotated factors. Its merit is to enable the researcher to see the hierarchical structure of studied phenomena. To interpret the results, one proceeds either by post-multiplying the primary factor pattern matrix by the higher-order factor pattern matrices (Gorsuch, 1983) and perhaps applying a Varimax rotation to the result (Thompson, 1990) or by using a Schmid-Leiman solution (SLS, Schmid & Leiman, 1957, also known as Schmid-Leiman transformation) which attributes the variation from the primary factors to the second-order factors.

Exploratory factor analysis (EFA) versus principal components analysis (PCA)

Factor analysis is related to principal component analysis (PCA), but the two are not identical. There has been significant controversy in the field over differences between the two techniques. PCA can be considered as a more basic version of exploratory factor analysis (EFA) that was developed in the early days prior to the advent of high-speed computers. Both PCA and factor analysis aim to reduce the dimensionality of a set of data, but the approaches taken to do so are different for the two techniques. Factor analysis is clearly designed with the objective to identify certain unobservable factors from the observed variables, whereas PCA does not directly address this objective; at best, PCA provides an approximation to the required factors. From the point of view of exploratory analysis, the eigenvalues of PCA are inflated component loadings, i.e., contaminated with error variance.

Whilst EFA and PCA are treated as synonymous techniques in some fields of statistics, this has been criticised. Factor analysis "deals with the assumption of an underlying causal structure: [it] assumes that the covariation in the observed variables is due to the presence of one or more latent variables (factors) that exert causal influence on these observed variables". In contrast, PCA neither assumes nor depends on such an underlying causal relationship. Researchers have argued that the distinctions between the two techniques may mean that there are objective benefits for preferring one over the other based on the analytic goal. If the factor model is incorrectly formulated or the assumptions are not met, then factor analysis will give erroneous results. Factor analysis has been used successfully where adequate understanding of the system permits good initial model formulations. PCA employs a mathematical transformation to the original data with no assumptions about the form of the covariance matrix. The objective of PCA is to determine linear combinations of the original variables and select a few that can be used to summarize the data set without losing much information.

Arguments contrasting PCA and EFA

Fabrigar et al. (1999) address a number of reasons used to suggest that PCA is not equivalent to factor analysis:

  1. It is sometimes suggested that PCA is computationally quicker and requires fewer resources than factor analysis. Fabrigar et al. suggest that readily available computer resources have rendered this practical concern irrelevant.
  2. PCA and factor analysis can produce similar results. This point is also addressed by Fabrigar et al.; in certain cases, whereby the communalities are low (e.g. 0.4), the two techniques produce divergent results. In fact, Fabrigar et al. argue that in cases where the data correspond to assumptions of the common factor model, the results of PCA are inaccurate results.
  3. There are certain cases where factor analysis leads to 'Heywood cases'. These encompass situations whereby 100% or more of the variance in a measured variable is estimated to be accounted for by the model. Fabrigar et al. suggest that these cases are actually informative to the researcher, indicating an incorrectly specified model or a violation of the common factor model. The lack of Heywood cases in the PCA approach may mean that such issues pass unnoticed.
  4. Researchers gain extra information from a PCA approach, such as an individual's score on a certain component; such information is not yielded from factor analysis. However, as Fabrigar et al. contend, the typical aim of factor analysis – i.e. to determine the factors accounting for the structure of the correlations between measured variables – does not require knowledge of factor scores and thus this advantage is negated. It is also possible to compute factor scores from a factor analysis.

Variance versus covariance

Factor analysis takes into account the random error that is inherent in measurement, whereas PCA fails to do so. This point is exemplified by Brown (2009), who indicated that, in respect to the correlation matrices involved in the calculations:

"In PCA, 1.00s are put in the diagonal meaning that all of the variance in the matrix is to be accounted for (including variance unique to each variable, variance common among variables, and error variance). That would, therefore, by definition, include all of the variance in the variables. In contrast, in EFA, the communalities are put in the diagonal meaning that only the variance shared with other variables is to be accounted for (excluding variance unique to each variable and error variance). That would, therefore, by definition, include only variance that is common among the variables."

— Brown (2009), Principal components analysis and exploratory factor analysis – Definitions, differences and choices

For this reason, Brown (2009) recommends using factor analysis when theoretical ideas about relationships between variables exist, whereas PCA should be used if the goal of the researcher is to explore patterns in their data.

Differences in procedure and results

The differences between PCA and factor analysis (FA) are further illustrated by Suhr (2009):

  • PCA results in principal components that account for a maximal amount of variance for observed variables; FA accounts for common variance in the data.
  • PCA inserts ones on the diagonals of the correlation matrix; FA adjusts the diagonals of the correlation matrix with the unique factors.
  • PCA minimizes the sum of squared perpendicular distance to the component axis; FA estimates factors that influence responses on observed variables.
  • The component scores in PCA represent a linear combination of the observed variables weighted by eigenvectors; the observed variables in FA are linear combinations of the underlying and unique factors.
  • In PCA, the components yielded are uninterpretable, i.e. they do not represent underlying ‘constructs’; in FA, the underlying constructs can be labelled and readily interpreted, given an accurate model specification.

In psychometrics

History

Charles Spearman was the first psychologist to discuss common factor analysis and did so in his 1904 paper. It provided few details about his methods and was concerned with single-factor models. He discovered that school children's scores on a wide variety of seemingly unrelated subjects were positively correlated, which led him to postulate that a single general mental ability, or g, underlies and shapes human cognitive performance.

The initial development of common factor analysis with multiple factors was given by Louis Thurstone in two papers in the early 1930s, summarized in his 1935 book, The Vector of Mind. Thurstone introduced several important factor analysis concepts, including communality, uniqueness, and rotation. He advocated for "simple structure", and developed methods of rotation that could be used as a way to achieve such structure.

In Q methodology, William Stephenson, a student of Spearman, distinguish between R factor analysis, oriented toward the study of inter-individual differences, and Q factor analysis oriented toward subjective intra-individual differences.

Raymond Cattell was a strong advocate of factor analysis and psychometrics and used Thurstone's multi-factor theory to explain intelligence. Cattell also developed the scree test and similarity coefficients.

Applications in psychology

Factor analysis is used to identify "factors" that explain a variety of results on different tests. For example, intelligence research found that people who get a high score on a test of verbal ability are also good on other tests that require verbal abilities. Researchers explained this by using factor analysis to isolate one factor, often called verbal intelligence, which represents the degree to which someone is able to solve problems involving verbal skills.

Factor analysis in psychology is most often associated with intelligence research. However, it also has been used to find factors in a broad range of domains such as personality, attitudes, beliefs, etc. It is linked to psychometrics, as it can assess the validity of an instrument by finding if the instrument indeed measures the postulated factors.

Advantages

  • Reduction of number of variables, by combining two or more variables into a single factor. For example, performance at running, ball throwing, batting, jumping and weight lifting could be combined into a single factor such as general athletic ability. Usually, in an item by people matrix, factors are selected by grouping related items. In the Q factor analysis technique, the matrix is transposed and factors are created by grouping related people. For example, liberals, libertarians, conservatives, and socialists might form into separate groups.
  • Identification of groups of inter-related variables, to see how they are related to each other. For example, Carroll used factor analysis to build his Three Stratum Theory. He found that a factor called "broad visual perception" relates to how good an individual is at visual tasks. He also found a "broad auditory perception" factor, relating to auditory task capability. Furthermore, he found a global factor, called "g" or general intelligence, that relates to both "broad visual perception" and "broad auditory perception". This means someone with a high "g" is likely to have both a high "visual perception" capability and a high "auditory perception" capability, and that "g" therefore explains a good part of why someone is good or bad in both of those domains.

Disadvantages

  • "...each orientation is equally acceptable mathematically. But different factorial theories proved to differ as much in terms of the orientations of factorial axes for a given solution as in terms of anything else, so that model fitting did not prove to be useful in distinguishing among theories." (Sternberg, 1977). This means all rotations represent different underlying processes, but all rotations are equally valid outcomes of standard factor analysis optimization. Therefore, it is impossible to pick the proper rotation using factor analysis alone.
  • Factor analysis can be only as good as the data allows. In psychology, where researchers often have to rely on less valid and reliable measures such as self-reports, this can be problematic.
  • Interpreting factor analysis is based on using a "heuristic", which is a solution that is "convenient even if not absolutely true". More than one interpretation can be made of the same data factored the same way, and factor analysis cannot identify causality.

In cross-cultural research

Factor analysis is a frequently used technique in cross-cultural research. It serves the purpose of extracting cultural dimensions. The best known cultural dimensions models are those elaborated by Geert Hofstede, Ronald Inglehart, Christian Welzel, Shalom Schwartz and Michael Minkov. A popular visualization is Inglehart and Welzel's cultural map of the world.

In political science

In an early 1965 study, political systems around the world are examined via factor analysis to construct related theoretical models and research, compare political systems, and create typological categories. For these purposes, in this study seven basic political dimensions are identified, which are related to a wide variety of political behaviour: these dimensions are Access, Differentiation, Consensus, Sectionalism, Legitimation, Interest, and Leadership Theory and Research.

Other political scientists explore the measurement of internal political efficacy using four new questions added to the 1988 National Election Study. Factor analysis is here used to find that these items measure a single concept distinct from external efficacy and political trust, and that these four questions provided the best measure of internal political efficacy up to that point in time.

In marketing

The basic steps are:

  • Identify the salient attributes consumers use to evaluate products in this category.
  • Use quantitative marketing research techniques (such as surveys) to collect data from a sample of potential customers concerning their ratings of all the product attributes.
  • Input the data into a statistical program and run the factor analysis procedure. The computer will yield a set of underlying attributes (or factors).
  • Use these factors to construct perceptual maps and other product positioning devices.

Information collection

The data collection stage is usually done by marketing research professionals. Survey questions ask the respondent to rate a product sample or descriptions of product concepts on a range of attributes. Anywhere from five to twenty attributes are chosen. They could include things like: ease of use, weight, accuracy, durability, colourfulness, price, or size. The attributes chosen will vary depending on the product being studied. The same question is asked about all the products in the study. The data for multiple products is coded and input into a statistical program such as R, SPSS, SAS, Stata, STATISTICA, JMP, and SYSTAT.

Analysis

The analysis will isolate the underlying factors that explain the data using a matrix of associations. Factor analysis is an interdependence technique. The complete set of interdependent relationships is examined. There is no specification of dependent variables, independent variables, or causality. Factor analysis assumes that all the rating data on different attributes can be reduced down to a few important dimensions. This reduction is possible because some attributes may be related to each other. The rating given to any one attribute is partially the result of the influence of other attributes. The statistical algorithm deconstructs the rating (called a raw score) into its various components and reconstructs the partial scores into underlying factor scores. The degree of correlation between the initial raw score and the final factor score is called a factor loading.

Advantages

  • Both objective and subjective attributes can be used provided the subjective attributes can be converted into scores.
  • Factor analysis can identify latent dimensions or constructs that direct analysis may not.
  • It is easy and inexpensive.

Disadvantages

  • Usefulness depends on the researchers' ability to collect a sufficient set of product attributes. If important attributes are excluded or neglected, the value of the procedure is reduced.
  • If sets of observed variables are highly similar to each other and distinct from other items, factor analysis will assign a single factor to them. This may obscure factors that represent more interesting relationships.
  • Naming factors may require knowledge of theory because seemingly dissimilar attributes can correlate strongly for unknown reasons.

In physical and biological sciences

Factor analysis has also been widely used in physical sciences such as geochemistry, hydrochemistryastrophysics and cosmology, as well as biological sciences, such as ecology, molecular biology, neuroscience and biochemistry.

In groundwater quality management, it is important to relate the spatial distribution of different chemical parameters to different possible sources, which have different chemical signatures. For example, a sulfide mine is likely to be associated with high levels of acidity, dissolved sulfates and transition metals. These signatures can be identified as factors through R-mode factor analysis, and the location of possible sources can be suggested by contouring the factor scores.

In geochemistry, different factors can correspond to different mineral associations, and thus to mineralisation.

In microarray analysis

Factor analysis can be used for summarizing high-density oligonucleotide DNA microarrays data at probe level for Affymetrix GeneChips. In this case, the latent variable corresponds to the RNA concentration in a sample.

Implementation

Factor analysis has been implemented in several statistical analysis programs since the 1980s:

Stand-alone

Timeline of the universe

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Timeline_of_the_universe   Diagram of Evol...