Search This Blog

Wednesday, February 9, 2022

Obstetrics and gynaecology

From Wikipedia, the free encyclopedia
 

Obstetrics and gynaecology (British English) or obstetrics and gynecology (American English) is the medical specialty that encompasses the two subspecialties of obstetrics (covering pregnancy, childbirth, and the postpartum period) and gynecology (covering the health of the female reproductive systemvagina, uterus, ovaries, and breasts). It is commonly abbreviated as OB-GYN or OB/GYN in US English and Canadian English, and as obs and gynae or O&G in British English.

Postgraduate training programs for both fields are usually combined, preparing the practicing obstetrician-gynecologist to be adept both at the care of female reproductive organs' health and at the management of pregnancy, although many doctors go on to develop subspecialty interests in one field or the other.

Education and training (residency)

United States

According to the American Board of Obstetrics and Gynecology (ABOG), which is responsible for issuing OB-GYN certifications in the United States, the first step to OB-GYN certification is completing medical school to receive an MD or DO degree. From there doctors must complete a four-year OB-GYN residency program approved by the Accreditation Council for Graduate Medical Education (ACGME). For the 2021 Electronic Residency Application Service (ERAS) match, there were 277 OB-GYN residency programs accepting applicants.

In their fourth year of residency, with an affidavit prom their director to confirm program completion, OB-GYN residents can choose whether to begin the board certification process by applying to take the ABOG Qualifying Exam, which is a written test. If residents pass the Qualifying Exam, demonstrating they possess the knowledge and skills to potentially become certified OB-GYNs, they are then eligible to sit for the oral Certification Exam. Prior to the Certification Exam, residents must also gather a list of patient cases they've worked on throughout their residency in order to demonstrate their competence and experience in OB-GYN patient care.

Residents then sit for the three-hour oral exam at ABOG's test center, and if they pass the exam they become "board certified" OB-GYNs. Since 2013 at least 82% of all Certifying Exam examinees have passed.

This adds up to 11–14 years of education and practical experience. The first 7–9 years are general medical training.

Experienced OB-GYN professionals can seek certifications in sub-specialty areas, including maternal and fetal medicine. See Fellowship (medicine).

United Kingdom

All doctors must first complete medical school and obtain a MBBS or equivalent certification. This portion typically takes five years. Following this, they are eligible for provisional registration with the General Medical Council. They then must complete a two years of foundation training. After the first year of training is complete, trainees are eligible for full registration with the General Medical Council. After the foundation training is complete applicants take the Part 1 MRCOG examination administered by the Royal College of Obstetricians and Gynaecologists. There are an additional seven years of training after this, and two more exams (Part 2 and Part 3 MRCOG exams) which adds up to nine years total minimum in training, although some trainees may take longer.

Subspecialties

Examples of subspecialty training available to physicians in the US are:

Of these, only the first four are truly recognized sub-specialties by the Accreditation Council for Graduate Medical Education (ACGME) and the American Board of Obstetrics and Gynecology (ABOG). The other subspecialties are recognized as informal concentrations of practice. To be recognized as a board-certified subspecialist by the American Board of Obstetrics and Gynecology or the American Osteopathic Board of Obstetrics and Gynecology, a practitioner must have completed an ACGME or AOA-accredited residency and obtained a Certificate of Added Qualifications (CAQ) which requires an additional standardized examination.

Additionally, physicians of other specialties may become trained in Advanced Life Support in Obstetrics (ALSO), a short certification that equips them to better manage emergent OB/GYN situations.

Common procedures

There are many procedures that can be provided to people by OB/GYN's. Some procedures may include:

  • Colposcopy: If the results of a cervical cancer screening test, such as Pap smear or HPV test, are abnormal this more thorough examination of the cervix and vaginal tissues may be needed.
  • Loop electrical excision procedure (LEEP): a procedure to quickly remove abnormal vaginal tissue within the cervix. A local anesthetic and a solution to enhance the points of removal visually is administered during the process. There is a chance of experiencing watery, pinkish discharge, brownish discharge, and mild cramping.
  • Endometrial biopsy: a procedure that collects a tissue sample from the endometrium lining of the uterus. The sample is tested and checked under a microscope for abnormals cells or indicators of cancer.
  • IUD insertion: an intrauterine device that is T-shaped and is placed in the uterus through the cervix. It is a reversible contraceptive that can be done in a doctor's office.
  • Nexplanon: is about a 4 cm implant that goes into the upper forearm. This implant releases birth control hormones into the body and can last up to three years. This type of birth control has a 99% success rate for pregnancy prevention.
  • Dilation and curettage (D&C): an out-patient procedure to open (dilate) the cervix to collect samples of endometrial tissue with a curette. A D&C can also be done to remove a fetus that was not passed naturally after a miscarriage or to induce an abortion.
  • Tubal ligation: a surgery to close the fallopian tubes for the prevention of pregnancy. It is also known as "tying the tubes".
  • Ovarian cystectomy: the removal of a cyst that either has a solid appearance, larger than three inches in diameter, has the possibility to become cancerous, or causes a constant pain. Cysts can be removed without removing an ovary. Women who do not take birth control produce small cysts every other month but they can disappear on their own.

Inclusive approaches to care

There is no global standard, or national U.S. standard, for OB-GYN curricula, during or after residency. Hospitals and universities implement their care and provider education with different priorities in mind, some institutions focusing on following the requirements from their governing body while others go beyond the set standards to set a precedent with the care they offer.

However, the American College of Obstetrics and Gynecology (ACOG) encourages OB-GYN care providers to offer care that is inclusive to all individuals, in the context of providing a safe space for lesbian and bisexual women, and transgender and gender diverse individuals. ACOG shares this recommendation through a series of Committee Opinions, the latest of which are #525 and #823, published 2012 and 2021 respectively. ACOG's first comment on the need for LGBTQ+ inclusive care from OB-GYNs came in 2011 through Committee Opinion #512, "Healthcare for Transgender Individuals", which has since been withdrawn and replaced by 2021 opinion #823.

ACOG recommends that OB-GYNs should:

  • Be a resource for both patients and families with health-oriented information on sexual orientation and gender issues
  • Provide gender-affirming surgeries if possible, and treat these procedures as medically necessary for patients with gender dysphoria
  • In the process of gender transition, fertility and parenting desires should be discussed early on, and patients should be offered all available information
  • Preventative screening procedures should be offered based on anatomical structures present, not gender identity
  • Create a structurally inclusive environment by making their offices "inviting to all individuals who need obstetric or gynecologic health care"
    • Educate all care providers and office personnel about LGBTQ+ health, including the use of inclusive language and the unique needs of LGBTQ+ patients
    • Post a nondiscrimination policy for the office in the reception area
  • Offer comprehensive patient education specific to LGBTQ+ individuals, like clarifying that gender hormone therapy is not a form of birth control

The need for LGBTQ+ affirming OB-GYNs

Transgender, nonbinary people, and lesbian or bisexual women frequently face additional challenges—including structural and overt discrimination, accessing insurance coverage stigma, and cisnormativity—when attempting to access gynecologic and reproductive healthcare services and providers. Obstetrics and gynecology is a field thought of as traditionally serving women because of its focus on the female reproductive system, leading care providers to make assumptions about patients' gender identity and expression in "women's health clinics" when many transgender or nonbinary patients may also seek care from OB-GYNs. Additionally, LGBTQ+ patients are at risk for and experience unique health conditions which care providers may not be prepared to address—i.e. the providers lack cultural competence—without LGBTQ+ specific healthcare training.

Interactions with care providers who are not prepared or knowledgeable enough to offer comprehensive and effective care can compound the "negative and traumatic experiences that many nonbinary and transgender people have had when accessing care", and drive them away from seeking necessary healthcare in the future. Even if transgender or nonbinary patients have not had a personally negative of traumatic experience, collective trauma can also play into a fear of medical settings and providers. This results in nearly 25% of transgender or nonbinary people reporting avoidance of healthcare services out of fear of being mistreated for their gender. Care typically offered by OB-GYNs, like pelvic exams and procedures, can be "particularly difficult or traumatic" for transgender and nonbinary patients, which the provider can offset by engaging a trauma-informed care approach to exams.

More research is gradually being done to establish the need for an LGBTQ+ healthcare curriculum for all medical students, residents, and doctors. Multiple researchers have also suggested that obstetrics & gynecology is the ideal field for this education to initially be implemented, since the nature of OB-GYN work ties it more directly to gender than other medical specialties. OB-GYNs also historically have more cultural competency training around gender based issues.

Some common services LGBTQ+ folks may seek from OB-GYNs include:

If providers are not educated on LGBTQ+ healthcare they will not be aware of some of the health risks LGBTQ+ patients may be more vulnerable too, placing their patients at greater risk. Common chronic diseases the LGBTQ+ population is at risk for:

Teaching LGBTQ+ healthcare

Most OB-GYN programs have a flexible curriculum that offers learning in a variety of contexts, including small group discussions, case study discussions, and structured patient exams. The multifaceted nature of this teaching structure makes it easier for programs to begin initiating LGBTQ+ healthcare instruction in informal voluntary contexts, while simultaneously advocating for long-term structural change that would incorporate LGBTQ+ healthcare into all OB-GYN and medical school curricula.

When initially instituting a LGBTQ+ healthcare curricula for OB-GYN residents or professionals, the Association of Professors of Gynecology and Obstetrics' (APGO) Undergraduate Medical Education Committee (UMEC) recommends that topics like patient education, screening standards, and common chronic diseases in the LGBTQ+ population are prioritized to ensure comprehensive and effective care for LGBTQ+ patients.

In residency

APGO's UMCE poses that academic health centers are in an ideal position to drive the "curricular change that is needed to remove health disparities" for all patient populations in the LGBTQ+ community.

Although not a requirement within the ACGME's standards for American OB-GYN residency programs curriculum, LGBTQ+ healthcare education is established in some OB-GYN residency programs throughout America. A web-based survey of approximately 100 Illinois OB-GYN residents a survey study found that 62% of their programs had spent 1-5 hours in the past year educating residents about lesbian and bisexual health, and 63% of the their programs had spent 1-5 hours in the past year educating residents about transgender health.

APGO's UMEC identifies the third year of residency or "clerkship" as an ideal time to institute education if it hasn't already occurred, and ideally to reinforce competencies with LGBTQ+ patients through clinical experiences.

Established OB-GYNs

In addition to instituting new educational policies for current and future OB-GYN residents, many programs are also looking to educate OB-GYNs that are already certified.

For example, the University of Michigan OB-GYN program utilizes what they name "Continuing Medical Education" to instruct their OB-GYN care providers on the nuances of caring for transgender patients. Daphna Strousma, MD, MPH, is responsible for instituting and developing UM's transgender healthcare curriculum—"Improving Care for Transgender & Non-Binary Individuals"—in coordination with Michigan Medicine and the Council on Resident Education in Obstetrics and Gynecology (CREOG). The curriculum is offered to UM's OB-GYNs through a series of video modules covering the following topics:

  1. Gender Identity and Care of Transgender and Gender Non-Conforming Patients
  2. Preventative Care for Transgender and Gender Non-Conforming Patients
  3. Gender Affirming Treatment & Transition Related Care
  4. Addressing Common Gynecologic Issues Among Transgender Patients
  5. Health Records, Billing, Insurance, and Legal Documents in Transgender Medicine

Challenges to instituting inclusive care

Although the ACOG has called upon OB-GYN care providers to provide gender-affirming and inclusive care for all individuals, that doesn't mean there hasn't been resistance, particularly because there is no formal legislation mandating the inclusion of LGBTQ+ healthcare in OB-GYN curricula.

Many OB-GYN providers do not currently feel comfortable offering care to LGBTQ+ patients, either because of their personal beliefs or due to a lack of education on LGBTQ+ healthcare. One study of approximately 100 Illinois OB-GYN residents found that 50% of residents felt unprepared to care for lesbian or bisexual patients, and 76% of residents felt unprepared to offer care to transgender patients. The two main areas the residents identified as preventing the implementation of LGBTQ+ healthcare training were curricular crowding, 85%, and lack of experienced faculty, 91%. However, 92% of residents wanted to receive more education on how to offer care to LGBTQ+ patients.

Steam reforming

From Wikipedia, the free encyclopedia

Illustrating inputs and outputs of steam reforming of natural gas, a process to produce hydrogen and CO2 greenhouse gas that may be captured with CCS

Steam reforming or steam methane reforming is a method for producing syngas (hydrogen and carbon monoxide) by reaction of hydrocarbons with water. Commonly natural gas is the feedstock. The main purpose of this technology is hydrogen production. The reaction is represented by this equilibrium:

The reaction is strongly endothermicHSR = 206 kJ/mol).

Hydrogen produced by steam reforming is termed 'grey hydrogen' when the waste carbon dioxide is released to the atmosphere and 'blue hydrogen' when the carbon dioxide is (mostly) captured and stored geologically - see carbon capture and storage. (Zero carbon 'green' hydrogen is produced by electrolysis using low- or zero-carbon electricity. Zero carbon emissions 'turquoise' hydrogen is produced by one-step methane pyrolysis of natural gas.)

Steam reforming of natural gas produces most of the world's hydrogen. Hydrogen is used in the industrial synthesis of ammonia and other chemicals.

Reactions

Steam reforming reaction kinetics, in particular using nickel-alumina catalysts, have been studied in detail since the 1950s.

Pre-reforming

Depiction of the general process flow of a typical steam reforming plant. From left to right: Desulphurisation, pre-reforming, steam reforming, shift conversion, and pressure-swing-adsorption.
Depiction of the general process flow of a typical steam reforming plant. (PSA = Pressure swing adsorption, NG = Natural gas)

The purpose of pre-reforming is to break down higher hydrocarbons such as propane, butane or naphta into methane (CH4), which allows for more efficient reforming downstream.

Steam reforming

The name-giving reaction is the steam reforming (SR) reaction and is expressed by the equation:

Via the water-gas shift reaction (WGSR), additional hydrogen is released by reaction of water with the carbon monoxide generated according to equation [1]:

Some additional reactions occurring within steam reforming processes have been studied. Commonly the direct steam reforming (DSR) reaction is also included:

As these reactions by themselves are highly endothermic (apart from WGSR, which is mildly exothermic), a large amount of heat needs to be added to the reactor to keep a constant temperature. Optimal SMR reactor operating conditions lie within a temperature range of 800 °C to 900 °C at medium pressures of 20-30 bar. High excess of steam is required, expressed by the (molar) steam-to-carbon (S/C) ratio. Typical S/C ratio values lie within the range 2.5:1 - 3:1.

Industrial practice

Global Hydrogen Production by Method

The reaction is conducted in multitubular packed bed reactors, a subtype of the plug flow reactor category. These reactors consist of an array of long and narrow tubes which are situated within the combustion chamber of a large industrial furnace, providing the necessary energy to keep the reactor at a constant temperature during operation. Furnace designs vary, depending on the burner configuration they are typically categorized into: top-fired, bottom-fired, and side-fired. A notable design is the Foster-Wheeler terrace wall reformer.

Inside the tubes, a mixture of steam and methane are put into contact with a nickel catalyst. Catalysts with high surface-area-to-volume ratio are preferred because of diffusion limitations due to high operating temperature. Examples of catalyst shapes used are spoked wheels, gear wheels, and rings with holes (see: Raschig rings). Additionally, these shapes have a low pressure drop which is advantageous for this application.

Steam reforming of natural gas is 65–75% efficient.

The United States produces 9–10 million tons of hydrogen per year, mostly with steam reforming of natural gas. The worldwide ammonia production, using hydrogen derived from steam reforming, was 144 million tonnes in 2018. The energy consumption has been reduced from 100 GJ/tonne of ammonia in 1920 to 27 GJ by 2019.

Globally, almost 50% of hydrogen is produced via steam reforming. It is currently the least expensive method for hydrogen production available in terms of its capital cost.

In an effort to decarbonise hydrogen production, carbon capture and storage (CCS) methods are being implemented within the industry, which have the potential to remove up to 90% of CO2 produced from the process. Despite this, implementation of this technology remains problematic, costly, and increases the price of the produced hydrogen significantly.

Autothermal reforming

Autothermal reforming (ATR) uses oxygen and carbon dioxide or steam in a reaction with methane to form syngas. The reaction takes place in a single chamber where the methane is partially oxidized. The reaction is exothermic due to the high energy of O2 with its relatively weak double bond. When the ATR uses carbon dioxide, the H2:CO ratio produced is 1:1; when the ATR uses steam, the H2:CO ratio produced is 2.5:1. The outlet temperature of the syngas is between 950–1100 °C and outlet pressure can be as high as 100 bar.

In addition to reactions [1] - [3], ATR introduces the following reaction:

The main difference between SMR and ATR is that SMR only uses air for combustion as a heat source to create steam, while ATR uses purified oxygen. The advantage of ATR is that the H2:CO ratio can be varied, which can be useful for producing specialty products. Due to the exothermic nature of some of the additional reactions occurring within ATR, the process can essentially be performed at a net enthalpy of zero (ΔH = 0).

Partial oxidation

Partial oxidation (POX) occurs when a sub-stoichiometric fuel-air mixture is partially combusted in a reformer creating hydrogen-rich syngas. POX is typically much faster than steam reforming and requires a smaller reactor vessel. POX produces less hydrogen per unit of the input fuel than steam reforming of the same fuel.

Steam reforming at small scale

The capital cost of steam reforming plants is considered prohibitive for small to medium size applications. The costs for these elaborate facilities do not scale down well. Conventional steam reforming plants operate at pressures between 200 and 600 psi (14–40 bar) with outlet temperatures in the range of 815 to 925 °C.

For combustion engines

Flared gas and vented volatile organic compounds (VOCs) are known problems in the offshore industry and in the on-shore oil and gas industry, since both release greenhouse gases into the atmosphere. Reforming for combustion engines utilizes steam reforming technology for converting waste gases into a source of energy.

Reforming for combustion engines is based on steam reforming, where non-methane hydrocarbons (NMHCs) of low quality gases are converted to synthesis gas (H2 + CO) and finally to methane (CH4), carbon dioxide (CO2) and hydrogen (H2) - thereby improving the fuel gas quality (methane number).

For fuel cells

There is also interest in the development of much smaller units based on similar technology to produce hydrogen as a feedstock for fuel cells. Small-scale steam reforming units to supply fuel cells are currently the subject of research and development, typically involving the reforming of methanol, but other fuels are also being considered such as propane, gasoline, autogas, diesel fuel, and ethanol.

Disadvantages

The reformer– the fuel-cell system is still being researched but in the near term, systems would continue to run on existing fuels, such as natural gas or gasoline or diesel. However, there is an active debate about whether using these fuels to make hydrogen is beneficial while global warming is an issue. Fossil fuel reforming does not eliminate carbon dioxide release into the atmosphere but reduces the carbon dioxide emissions and nearly eliminates carbon monoxide emissions as compared to the burning of conventional fuels due to increased efficiency and fuel cell characteristics. However, by turning the release of carbon dioxide into a point source rather than distributed release, carbon capture and storage becomes a possibility, which would prevent the carbon dioxide's release to the atmosphere, while adding to the cost of the process.

The cost of hydrogen production by reforming fossil fuels depends on the scale at which it is done, the capital cost of the reformer, and the efficiency of the unit, so that whilst it may cost only a few dollars per kilogram of hydrogen at an industrial scale, it could be more expensive at the smaller scale needed for fuel cells.

Challenges with reformers supplying fuel cells

There are several challenges associated with this technology:

  • The reforming reaction takes place at high temperatures, making it slow to start up and requiring costly high-temperature materials.
  • Sulfur compounds in the fuel will poison certain catalysts, making it difficult to run this type of system from ordinary gasoline. Some new technologies have overcome this challenge with sulfur-tolerant catalysts.
  • Coking would be another cause of catalyst deactivation during steam reforming. High reaction temperatures, low steam-to-carbon ratio (S/C), and the complex nature of sulfur-containing commercial hydrocarbon fuels make coking especially favorable. Olefins, typically ethylene, and aromatics are well-known carbon-precursors, hence their formation must be reduced during steam reforming. Additionally, catalysts with lower acidity were reported to be less prone to coking by suppressing dehydrogenation reactions. H2S, the main product in the reforming of organic sulfur, can bind to all transition metal catalysts to form metal–sulfur bonds and subsequently reduce catalyst activity by inhibiting the chemisorption of reforming reactants. Meanwhile, the adsorbed sulfur species increases the catalyst acidity, and hence indirectly promotes coking. Precious metal catalysts such as Rh and Pt have lower tendencies to form bulk sulfides than other metal catalysts such as Ni. Rh and Pt are less prone to sulfur poisoning by only chemisorbing sulfur rather than forming metal sulfides.
  • Low temperature polymer fuel cell membranes can be poisoned by the carbon monoxide (CO) produced by the reactor, making it necessary to include complex CO-removal systems. Solid oxide fuel cells (SOFC) and molten carbonate fuel cells (MCFC) do not have this problem, but operate at higher temperatures, slowing start-up time, and requiring costly materials and bulky insulation.
  • The thermodynamic efficiency of the process is between 70% and 85% (LHV basis) depending on the purity of the hydrogen product.

Tuesday, February 8, 2022

p-value

From Wikipedia, the free encyclopedia
https://en.wikipedia.org/wiki/P-value

In null hypothesis significance testing, the p-value is the probability of obtaining test results at least as extreme as the results actually observed, under the assumption that the null hypothesis is correct. A very small p-value means that such an extreme observed outcome would be very unlikely under the null hypothesis. Reporting p-values of statistical tests is common practice in academic publications of many quantitative fields. Since the precise meaning of p-value is hard to grasp, misuse is widespread and has been a major topic in metascience.

Basic concepts

In statistics, every conjecture concerning the unknown probability distribution of a collection of random variables representing the observed data in some study is called a statistical hypothesis. If we state one hypothesis only and the aim of the statistical test is to see whether this hypothesis is tenable, but not to investigate other specific hypotheses, then such a test is called a null hypothesis test.

As our statistical hypothesis will, by definition, state some property of the distribution, the null hypothesis is the default hypothesis under which that property does not exist. The null hypothesis is typically that some parameter (such as a correlation or a difference between means) in the populations of interest is zero. Note that our hypothesis might specify the probability distribution of precisely, or it might only specify that it belongs to some class of distributions. Often, we reduce the data to a single numerical statistic, e.g., , whose marginal probability distribution is closely connected to a main question of interest in the study.

The p-value is used in the context of null hypothesis testing in order to quantify the statistical significance of a result, the result being the observed value of the chosen statistic .  The lower the p-value is, the lower the probability of getting that result if the null hypothesis were true. A result is said to be statistically significant if it allows us to reject the null hypothesis. All other things being equal, smaller p-values are taken as stronger evidence against the null hypothesis

Loosely speaking, rejection of the null hypothesis implies that there is sufficient evidence against it.

As a particular example, if a null hypothesis states that a certain summary statistic follows the standard normal distribution N(0,1), then the rejection of this null hypothesis could mean that (i) the mean of is not 0, or (ii) the variance of is not 1, or (iii) is not normally distributed. Different tests of the same null hypothesis would be more or less sensitive to different alternatives. However, even if we do manage to reject the null hypothesis for all 3 alternatives, and even if we know the distribution is normal and variance is 1, the null hypothesis test does not tell us which non-zero values of the mean are now most plausible. The more independent observations from the same probability distribution one has, the more accurate the test will be, and the higher the precision with which one will be able to determine the mean value and show that it is not equal to zero; but this will also increase the importance of evaluating the real-world or scientific relevance of this deviation.

Definition and interpretation

General

P-value in statistical significance testing.svg

Consider an observed test-statistic from unknown distribution . Then the p-value is what the prior probability would be of observing a test-statistic value at least as "extreme" as if null hypothesis were true. That is:

  • for a one-sided right-tail test,
  • for a one-sided left-tail test,
  • for a two-sided test. If distribution is symmetric about zero, then

If the p-value is very small, then either the null hypothesis is false or something unlikely has occurred. In a formal significance test, the null hypothesis is rejected if the p-value is less than a pre-defined threshold value , which is referred to as the alpha level or significance level. The value of is instead set by the researcher before examining the data. defines the proportion of the distribution, , that is said to define such a narrow range of all the possible outcomes of that if 's value is within that range its value is unlikely to have occurred by chance. Intuitively, this means that if is set to be 0.10, only 1/10th of the distribution of is defined by , so if falls within that range it is already occurring over a number of outcomes that happen a rare 1/10th of the time, thus suggesting this is unlikely to occur randomly. By convention, is commonly set to 0.05, though lower alpha levels are sometimes used.

The p-value is a function of the chosen test statistic and is therefore a random variable. If the null hypothesis fixes the probability distribution of precisely, and if that distribution is continuous, then when the null-hypothesis is true, the p-value is uniformly distributed between 0 and 1. Thus, the p-value is not fixed. If the same test is repeated independently with fresh data (always with the same probability distribution), one will obtain a different p-value in each iteration. If the null-hypothesis is composite, or the distribution of the statistic is discrete, the probability of obtaining a p-value less than or equal to any number between 0 and 1 is less than or equal to that number, if the null-hypothesis is true. It remains the case that very small values are relatively unlikely if the null-hypothesis is true, and that a significance test at level is obtained by rejecting the null-hypothesis if the significance level is less than or equal to .

Different p-values based on independent sets of data can be combined, for instance using Fisher's combined probability test.

Distribution

When the null hypothesis is true, if it takes the form , and the underlying random variable is continuous, then the probability distribution of the p-value is uniform on the interval [0,1]. By contrast, if the alternative hypothesis is true, the distribution is dependent on sample size and the true value of the parameter being studied.

The distribution of p-values for a group of studies is sometimes called a p-curve. A p-curve can be used to assess the reliability of scientific literature, such as by detecting publication bias or p-hacking.

For composite hypothesis

In parametric hypothesis testing problems, a simple or point hypothesis refers to a hypothesis where the parameter's value is assumed to be a single number. In contrast, in a composite hypothesis the parameter's value is given by a set of numbers. For example, when testing the null hypothesis that a distribution is normal with a mean less than or equal to zero against the alternative that the mean is greater than zero (variance known), the null hypothesis does not specify the probability distribution of the appropriate test statistic. In the just mentioned example that would be the Z-statistic belonging to the one-sided one-sample Z-test. For each possible value of the theoretical mean, the Z-test statistic has a different probability distribution. In these circumstances (the case of a so-called composite null hypothesis) the p-value is defined by taking the least favourable null-hypothesis case, which is typically on the border between null and alternative.

This definition ensures the complementarity of p-values and alpha-levels. If we set the significance level alpha to 0.05, and only reject the null hypothesis if the p-value is less than or equal to 0.05, then our hypothesis test will indeed have significance level (maximal type 1 error rate) 0.05. As Neyman wrote: “The error that a practising statistician would consider the more important to avoid (which is a subjective judgment) is called the error of the first kind. The first demand of the mathematical theory is to deduce such test criteria as would ensure that the probability of committing an error of the first kind would equal (or approximately equal, or not exceed) a preassigned number α, such as α = 0.05 or 0.01, etc. This number is called the level of significance”; Neyman 1976, p. 161 in "The Emergence of Mathematical Statistics: A Historical Sketch with Particular Reference to the United States","On the History of Statistics and Probability", ed. D.B. Owen, New York: Marcel Dekker, pp. 149-193. See also "Confusion Over Measures of Evidence (p's) Versus Errors (a's) in Classical Statistical Testing", Raymond Hubbard and M. J. Bayarri, The American Statistician, August 2003, Vol. 57, No 3, 171--182 (with discussion). For a concise modern statement see Chapter 10 of "All of Statistics: A Concise Course in Statistical Inference", Springer; 1st Corrected ed. 20 edition (September 17, 2004). Larry Wasserman.

Usage

The p-value is widely used in statistical hypothesis testing, specifically in null hypothesis significance testing. In this method, before conducting the study, one first chooses a model (the null hypothesis) and the alpha level α (most commonly .05). After analyzing the data, if the p-value is less than α, that is taken to mean the observed data is sufficiently inconsistent with the null hypothesis for the null hypothesis to be rejected. However, that does not prove that the null hypothesis is false. The p-value does not, in itself, establish probabilities of hypotheses. Rather, it is a tool for deciding whether to reject the null hypothesis.

Misuse

According to the ASA, there is widespread agreement that p-values are often misused and misinterpreted. One practice that has been particularly criticized is accepting the alternative hypothesis for any p-value nominally less than .05 without other supporting evidence. Although p-values are helpful in assessing how incompatible the data are with a specified statistical model, contextual factors must also be considered, such as "the design of a study, the quality of the measurements, the external evidence for the phenomenon under study, and the validity of assumptions that underlie the data analysis". Another concern is that the p-value is often misunderstood as being the probability that the null hypothesis is true.

Some statisticians have proposed abandoning p-values and focusing more on other inferential statistics, such as confidence intervals, likelihood ratios, or Bayes factors, but there is heated debate on the feasibility of these alternatives. Others have suggested to remove fixed significance thresholds and to interpret p-values as continuous indices of the strength of evidence against the null hypothesis. Yet others suggested to report alongside p-values the prior probability of a real effect that would be required to obtain a false positive risk (i.e. the probability that there is no real effect) below a pre-specified threshold (e.g. 5%).

Calculation

Usually, is a test statistic. A test statistic is the output of a scalar function of all the observations. This statistic provides a single number, such as a t-statistic or an F-statistic. As such, the test statistic follows a distribution determined by the function used to define that test statistic and the distribution of the input observational data.

For the important case in which the data are hypothesized to be a random sample from a normal distribution, depending on the nature of the test statistic and the hypotheses of interest about its distribution, different null hypothesis tests have been developed. Some such tests are the z-test for hypotheses concerning the mean of a normal distribution with known variance, the t-test based on Student's t-distribution of a suitable statistic for hypotheses concerning the mean of a normal distribution when the variance is unknown, the F-test based on the F-distribution of yet another statistic for hypotheses concerning the variance. For data of other nature, for instance categorical (discrete) data, test statistics might be constructed whose null hypothesis distribution is based on normal approximations to appropriate statistics obtained by invoking the central limit theorem for large samples, as in the case of Pearson's chi-squared test.

Thus computing a p-value requires a null hypothesis, a test statistic (together with deciding whether the researcher is performing a one-tailed test or a two-tailed test), and data. Even though computing the test statistic on given data may be easy, computing the sampling distribution under the null hypothesis, and then computing its cumulative distribution function (CDF) is often a difficult problem. Today, this computation is done using statistical software, often via numeric methods (rather than exact formulae), but, in the early and mid 20th century, this was instead done via tables of values, and one interpolated or extrapolated p-values from these discrete values. Rather than using a table of p-values, Fisher instead inverted the CDF, publishing a list of values of the test statistic for given fixed p-values; this corresponds to computing the quantile function (inverse CDF).

Example

As an example of a statistical test, an experiment is performed to determine whether a coin flip is fair (equal chance of landing heads or tails) or unfairly biased (one outcome being more likely than the other).

Suppose that the experimental results show the coin turning up heads 14 times out of 20 total flips. The full data would be a sequence of twenty times the symbol "H" or "T". The statistic on which one might focus, could be the total number of heads. The null hypothesis is that the coin is fair, and coin tosses are independent of one another. If a right-tailed test is considered, which would be the case if one is actually interested in the possibility that the coin is biased towards falling heads, then the p-value of this result is the chance of a fair coin landing on heads at least 14 times out of 20 flips. That probability can be computed from binomial coefficients as

This probability is the p-value, considering only extreme results that favor heads. This is called a one-tailed test. However, one might be interested in deviations in either direction, favoring either heads or tails. The two-tailed p-value, which considers deviations favoring either heads or tails, may instead be calculated. As the binomial distribution is symmetrical for a fair coin, the two-sided p-value is simply twice the above calculated single-sided p-value: the two-sided p-value is 0.115.

In the above example:

  • Null hypothesis (H0): The coin is fair, with Pr(heads) = 0.5
  • Test statistic: Number of heads
  • Alpha level (designated threshold of significance): 0.05
  • Observation O: 14 heads out of 20 flips; and
  • Two-tailed p-value of observation O given H0 = 2 × min(Pr(no. of heads ≥ 14 heads), Pr(no. of heads ≤ 14 heads)) = 2 × min(0.058, 0.978) = 2*0.058 = 0.115.

Note that the Pr (no. of heads ≤ 14 heads) = 1 - Pr(no. of heads ≥ 14 heads) + Pr (no. of head = 14) = 1 - 0.058 + 0.036 = 0.978; however, symmetry of the binomial distribution makes it an unnecessary computation to find the smaller of the two probabilities. Here, the calculated p-value exceeds .05, meaning that the data falls within the range of what would happen 95% of the time were the coin in fact fair. Hence, the null hypothesis is not rejected at the .05 level.

However, had one more head been obtained, the resulting p-value (two-tailed) would have been 0.0414 (4.14%), in which case the null hypothesis would be rejected at the .05 level.

History

Chest high painted portrait of man wearing a brown robe and head covering
Man seated at his desk looking up at the camera
Sepia toned photo of young man wearing a suit, a medal, and wire-rimmed eyeglasses

Computations of p-values date back to the 1700s, where they were computed for the human sex ratio at birth, and used to compute statistical significance compared to the null hypothesis of equal probability of male and female births. John Arbuthnot studied this question in 1710, and examined birth records in London for each of the 82 years from 1629 to 1710. In every year, the number of males born in London exceeded the number of females. Considering more male or more female births as equally likely, the probability of the observed outcome is 1/282, or about 1 in 4,836,000,000,000,000,000,000,000; in modern terms, the p-value. This is vanishingly small, leading Arbuthnot that this was not due to chance, but to divine providence: "From whence it follows, that it is Art, not Chance, that governs." In modern terms, he rejected the null hypothesis of equally likely male and female births at the p = 1/282 significance level. This and other work by Arbuthnot is credited as "… the first use of significance tests …" the first example of reasoning about statistical significance, and "… perhaps the first published report of a nonparametric test …", specifically the sign test; see details at Sign test § History.

The same question was later addressed by Pierre-Simon Laplace, who instead used a parametric test, modeling the number of male births with a binomial distribution:

In the 1770s Laplace considered the statistics of almost half a million births. The statistics showed an excess of boys compared to girls. He concluded by calculation of a p-value that the excess was a real, but unexplained, effect.

The p-value was first formally introduced by Karl Pearson, in his Pearson's chi-squared test, using the chi-squared distribution and notated as capital P. The p-values for the chi-squared distribution (for various values of χ2 and degrees of freedom), now notated as P, were calculated in (Elderton 1902), collected in (Pearson 1914, pp. xxxi–xxxiii, 26–28, Table XII).

The use of the p-value in statistics was popularized by Ronald Fisher, and it plays a central role in his approach to the subject. In his influential book Statistical Methods for Research Workers (1925), Fisher proposed the level p = 0.05, or a 1 in 20 chance of being exceeded by chance, as a limit for statistical significance, and applied this to a normal distribution (as a two-tailed test), thus yielding the rule of two standard deviations (on a normal distribution) for statistical significance (see 68–95–99.7 rule).

He then computed a table of values, similar to Elderton but, importantly, reversed the roles of χ2 and p. That is, rather than computing p for different values of χ2 (and degrees of freedom n), he computed values of χ2 that yield specified p-values, specifically 0.99, 0.98, 0.95, 0,90, 0.80, 0.70, 0.50, 0.30, 0.20, 0.10, 0.05, 0.02, and 0.01. That allowed computed values of χ2 to be compared against cutoffs and encouraged the use of p-values (especially 0.05, 0.02, and 0.01) as cutoffs, instead of computing and reporting p-values themselves. The same type of tables were then compiled in (Fisher & Yates 1938), which cemented the approach.

As an illustration of the application of p-values to the design and interpretation of experiments, in his following book The Design of Experiments (1935), Fisher presented the lady tasting tea experiment, which is the archetypal example of the p-value.

To evaluate a lady's claim that she (Muriel Bristol) could distinguish by taste how tea is prepared (first adding the milk to the cup, then the tea, or first tea, then milk), she was sequentially presented with 8 cups: 4 prepared one way, 4 prepared the other, and asked to determine the preparation of each cup (knowing that there were 4 of each). In that case, the null hypothesis was that she had no special ability, the test was Fisher's exact test, and the p-value was so Fisher was willing to reject the null hypothesis (consider the outcome highly unlikely to be due to chance) if all were classified correctly. (In the actual experiment, Bristol correctly classified all 8 cups.)

Fisher reiterated the p = 0.05 threshold and explained its rationale, stating:

It is usual and convenient for experimenters to take 5 per cent as a standard level of significance, in the sense that they are prepared to ignore all results which fail to reach this standard, and, by this means, to eliminate from further discussion the greater part of the fluctuations which chance causes have introduced into their experimental results.

He also applies this threshold to the design of experiments, noting that had only 6 cups been presented (3 of each), a perfect classification would have only yielded a p-value of which would not have met this level of significance. Fisher also underlined the interpretation of p, as the long-run proportion of values at least as extreme as the data, assuming the null hypothesis is true.

In later editions, Fisher explicitly contrasted the use of the p-value for statistical inference in science with the Neyman–Pearson method, which he terms "Acceptance Procedures". Fisher emphasizes that while fixed levels such as 5%, 2%, and 1% are convenient, the exact p-value can be used, and the strength of evidence can and will be revised with further experimentation. In contrast, decision procedures require a clear-cut decision, yielding an irreversible action, and the procedure is based on costs of error, which, he argues, are inapplicable to scientific research.

Related indices

The E-value corresponds to the expected number of times in multiple testing that one expects to obtain a test statistic at least as extreme as the one that was actually observed if one assumes that the null hypothesis is true. The E-value is the product of the number of tests and the p-value.

The q-value is the analog of the p-value with respect to the positive false discovery rate. It is used in multiple hypothesis testing to maintain statistical power while minimizing the false positive rate.

The Probability of Direction (pd) is the Bayesian numerical equivalent of the p-value. It corresponds to the proportion of the posterior distribution that is of the median's sign, typically varying between 50% and 100%, and representing the certainty with which an effect is positive or negative.

Operator (computer programming)

From Wikipedia, the free encyclopedia https://en.wikipedia.org/wiki/Operator_(computer_programmin...