From Wikipedia, the free encyclopedia
  In
 statistics, every conjecture concerning the unknown probability 
distribution of a collection of random variables representing the 
observed data 
 in some study is called a statistical hypothesis.
 If we state one hypothesis only and the aim of the statistical test is 
to see whether this hypothesis is tenable, but not to investigate other 
specific hypotheses, then such a test is called a null hypothesis test.
As our statistical hypothesis will, by definition, state some property of the distribution, the null hypothesis
 is the default hypothesis under which that property does not exist. The
 null hypothesis is typically that some parameter (such as a correlation
 or a difference between means) in the populations of interest is zero. 
Note that our hypothesis might specify the probability distribution of 
 precisely, or it might only specify that it belongs to some class of 
distributions. Often, we reduce the data to a single numerical 
statistic, e.g., 
, whose marginal probability distribution is closely connected to a main question of interest in the study.
The p-value is used in the context of null hypothesis testing in order to quantify the statistical significance of a result, the result being the observed value of the chosen statistic 
.  The lower the p-value is, the lower the probability of getting that result if the null hypothesis were true. A result is said to be statistically significant
 if it allows us to reject the null hypothesis. All other things being 
equal, smaller p-values are taken as stronger evidence against the null 
hypothesis
Loosely speaking, rejection of the null hypothesis implies that there is sufficient evidence against it.
As a particular example, if a null hypothesis states that a certain summary statistic 
 follows the standard normal distribution N(0,1), then the rejection of this null hypothesis could mean that (i) the mean of 
 is not 0, or (ii) the variance of 
 is not 1, or (iii) 
 is not normally distributed. Different tests of the same null 
hypothesis would be more or less sensitive to different alternatives. 
However, even if we do manage to reject the null hypothesis for all 3 
alternatives, and even if we know the distribution is normal and 
variance is 1, the null hypothesis test does not tell us which non-zero 
values of the mean are now most plausible. The more independent 
observations from the same probability distribution one has, the more 
accurate the test will be, and the higher the precision with which one 
will be able to determine the mean value and show that it is not equal 
to zero; but this will also increase the importance of evaluating the 
real-world or scientific relevance of this deviation.
Definition and interpretation
General
Consider an observed test-statistic 
 from unknown distribution 
. Then the p-value 
 is what the prior probability would be of observing a test-statistic value at least as "extreme" as 
 if null hypothesis 
 were true. That is:
 for a one-sided right-tail test,
 for a one-sided left-tail test,
 for a two-sided test. If distribution 
 is symmetric about zero, then  
If the p-value is very small, then either the null hypothesis is false or something unlikely has occurred. In a formal significance test, the null hypothesis 
 is rejected if the p-value is less than a pre-defined threshold value 
, which is referred to as the alpha level or significance level. The value of 
 is instead set by the researcher before examining the data. 
 defines the proportion of the distribution, 
, that is said to define such a narrow range of all the possible outcomes of 
 that if 
's value is within that range its value is unlikely to have occurred by chance. Intuitively, this means that if 
 is set to be 0.10, only 1/10th of the distribution of 
 is defined by 
, so if 
 falls within that range it is already occurring over a number of 
outcomes that happen a rare 1/10th of the time, thus suggesting this is 
unlikely to occur randomly. By convention, 
 is commonly set to 0.05, though lower alpha levels are sometimes used.
The p-value is a function of the chosen test statistic 
 and is therefore a random variable. If the null hypothesis fixes the probability distribution of 
 precisely, and if that distribution is continuous, then when the 
null-hypothesis is true, the p-value is uniformly distributed between 0 
and 1. Thus, the p-value is not fixed. If the same test is 
repeated independently with fresh data (always with the same probability
 distribution), one will obtain a different p-value in each 
iteration. If the null-hypothesis is composite, or the distribution of 
the statistic is discrete, the probability of obtaining a p-value
 less than or equal to any number between 0 and 1 is less than or equal 
to that number, if the null-hypothesis is true. It remains the case that
 very small values are relatively unlikely if the null-hypothesis is 
true, and that a significance test at level 
 is obtained by rejecting the null-hypothesis if the significance level is less than or equal to 
.
Different p-values based on independent sets of data can be combined, for instance using Fisher's combined probability test.
Distribution
When the null hypothesis is true, if it takes the form 
, and the underlying random variable is continuous, then the probability distribution of the p-value is uniform
 on the interval [0,1]. By contrast, if the alternative hypothesis is 
true, the distribution is dependent on sample size and the true value of
 the parameter being studied.
The distribution of p-values for a group of studies is sometimes called a p-curve. A p-curve can be used to assess the reliability of scientific literature, such as by detecting publication bias or p-hacking.
For composite hypothesis
In parametric hypothesis testing problems, a simple or point hypothesis refers to a hypothesis where the parameter's value is assumed to be a single number. In contrast, in a composite hypothesis
 the parameter's value is given by a set of numbers. For example, when 
testing the null hypothesis that a distribution is normal with a mean 
less than or equal to zero against the alternative that the mean is 
greater than zero (variance known), the null hypothesis does not specify
 the probability distribution of the appropriate test statistic. In the 
just mentioned example that would be the Z-statistic belonging to the one-sided one-sample Z-test. For each possible value of the theoretical mean, the Z-test
 statistic has a different probability distribution. In these 
circumstances (the case of a so-called composite null hypothesis) the p-value
 is defined by taking the least favourable null-hypothesis case, which 
is typically on the border between null and alternative. 
This definition ensures the complementarity of p-values and 
alpha-levels. If we set the significance level alpha to 0.05, and only 
reject the null hypothesis if the p-value is less than or equal to 0.05,
 then our hypothesis test will indeed have significance level (maximal 
type 1 error rate) 0.05. As Neyman wrote: “The error that a practising 
statistician would consider the more important to avoid (which is a 
subjective judgment) is called the error of the first kind. The first 
demand of the mathematical theory is to deduce such test criteria as 
would ensure that the probability of committing an error of the first 
kind would equal (or approximately equal, or not exceed) a preassigned 
number α, such as α = 0.05 or 0.01, etc. This number is called the level
 of significance”; Neyman 1976, p. 161 in "The Emergence of Mathematical
 Statistics: A Historical Sketch with Particular Reference to the United
 States","On the History of Statistics and Probability", ed. D.B. Owen, 
New York: Marcel Dekker, pp. 149-193. See also "Confusion Over Measures 
of Evidence (p's) Versus Errors (a's) in Classical Statistical Testing",
 Raymond Hubbard and M. J. Bayarri,  The American Statistician, August 
2003, Vol. 57, No 3, 171--182 (with discussion). For a concise modern 
statement see Chapter 10 of "All of Statistics: A Concise Course in 
Statistical Inference", Springer; 1st Corrected ed. 20 edition 
(September 17, 2004). Larry Wasserman.
Usage
The p-value is widely used in statistical hypothesis testing,
 specifically in null hypothesis significance testing. In this method, 
before conducting the study, one first chooses a model (the null hypothesis) and the alpha level α (most commonly .05). After analyzing the data, if the p-value is less than α, that is taken to mean the observed data is sufficiently inconsistent with the null hypothesis for the null hypothesis to be rejected. However, that does not prove that the null hypothesis is false. The p-value
 does not, in itself, establish probabilities of hypotheses. Rather, it 
is a tool for deciding whether to reject the null hypothesis.
Misuse
According to the ASA, there is widespread agreement that p-values are often misused and misinterpreted. One practice that has been particularly criticized is accepting the alternative hypothesis for any p-value nominally less than .05 without other supporting evidence. Although p-values
 are helpful in assessing how incompatible the data are with a specified
 statistical model, contextual factors must also be considered, such as 
"the design of a study, the quality of the measurements, the external 
evidence for the phenomenon under study, and the validity of assumptions
 that underlie the data analysis". Another concern is that the p-value is often misunderstood as being the probability that the null hypothesis is true. 
Some statisticians have proposed abandoning p-values and focusing more on other inferential statistics, such as confidence intervals, likelihood ratios, or Bayes factors, but there is heated debate on the feasibility of these alternatives. Others have suggested to remove fixed significance thresholds and to interpret p-values as continuous indices of the strength of evidence against the null hypothesis. Yet others suggested to report alongside p-values the prior probability
 of a real effect that would be required to obtain a false positive risk
 (i.e. the probability that there is no real effect) below a 
pre-specified threshold (e.g. 5%).
Calculation
Usually, 
 is a test statistic. A test statistic is the output of a scalar
 function of all the observations. This statistic provides a single 
number, such as a t-statistic or an F-statistic. As such, the test 
statistic follows a distribution determined by the function used to 
define that test statistic and the distribution of the input 
observational data.
For the important case in which the data are hypothesized to be a
 random sample from a normal distribution, depending on the nature of 
the test statistic and the hypotheses of interest about its 
distribution, different null hypothesis tests have been developed. Some 
such tests are the z-test for hypotheses concerning the mean of a normal distribution with known variance, the t-test based on Student's t-distribution of a suitable statistic for hypotheses concerning the mean of a normal distribution when the variance is unknown, the F-test based on the F-distribution
 of yet another statistic for hypotheses concerning the variance. For 
data of other nature, for instance categorical (discrete) data, test 
statistics might be constructed whose null hypothesis distribution is 
based on normal approximations to appropriate statistics obtained by 
invoking the central limit theorem for large samples, as in the case of Pearson's chi-squared test.
Thus computing a p-value requires a null hypothesis, a test statistic (together with deciding whether the researcher is performing a one-tailed test or a two-tailed test),
 and data. Even though computing the test statistic on given data may be
 easy, computing the sampling distribution under the null hypothesis, 
and then computing its cumulative distribution function
 (CDF) is often a difficult problem. Today, this computation is done 
using statistical software, often via numeric methods (rather than exact
 formulae), but, in the early and mid 20th century, this was instead 
done via tables of values, and one interpolated or extrapolated p-values from these discrete values. Rather than using a table of p-values, Fisher instead inverted the CDF, publishing a list of values of the test statistic for given fixed p-values; this corresponds to computing the quantile function (inverse CDF).
Example
As an example of a statistical test, an experiment is performed to determine whether a coin flip is fair (equal chance of landing heads or tails) or unfairly biased (one outcome being more likely than the other).
Suppose that the experimental results show the coin turning up heads 14 times out of 20 total flips. The full data 
 would be a sequence of twenty times the symbol "H" or "T". The statistic on which one might focus, could be the total number 
 of heads. The null hypothesis is that the coin is fair, and coin tosses
 are independent of one another. If a right-tailed test is considered, 
which would be the case if one is actually interested in the possibility
 that the coin is biased towards falling heads, then the p-value of this result is the chance of a fair coin landing on heads at least 14 times out of 20 flips. That probability can be computed from binomial coefficients as
![{\displaystyle {\begin{aligned}&\Pr(14{\text{ heads}})+\Pr(15{\text{ heads}})+\cdots +\Pr(20{\text{ heads}})\\&={\frac {1}{2^{20}}}\left[{\binom {20}{14}}+{\binom {20}{15}}+\cdots +{\binom {20}{20}}\right]={\frac {60,\!460}{1,\!048,\!576}}\approx 0.058\end{aligned}}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/f2c59548d359fb637177e73515aea16368929436)
This probability is the p-value, considering only extreme results that favor heads. This is called a one-tailed test. However, one might be interested in deviations in either direction, favoring either heads or tails. The two-tailed p-value, which considers deviations favoring either heads or tails, may instead be calculated. As the binomial distribution is symmetrical for a fair coin, the two-sided p-value is simply twice the above calculated single-sided p-value: the two-sided p-value is 0.115.
In the above example:
- Null hypothesis (H0): The coin is fair, with Pr(heads) = 0.5
 - Test statistic: Number of heads
 - Alpha level (designated threshold of significance): 0.05
 - Observation O: 14 heads out of 20 flips; and
 - Two-tailed p-value of observation O given H0 = 2 × min(Pr(no. of heads ≥ 14 heads), Pr(no. of heads ≤ 14 heads)) = 2 × min(0.058, 0.978) = 2*0.058 = 0.115.
 
Note that the Pr (no. of heads ≤ 14 heads) = 1 - Pr(no. of heads 
≥ 14 heads) + Pr (no. of head = 14) = 1 - 0.058 + 0.036 = 0.978; 
however, symmetry of the binomial distribution makes it an unnecessary 
computation to find the smaller of the two probabilities. Here, the 
calculated p-value exceeds .05, meaning that the data falls 
within the range of what would happen 95% of the time were the coin in 
fact fair. Hence, the null hypothesis is not rejected at the .05 level.
However, had one more head been obtained, the resulting p-value (two-tailed) would have been 0.0414 (4.14%), in which case the null hypothesis would be rejected at the .05 level.
History
   
Computations of p-values date back to the 1700s, where they were computed for the human sex ratio
 at birth, and used to compute statistical significance compared to the 
null hypothesis of equal probability of male and female births. John Arbuthnot studied this question in 1710,
 and examined birth records in London for each of the 82 years from 1629
 to 1710. In every year, the number of males born in London exceeded the
 number of females. Considering more male or more female births as 
equally likely, the probability of the observed outcome is 1/282, or about 1 in 4,836,000,000,000,000,000,000,000; in modern terms, the p-value.
 This is vanishingly small, leading Arbuthnot that this was not due to 
chance, but to divine providence: "From whence it follows, that it is 
Art, not Chance, that governs." In modern terms, he rejected the null 
hypothesis of equally likely male and female births at the p = 1/282 significance level. This and other work by Arbuthnot is credited as "… the first use of significance tests …" the first example of reasoning about statistical significance, and "… perhaps the first published report of a nonparametric test …", specifically the sign test; see details at Sign test § History.
The same question was later addressed by Pierre-Simon Laplace, who instead used a parametric test, modeling the number of male births with a binomial distribution:
In the 1770s Laplace considered the
 statistics of almost half a million births. The statistics showed an 
excess of boys compared to girls. He concluded by calculation of a p-value that the excess was a real, but unexplained, effect.
The p-value was first formally introduced by Karl Pearson, in his Pearson's chi-squared test, using the chi-squared distribution and notated as capital P. The p-values for the chi-squared distribution (for various values of χ2 and degrees of freedom), now notated as P, were calculated in (Elderton 1902), collected in (Pearson 1914, pp. xxxi–xxxiii, 26–28, Table XII).
The use of the p-value in statistics was popularized by Ronald Fisher, and it plays a central role in his approach to the subject. In his influential book Statistical Methods for Research Workers (1925), Fisher proposed the level p = 0.05, or a 1 in 20 chance of being exceeded by chance, as a limit for statistical significance,
 and applied this to a normal distribution (as a two-tailed test), thus 
yielding the rule of two standard deviations (on a normal distribution) 
for statistical significance (see 68–95–99.7 rule).
He then computed a table of values, similar to Elderton but, importantly, reversed the roles of χ2 and p. That is, rather than computing p for different values of χ2 (and degrees of freedom n), he computed values of χ2 that yield specified p-values, specifically 0.99, 0.98, 0.95, 0,90, 0.80, 0.70, 0.50, 0.30, 0.20, 0.10, 0.05, 0.02, and 0.01. That allowed computed values of χ2 to be compared against cutoffs and encouraged the use of p-values (especially 0.05, 0.02, and 0.01) as cutoffs, instead of computing and reporting p-values themselves. The same type of tables were then compiled in (Fisher & Yates 1938), which cemented the approach.
As an illustration of the application of p-values to the design and interpretation of experiments, in his following book The Design of Experiments (1935), Fisher presented the lady tasting tea experiment, which is the archetypal example of the p-value.
To evaluate a lady's claim that she (Muriel Bristol)
 could distinguish by taste how tea is prepared (first adding the milk 
to the cup, then the tea, or first tea, then milk), she was sequentially
 presented with 8 cups: 4 prepared one way, 4 prepared the other, and 
asked to determine the preparation of each cup (knowing that there were 4
 of each). In that case, the null hypothesis was that she had no special
 ability, the test was Fisher's exact test, and the p-value was 
 so Fisher was willing to reject the null hypothesis (consider the 
outcome highly unlikely to be due to chance) if all were classified 
correctly. (In the actual experiment, Bristol correctly classified all 8
 cups.)
Fisher reiterated the p = 0.05 threshold and explained its rationale, stating:
It is usual and convenient for 
experimenters to take 5 per cent as a standard level of significance, in
 the sense that they are prepared to ignore all results which fail to 
reach this standard, and, by this means, to eliminate from further 
discussion the greater part of the fluctuations which chance causes have
 introduced into their experimental results.
He also applies this threshold to the design of experiments, noting 
that had only 6 cups been presented (3 of each), a perfect 
classification would have only yielded a p-value of 
 which would not have met this level of significance. Fisher also underlined the interpretation of p, as the long-run proportion of values at least as extreme as the data, assuming the null hypothesis is true.
In later editions, Fisher explicitly contrasted the use of the p-value for statistical inference in science with the Neyman–Pearson method, which he terms "Acceptance Procedures". Fisher emphasizes that while fixed levels such as 5%, 2%, and 1% are convenient, the exact p-value
 can be used, and the strength of evidence can and will be revised with 
further experimentation. In contrast, decision procedures require a 
clear-cut decision, yielding an irreversible action, and the procedure 
is based on costs of error, which, he argues, are inapplicable to 
scientific research.
Related indices
The E-value corresponds to the expected number of times in multiple testing
 that one expects to obtain a test statistic at least as extreme as the 
one that was actually observed if one assumes that the null hypothesis 
is true. The E-value is the product of the number of tests and the p-value.
The q-value is the analog of the p-value with respect to the positive false discovery rate. It is used in multiple hypothesis testing to maintain statistical power while minimizing the false positive rate.
The Probability of Direction (pd) is the  Bayesian numerical equivalent of the p-value. It corresponds to the proportion of the  posterior distribution
 that is of the median's sign, typically varying between 50% and 100%, 
and representing the certainty with which an effect is positive or 
negative.