Test Statistic

Test statistic is a quantity derived from the sample for statistical hypothesis testing.

A hypothesis test is typically specified in terms of a test statistic, considered as a numerical summary of a data-set that reduces the data to one value that can be used to perform the hypothesis test. In general, a test statistic is selected or defined in such a way as to quantify, within observed data, behaviours that would distinguish the null from the alternative hypothesis, where such an alternative is prescribed, or that would characterize the null hypothesis if there is no explicitly stated alternative hypothesis.

Test Statistic
The above image shows a table with some of the most common test statistics and their corresponding statistical tests or models.

An important property of a test statistic is that its sampling distribution under the null hypothesis must be calculable, either exactly or approximately, which allows p-values to be calculated. A test statistic shares some of the same qualities of a descriptive statistic, and many statistics can be used as both test statistics and descriptive statistics. However, a test statistic is specifically intended for use in statistical testing, whereas the main quality of a descriptive statistic is that it is easily interpretable. Some informative descriptive statistics, such as the sample range, do not make good test statistics since it is difficult to determine their sampling distribution.

Two widely used test statistics are the t-statistic and the F-test.

Example

Suppose the task is to test whether a coin is fair (i.e. has equal probabilities of producing a head or a tail). If the coin is flipped 100 times and the results are recorded, the raw data can be represented as a sequence of 100 heads and tails. If there is interest in the marginal probability of obtaining a tail, only the number T out of the 100 flips that produced a tail needs to be recorded. But T can also be used as a test statistic in one of two ways:

  • the exact sampling distribution of T under the null hypothesis is the binomial distribution with parameters 0.5 and 100.
  • the value of T can be compared with its expected value under the null hypothesis of 50, and since the sample size is large, a normal distribution can be used as an approximation to the sampling distribution either for T or for the revised test statistic T−50.

Using one of these sampling distributions, it is possible to compute either a one-tailed or two-tailed p-value for the null hypothesis that the coin is fair. The test statistic in this case reduces a set of 100 numbers to a single numerical summary that can be used for testing.

Common test statistics

One-sample tests are appropriate when a sample is being compared to the population from a hypothesis. The population characteristics are known from theory or are calculated from the population.

Two-sample tests are appropriate for comparing two samples, typically experimental and control samples from a scientifically controlled experiment.

Paired tests are appropriate for comparing two samples where it is impossible to control important variables. Rather than comparing two sets, members are paired between samples so the difference between the members becomes the sample. Typically the mean of the differences is then compared to zero. The common example scenario for when a paired difference test is appropriate is when a single set of test subjects has something applied to them and the test is intended to check for an effect.

Z-tests are appropriate for comparing means under stringent conditions regarding normality and a known standard deviation.

A t-test is appropriate for comparing means under relaxed conditions (less is assumed).

Tests of proportions are analogous to tests of means (the 50% proportion).

Chi-squared tests use the same calculations and the same probability distribution for different applications:

  • Chi-squared tests for variance are used to determine whether a normal population has a specified variance. The null hypothesis is that it does.
  • Chi-squared tests of independence are used for deciding whether two variables are associated or are independent. The variables are categorical rather than numeric. It can be used to decide whether left-handedness is correlated with height (or not). The null hypothesis is that the variables are independent. The numbers used in the calculation are the observed and expected frequencies of occurrence (from contingency tables).
  • Chi-squared goodness of fit tests are used to determine the adequacy of curves fit to data. The null hypothesis is that the curve fit is adequate. It is common to determine curve shapes to minimize the mean square error, so it is appropriate that the goodness-of-fit calculation sums the squared errors.

F-tests (analysis of variance, ANOVA) are commonly used when deciding whether groupings of data by category are meaningful. If the variance of test scores of the left-handed in a class is much smaller than the variance of the whole class, then it may be useful to study lefties as a group. The null hypothesis is that two variances are the same – so the proposed grouping is not meaningful.

In the table below, the symbols used are defined at the bottom of the table. Many other tests can be found in other articles. Proofs exist that the test statistics are appropriate.

Name Formula Assumptions or notes
One-sample Test Statistic  -test Test Statistic  (Normal population or n large) and σ known.

(z is the distance from the mean in relation to the standard deviation of the mean). For non-normal distributions it is possible to calculate a minimum proportion of a population that falls within k standard deviations for any k (see: Chebyshev's inequality).

Two-sample z-test Test Statistic  Normal population and independent observations and σ1 and σ2 are known where Test Statistic  is the value of Test Statistic  under the null hypothesis
One-sample t-test Test Statistic 
Test Statistic 
(Normal population or n large) and Test Statistic  unknown
Paired t-test Test Statistic 

Test Statistic 

(Normal population of differences or n large) and Test Statistic  unknown
Two-sample pooled t-test, equal variances Test Statistic 

Test Statistic 
Test Statistic 

(Normal populations or n1 + n2 > 40) and independent observations and σ1 = σ2 unknown
Two-sample unpooled t-test, unequal variances (Welch's t-test) Test Statistic 

Test Statistic 

(Normal populations or n1 + n2 > 40) and independent observations and σ1 ≠ σ2 both unknown
One-proportion z-test Test Statistic  n .p0 > 10 and n (1 − p0) > 10 and it is a SRS (Simple Random Sample), see notes.
Two-proportion z-test, pooled for Test Statistic  Test Statistic 

Test Statistic 

n1 p1 > 5 and n1(1 − p1) > 5 and n2 p2 > 5 and n2(1 − p2) > 5 and independent observations, see notes.
Two-proportion z-test, unpooled for Test Statistic  Test Statistic  n1 p1 > 5 and n1(1 − p1) > 5 and n2 p2 > 5 and n2(1 − p2) > 5 and independent observations, see notes.
Chi-squared test for variance Test Statistic  df = n-1

• Normal population

Chi-squared test for goodness of fit Test Statistic  df = k − 1 − # parameters estimated, and one of these must hold.

• All expected counts are at least 5.

• All expected counts are > 1 and no more than 20% of expected counts are less than 5

Two-sample F test for equality of variances Test Statistic  Normal populations
Arrange so Test Statistic  and reject H0 for Test Statistic 
Regression t-test of Test Statistic  Test Statistic  Reject H0 for Test Statistic 
*Subtract 1 for intercept; k terms contain independent variables.
In general, the subscript 0 indicates a value taken from the null hypothesis, H0, which should be used as much as possible in constructing its test statistic. ... Definitions of other symbols:
  • Test Statistic  = sample variance
  • Test Statistic  = sample 1 standard deviation
  • Test Statistic  = sample 2 standard deviation
  • Test Statistic  = t statistic
  • Test Statistic  = degrees of freedom
  • Test Statistic  = sample mean of differences
  • Test Statistic  = hypothesized population mean difference
  • Test Statistic  = standard deviation of differences
  • Test Statistic  = Chi-squared statistic
  • Test Statistic  = sample proportion, unless specified otherwise
  • Test Statistic  = hypothesized population proportion
  • Test Statistic  = proportion 1
  • Test Statistic  = proportion 2
  • Test Statistic  = hypothesized difference in proportion
  • Test Statistic  = minimum of Test Statistic  and Test Statistic 
  • Test Statistic 
  • Test Statistic 
  • Test Statistic  = F statistic

See also

References

Tags:

Test Statistic ExampleTest Statistic Common test statisticsTest StatisticAlternative hypothesisNull hypothesisSample (statistics)StatisticStatistical hypothesis testing

🔥 Trending searches on Wiki English:

Emma CorrinList of countries by GDP (nominal)Arthur the KingMoonRichard Armitage (actor)Dave BautistaItalyThe BeatlesHugh JackmanCanvaBernard ArnaultCD-ROMRonald ReaganMark ZuckerbergMurder trial of O. J. SimpsonAngelina JolieJennifer PanDrop bearPassoverCryptocurrencyAnunnakiDhruv RatheeMegan Thee StallionMercury SevenList of United States cities by populationSingaporeX (2022 film)2019 Indian general electionMichael DouglasChappell RoanOracle CorporationTom Cleverley2023–24 AFC Champions LeagueNicole Mitchell (meteorologist)Tom CruiseTemperatureList of prime ministers of IndiaJohn CenaIker MuniainXNXXJoe BidenAndrew TateShou Zi ChewVladimir PutinApril 24Brad PittThe Zone of Interest (film)Charles IIIAnne HathawayKim Ji-won (actress)Heart (band)Kaya Scodelario2024 World Snooker ChampionshipJoe ManganielloErin MoranDenzel WashingtonRonnie O'SullivanShōgun (1980 miniseries)Armenian Genocide Remembrance DayReggie BushFloyd Mayweather Jr.ShogunLina KhanList of James Bond filmsGermanyShaquille O'NealMarlon BrandoJason KiddFallout 4Candidates Tournament 2024Anne HecheAngus CloudXXX (2002 film)Sexual intercourseTupac ShakurMin Hee-jin🡆 More