Score Test

In statistics, the score test assesses constraints on statistical parameters based on the gradient of the likelihood function—known as the score—evaluated at the hypothesized parameter value under the null hypothesis.

Intuitively, if the restricted estimator is near the maximum of the likelihood function, the score should not differ from zero by more than sampling error. While the finite sample distributions of score tests are generally unknown, they have an asymptotic χ2-distribution under the null hypothesis as first proved by C. R. Rao in 1948, a fact that can be used to determine statistical significance.

Since function maximization subject to equality constraints is most conveniently done using a Lagrangean expression of the problem, the score test can be equivalently understood as a test of the magnitude of the Lagrange multipliers associated with the constraints where, again, if the constraints are non-binding at the maximum likelihood, the vector of Lagrange multipliers should not differ from zero by more than sampling error. The equivalence of these two approaches was first shown by S. D. Silvey in 1959, which led to the name Lagrange multiplier test that has become more commonly used, particularly in econometrics, since Breusch and Pagan's much-cited 1980 paper.

The main advantage of the score test over the Wald test and likelihood-ratio test is that the score test only requires the computation of the restricted estimator. This makes testing feasible when the unconstrained maximum likelihood estimate is a boundary point in the parameter space.[citation needed] Further, because the score test only requires the estimation of the likelihood function under the null hypothesis, it is less specific than the likelihood ratio test about the alternative hypothesis.

Single-parameter test

The statistic

Let Score Test  be the likelihood function which depends on a univariate parameter Score Test  and let Score Test  be the data. The score Score Test  is defined as

    Score Test 

The Fisher information is

    Score Test 

where ƒ is the probability density.

The statistic to test Score Test  is Score Test 

which has an asymptotic distribution of Score Test , when Score Test  is true. While asymptotically identical, calculating the LM statistic using the outer-gradient-product estimator of the Fisher information matrix can lead to bias in small samples.

Note on notation

Note that some texts use an alternative notation, in which the statistic Score Test  is tested against a normal distribution. This approach is equivalent and gives identical results.

As most powerful test for small deviations

    Score Test 

where Score Test  is the likelihood function, Score Test  is the value of the parameter of interest under the null hypothesis, and Score Test  is a constant set depending on the size of the test desired (i.e. the probability of rejecting Score Test  if Score Test  is true; see Type I error).

The score test is the most powerful test for small deviations from Score Test . To see this, consider testing Score Test  versus Score Test . By the Neyman–Pearson lemma, the most powerful test has the form

    Score Test 

Taking the log of both sides yields

    Score Test 

The score test follows making the substitution (by Taylor series expansion)

    Score Test 

and identifying the Score Test  above with Score Test .

Relationship with other hypothesis tests

If the null hypothesis is true, the likelihood ratio test, the Wald test, and the Score test are asymptotically equivalent tests of hypotheses. When testing nested models, the statistics for each test then converge to a Chi-squared distribution with degrees of freedom equal to the difference in degrees of freedom in the two models. If the null hypothesis is not true, however, the statistics converge to a noncentral chi-squared distribution with possibly different noncentrality parameters.

Multiple parameters

A more general score test can be derived when there is more than one parameter. Suppose that Score Test  is the maximum likelihood estimate of Score Test  under the null hypothesis Score Test  while Score Test  and Score Test  are respectively, the score vector and the Fisher information matrix. Then

    Score Test 

asymptotically under Score Test , where Score Test  is the number of constraints imposed by the null hypothesis and

    Score Test 

and

    Score Test 

This can be used to test Score Test .

The actual formula for the test statistic depends on which estimator of the Fisher information matrix is being used.

Special cases

In many situations, the score statistic reduces to another commonly used statistic.

In linear regression, the Lagrange multiplier test can be expressed as a function of the F-test.

When the data follows a normal distribution, the score statistic is the same as the t statistic.[clarification needed]

When the data consists of binary observations, the score statistic is the same as the chi-squared statistic in the Pearson's chi-squared test.

See also

References

Further reading

  • Buse, A. (1982). "The Likelihood Ratio, Wald, and Lagrange Multiplier Tests: An Expository Note". The American Statistician. 36 (3a): 153–157. doi:10.1080/00031305.1982.10482817.
  • Godfrey, L. G. (1988). "The Lagrange Multiplier Test and Testing for Misspecification : An Extended Analysis". Misspecification Tests in Econometrics. New York: Cambridge University Press. pp. 69–99. ISBN 0-521-26616-5.
  • Ma, Jun; Nelson, Charles R. (2016). "The superiority of the LM test in a class of econometric models where the Wald test performs poorly". Unobserved Components and Time Series Econometrics. Oxford University Press. pp. 310–330. doi:10.1093/acprof:oso/9780199683666.003.0014. ISBN 978-0-19-968366-6.
  • Rao, C. R. (2005). "Score Test: Historical Review and Recent Developments". Advances in Ranking and Selection, Multiple Comparisons, and Reliability. Boston: Birkhäuser. pp. 3–20. ISBN 978-0-8176-3232-8.

Tags:

Score Test Single-parameter testScore Test Multiple parametersScore Test Special casesScore Test Further readingScore TestC. R. RaoChi-squared distributionConstraint (mathematics)GradientLikelihood functionMaxima and minimaNull hypothesisSampling distributionSampling errorScore (statistics)Statistical parameterStatistical significanceStatistics

🔥 Trending searches on Wiki English:

Mike TysonRomelu LukakuJake PaulFuture (rapper)Sergey BrinSingaporeAaron HernandezMathias BoeYouTubeList of Royal Challengers Bangalore recordsAAlex JonesParakala PrabhakarThe Bear (TV series)Liu CixinPier Gerlofs DoniaNeatsville, KentuckyThe Dark ForestOlivia RodrigoMahatma GandhiList of presidents of the United StatesWes MooreRoyal Challengers BangaloreMichael JacksonCultural RevolutionAadhaarTaapsee PannuElon MuskElvis Presley2024 Formula One World ChampionshipKobbie MainooSiddharth (actor)Cosmo JarvisSocial mediaBradley CooperTom HardyWiki FoundationKwena MaphakaList of states and territories of the United StatesYorgos LanthimosRaindrop cakeJustin TimberlakeTiger WoodsCorfuNational September 11 Memorial & MuseumNickelodeonTilak VarmaThe BeatlesDeclan RiceShaquille O'NealHTTP 404MonsterVerse2024 in filmPaul AtreidesBarack ObamaU.S. stateTwitterOnce Upon a Time in HollywoodIndonesiaPrithviraj SukumaranMaster X MasterBlake LivelyLindsay Lohan2024 Indian general election in MaharashtraLuna Snow50 CentJustin BieberBob MarleyShane GillisKen MilesAngela ChaoPrince (musician)Amazon (company)Vinayak Damodar SavarkarPortugalAlexander the Great🡆 More