Score test
Rao's score test, also known as the score test or the Lagrange multiplier test (LM test) in econometrics,[1][2] is a statistical test of a simple null hypothesis that a parameter of interest is equal to some particular value . It is the most powerful test when the true value of is close to . The main advantage of the score test is that it does not require an estimate of the information under the alternative hypothesis or unconstrained maximum likelihood. This constitutes a potential advantage in comparison to other tests, such as the Wald test and the generalized likelihood ratio test (GLRT). This makes testing feasible when the unconstrained maximum likelihood estimate is a boundary point in the parameter space.
Single parameter test
The statistic
Let be the likelihood function which depends on a univariate parameter and let be the data. The score is defined as
- Failed to parse (syntax error): {\displaystyle U(\theta)=\frac{\partial \log L(\theta \mid x)}{\partial \left[\left. \frac{\partial^2}{\partial\theta^2} \log L(X;\theta)\right|\theta \right]\,. }
The statistic to test is
which has an asymptotic distribution of , when is true.
Note on notation
Note that some texts use an alternative notation, in which the statistic is tested against a normal distribution. This approach is equivalent and gives identical results.
Justification
![]() | This section needs expansion. You can help by adding to it. (June 2008) |
The case of a likelihood with nuisance parameters
![]() | This section is empty. You can help by adding to it. (June 2008) |
As most powerful test for small deviations
Where is the likelihood function, is the value of the parameter of interest under the null hypothesis, and is a constant set depending on the size of the test desired (i.e. the probability of rejecting if is true; see Type I error).
The score test is the most powerful test for small deviations from . To see this, consider testing versus . By the Neyman–Pearson lemma, the most powerful test has the form
Taking the log of both sides yields
The score test follows making the substitution (by Taylor series expansion)
- is the maximum likelihood estimate of under the null hypothesis while and are respectively, the score and the Fisher information matrices under the alternative hypothesis. Then
asymptotically under , where is the number of constraints imposed by the null hypothesis and
and
This can be used to test .
Special cases
In many situations, the score statistic reduces to another commonly used statistic.[3]
When the data follows a normal distribution, the score statistic is the same as the t statistic.[clarification needed]
When the data consists of binary observations, the score statistic is the same as the chi-squared statistic in the Pearson's chi-squared test.
When the data consists of failure time data in two groups, the score statistic for the Cox partial likelihood is the same as the log-rank statistic in the log-rank test. Hence the log-rank test for difference in survival between two groups is most powerful when the proportional hazards assumption holds.
See also
References
- ^ Bera, Anil K.; Bilias, Yannis (2001). "Rao's score, Neyman's C(α) and Silvey's LM tests: An essay on historical developments and some new results". Journal of Statistical Planning and Inference. 97: 9–44. doi:10.1016/S0378-3758(00)00343-8.
- ^ Engle, Robert F. (1984). "Chapter 13: Wald, likelihood ratio, and Lagrange multiplier tests in econometrics". Handbook of Econometrics. 2: 775–826. doi:10.1016/S1573-4412(84)02005-5.
- ^ Cook, T. D.; DeMets, D. L., eds. (2007). Introduction to Statistical Methods for Clinical Trials. Chapman and Hall. pp. 296–297. ISBN 1-58488-027-9.