Electric dipole transition: Difference between revisions
No edit summary |
en>ChrisGualtieri m General Fixes using AWB |
||
Line 1: | Line 1: | ||
{{DISPLAYTITLE:Generalized ''p''-value}} | |||
{{Expert-subject|statistics|date=March 2011}} | |||
{{Primary sources|date=October 2010}} | |||
In [[statistics]], a '''generalized ''p''-value''' is an extended version of the classical [[p-value|''p''-value]], which except in a limited number of applications, provides only approximate solutions. | |||
Conventional statistical methods do not provide exact solutions to many statistical problems, such as those arising in [[mixed model]]s and MANOVA, especially when the problem involves many [[nuisance parameter]]s. As a result, practitioners often resort to approximate statistical methods or [[Asymptotic theory (statistics)|asymptotic statistical methods]] that are valid only with large samples. With small samples, such methods often have poor performance.<ref name=WE/> Use of approximate and asymptotic methods may lead to misleading conclusions or may fail to detect truly [[Statistical significance|significant]] results from [[experiment]]s. | |||
Tests based on generalized ''p''-values are exact statistical methods in that they are based on exact probability statements. While conventional statistical methods do not provide exact solutions to such problems as testing [[variance components]] or [[ANOVA]] under unequal variances, exact tests for such problems can be obtained based on generalized ''p''-values.<ref name=WE>Weerahandi (1995)</ref><ref name=TW>Tsui & Weerahandi (1989)</ref> | |||
In order to overcome the shortcomings of the classical ''p''-values, Tsui and Weerahandi<ref name=TW/> extended the classical definition so that one can obtain exact solutions for such problems as the [[Behrens–Fisher problem]] and testing variance components. This is accomplished by allowing test variables to depend on observable random vectors as well as their observed values, as in the Bayesian treatment of the problem, but without having to treat constant parameters as random variables. | |||
==A simple case== | |||
To describe the idea of generalized ''p''-values in a simple example, consider a situation of sampling from a normal population with mean <math>\mu</math>, and variance <math>\sigma ^2</math>, suppose <math>\overline{X}</math> and <math>S ^2</math> are the sample mean and the sample variance. Inferences on all unknown parameters can be based on the distributional results | |||
:<math> Z = \sqrt{n}(\overline{X} - \mu)/ \sigma \sim N(0,1)</math> | |||
and | |||
:<math>U = n S^2 / \sigma^2 \sim \chi^2 _ {n-1} .</math> | |||
Now suppose we need to test the coefficient of variation, <math>\rho = \mu /\sigma </math>. While the problem is not trivial with conventional ''p''-values, the task can be easily accomplished based on the generalized test variable | |||
:<math>R = \frac {\overline{x} S} {s \sigma} - \frac{\overline{X}- \mu} {\sigma} | |||
= \frac {\overline{x}} {s} \frac {\sqrt{U}} {\sqrt{n}} ~-~ \frac {Z} {\sqrt{n}} ,</math> | |||
where <math>\overline{x}</math> is the observed value of <math>\overline{X}</math> and <math>s</math> is the observed value of <math>S</math>. Note that the distribution of <math>R</math> and its observed value are both free of nuisance parameters. Therefore, a test of a hypothesis with a one-sided alternative such as <math> H_A : \rho < \rho_0 </math> can be based on the generalized ''p''-value <math> p = Pr( R \ge \rho_0 )</math>, a quantity that can be evaluated via Monte Carlo simulation or using the non-central t-distribution. | |||
==Notes== | |||
{{Reflist}} | |||
==References== | |||
*Tsui, K. and Weerahandi, S. (1989): [http://www.jstor.org/stable/2289949 "Generalized ''p''-values in significance testing of hypotheses in the presence of nuisance parameters"]. ''[[Journal of the American Statistical Association]]'', 84, 602–607 | |||
*Weerahandi, S. (1995) [http://www.springer.com/statistics/statistical+theory+and+methods/book/978-0-387-40621-3 ''Exact Statistical Methods for Data Analysis'' ] Springer-Verlag, New York. ISBN 978-0-387-40621-3 | |||
==External links== | |||
*[http://www.x-techniques.com/ XPro, Free software package for exact parametric statistics] | |||
{{DEFAULTSORT:Generalized P-Value}} | |||
[[Category:Hypothesis testing]] |
Revision as of 01:15, 26 October 2013
Template:Expert-subject Template:Primary sources In statistics, a generalized p-value is an extended version of the classical p-value, which except in a limited number of applications, provides only approximate solutions.
Conventional statistical methods do not provide exact solutions to many statistical problems, such as those arising in mixed models and MANOVA, especially when the problem involves many nuisance parameters. As a result, practitioners often resort to approximate statistical methods or asymptotic statistical methods that are valid only with large samples. With small samples, such methods often have poor performance.[1] Use of approximate and asymptotic methods may lead to misleading conclusions or may fail to detect truly significant results from experiments.
Tests based on generalized p-values are exact statistical methods in that they are based on exact probability statements. While conventional statistical methods do not provide exact solutions to such problems as testing variance components or ANOVA under unequal variances, exact tests for such problems can be obtained based on generalized p-values.[1][2]
In order to overcome the shortcomings of the classical p-values, Tsui and Weerahandi[2] extended the classical definition so that one can obtain exact solutions for such problems as the Behrens–Fisher problem and testing variance components. This is accomplished by allowing test variables to depend on observable random vectors as well as their observed values, as in the Bayesian treatment of the problem, but without having to treat constant parameters as random variables.
A simple case
To describe the idea of generalized p-values in a simple example, consider a situation of sampling from a normal population with mean , and variance , suppose and are the sample mean and the sample variance. Inferences on all unknown parameters can be based on the distributional results
and
Now suppose we need to test the coefficient of variation, . While the problem is not trivial with conventional p-values, the task can be easily accomplished based on the generalized test variable
where is the observed value of and is the observed value of . Note that the distribution of and its observed value are both free of nuisance parameters. Therefore, a test of a hypothesis with a one-sided alternative such as can be based on the generalized p-value , a quantity that can be evaluated via Monte Carlo simulation or using the non-central t-distribution.
Notes
43 year old Petroleum Engineer Harry from Deep River, usually spends time with hobbies and interests like renting movies, property developers in singapore new condominium and vehicle racing. Constantly enjoys going to destinations like Camino Real de Tierra Adentro.
References
- Tsui, K. and Weerahandi, S. (1989): "Generalized p-values in significance testing of hypotheses in the presence of nuisance parameters". Journal of the American Statistical Association, 84, 602–607
- Weerahandi, S. (1995) Exact Statistical Methods for Data Analysis Springer-Verlag, New York. ISBN 978-0-387-40621-3