Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Jump to content

Behrens–Fisher distribution: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Citation bot (talk | contribs)
Add: s2cid, authors 1-1. Removed parameters. Some additions/deletions were parameter name changes. | Use this bot. Report bugs. | Suggested by SemperIocundus | #UCB_webform 1719/2500
 
(28 intermediate revisions by 22 users not shown)
Line 1: Line 1:
{{Short description|Probability distribution}}
In [[statistics]], the '''Behrens–Fisher distribution''', named after [[Ronald Fisher]] and [[W. V. Behrens]], is a [[parameter]]ized family of [[probability distribution]]s arising from the solution of the [[Behrens–Fisher problem]] proposed first by Behrens and several years later by Fisher. The Behrens–Fisher problem is that of [[statistical inference]] concerning the difference between the means of two [[normal distribution|normally distributed]] [[Statistical population|populations]] when the [[ratio]] of their [[variance]]s is not known (and in particular, it is not known that their variances are equal).
{{more citations needed|date=July 2022}}

In [[statistics]], the '''Behrens–Fisher distribution''', named after [[Ronald Fisher]] and [[Walter Behrens (statistician)|Walter Behrens]], is a [[parameter]]ized family of [[probability distribution]]s arising from the solution of the [[Behrens–Fisher problem]] proposed first by Behrens and several years later by Fisher. The Behrens–Fisher problem is that of [[statistical inference]] concerning the difference between the means of two [[normal distribution|normally distributed]] [[Statistical population|populations]] when the [[ratio]] of their [[variance]]s is not known (and in particular, it is not known that their variances are equal).<ref>{{Cite journal |last1=Kim |first1=Seock-Ho |last2=Cohen |first2=Allan S. |date=December 1998 |title=On the Behrens-Fisher Problem: A Review |url=http://journals.sagepub.com/doi/10.3102/10769986023004356 |journal=Journal of Educational and Behavioral Statistics |language=en |volume=23 |issue=4 |pages=356–377 |doi=10.3102/10769986023004356 |s2cid=85462934 |issn=1076-9986}}</ref>


== Definition ==
== Definition ==
Line 7: Line 10:
: <math> T_2 \cos\theta - T_1\sin\theta \, </math>
: <math> T_2 \cos\theta - T_1\sin\theta \, </math>


where ''T''<sub>1</sub> and ''T''<sub>2</sub> are [[independence (probability theory)|independent]] [[random variable]]s each with a Student's [[t-distribution]], with respective degrees of freedom ''&nu;''<sub>1</sub>&nbsp;=&nbsp;''n''<sub>1</sub>&nbsp;&minus;&nbsp;1 and ''&nu;''<sub>2</sub>&nbsp;=&nbsp;''n''<sub>2</sub>&nbsp;&minus;&nbsp;1, and ''&theta;'' is a constant. Thus the family of Behrens–Fisher distributions is parametrized by ''&nu;''<sub>1</sub>,&nbsp;''&nu;''<sub>2</sub>,&nbsp;and&nbsp;''&theta;''.
where ''T''<sub>1</sub> and ''T''<sub>2</sub> are [[independence (probability theory)|independent]] [[random variable]]s each with a Student's [[Student's t-distribution|t-distribution]], with respective degrees of freedom ''&nu;''<sub>1</sub>&nbsp;=&nbsp;''n''<sub>1</sub>&nbsp;&minus;&nbsp;1 and ''&nu;''<sub>2</sub>&nbsp;=&nbsp;''n''<sub>2</sub>&nbsp;&minus;&nbsp;1, and ''&theta;'' is a constant. Thus the family of Behrens–Fisher distributions is parametrized by ''&nu;''<sub>1</sub>,&nbsp;''&nu;''<sub>2</sub>,&nbsp;and&nbsp;''&theta;''.


== Derivation ==
== Derivation ==
Line 39: Line 42:
: <math> \frac{(\mu_2-\mu_1)-(\bar X_2 - \bar X_1)}{\displaystyle\sqrt{\frac{S^2_\mathrm{pooled}}{n_1} + \frac{S^2_\mathrm{pooled}}{n_2} }} </math>
: <math> \frac{(\mu_2-\mu_1)-(\bar X_2 - \bar X_1)}{\displaystyle\sqrt{\frac{S^2_\mathrm{pooled}}{n_1} + \frac{S^2_\mathrm{pooled}}{n_2} }} </math>


has a [[t-distribution]] with ''n''<sub>1</sub>&nbsp;+&nbsp;''n''<sub>2</sub>&nbsp;&minus;&nbsp;2 [[Degrees of freedom (statistics)|degrees of freedom]]. Accordingly, one can find a [[confidence interval]] for ''&mu;''<sub>2</sub>&nbsp;&minus;&nbsp;''&mu;''<sub>1</sub> whose endpoints are
has a [[Student's t-distribution|t-distribution]] with ''n''<sub>1</sub>&nbsp;+&nbsp;''n''<sub>2</sub>&nbsp;&minus;&nbsp;2 [[Degrees of freedom (statistics)|degrees of freedom]]. Accordingly, one can find a [[confidence interval]] for ''&mu;''<sub>2</sub>&nbsp;&minus;&nbsp;''&mu;''<sub>1</sub> whose endpoints are


: <math> \bar{X}_2 - \bar{X_1} \pm A \cdot S_\mathrm{pooled} \sqrt{\frac{1}{n_1} +\frac{1}{n_2}}, </math>
: <math> \bar{X}_2 - \bar{X_1} \pm A \cdot S_\mathrm{pooled} \sqrt{\frac{1}{n_1} +\frac{1}{n_2}}, </math>


where ''A'' is an appropriate percentage point of the t-distribution.
where ''A'' is an appropriate quantile of the t-distribution.


However, in the Behrens–Fisher problem, the two population variances are not known to be equal, nor is their ratio known. Fisher considered{{fact}} the pivotal quantity
However, in the Behrens–Fisher problem, the two population variances are not known to be equal, nor is their ratio known. Fisher considered{{citation needed|date=August 2012}} the pivotal quantity


: <math> \frac{(\mu_2-\mu_1)-(\bar X_2 - \bar X_1)}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }}. </math>
: <math> \frac{(\mu_2-\mu_1)-(\bar X_2 - \bar X_1)}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }}. </math>
Line 66: Line 69:
\begin{align}
\begin{align}
\frac{(\mu_2-\mu_1)-(\bar X_2 - \bar X_1)}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} & = \frac{\mu_2-\bar{X}_2}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} - \frac{\mu_1-\bar{X}_1}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} \\[10pt]
\frac{(\mu_2-\mu_1)-(\bar X_2 - \bar X_1)}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} & = \frac{\mu_2-\bar{X}_2}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} - \frac{\mu_1-\bar{X}_1}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} \\[10pt]
& = \underbrace{\frac{\mu_2-\bar{X}_2}{S_2/\sqrt{n_2}}}_{\text{This is }T_2} \cdot \underbrace{\left( \frac{S_2/\sqrt{n_2}}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} \right)}_{\text{This is }\cos\theta} - \underbrace{\frac{\mu_1-\bar{X}_1}{S_1/\sqrt{n_1}}}_{\text{This is }T_1}\cdot\underbrace{\left( \frac{S_1/\sqrt{n_1}}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} \right)}_{\text{This is }\sin\theta}.
& = \underbrace{\frac{\mu_2-\bar{X}_2}{S_2/\sqrt{n_2}}}_{\text{This is }T_2} \cdot \underbrace{\left( \frac{S_2/\sqrt{n_2}}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} \right)}_{\text{This is }\cos\theta} - \underbrace{\frac{\mu_1-\bar{X}_1}{S_1/\sqrt{n_1}}}_{\text{This is }T_1}\cdot\underbrace{\left( \frac{S_1/\sqrt{n_1}}{\displaystyle\sqrt{\frac{S^2_1}{n_1} + \frac{S^2_2}{n_2} }} \right)}_{\text{This is }\sin\theta}.\qquad\qquad\qquad (1)
\end{align}
\end{align}
</math>
</math>


The fact that the sum of the squares of the expressions in parentheses above is&nbsp;1 implies that they are the cosine and sine of some angle.
The fact that the sum of the squares of the expressions in parentheses above is&nbsp;1 implies that they are the squared cosine and squared sine of some angle.

The Behren–Fisher distribution is actually the [[conditional distribution]] of the quantity (1) above, ''given'' the values of the quantities labeled cos&nbsp;''&theta;'' and sin&nbsp;''&theta;''. In effect, Fisher [[conditioning on an ancillary statistic|conditions on ancillary information]].


Fisher then found the "[[fiducial inference|fiducial]] interval" whose endpoints are
Fisher then found the "[[fiducial inference|fiducial]] interval" whose endpoints are
Line 76: Line 81:
: <math> \bar{X}_2-\bar{X}_1 \pm A \sqrt{\frac{S_1^2}{n_1} + \frac{S_2^2}{n_2} } </math>
: <math> \bar{X}_2-\bar{X}_1 \pm A \sqrt{\frac{S_1^2}{n_1} + \frac{S_2^2}{n_2} } </math>


where ''A'' is the appropriate percentage point of the Behrens–Fisher distribution. Fihser claimed{{fact}} that the probability that ''&mu;''<sub>2</sub>&nbsp;&minus;&nbsp;''&mu;''<sub>1</sub> is in this interval, given the data (ultimately the ''X''s) is the probability that a Behrens–Fisher-distributed random variable is between &minus;''A'' and&nbsp;''A''.
where ''A'' is the appropriate percentage point of the Behrens–Fisher distribution. Fisher claimed{{citation needed|date=August 2012}} that the probability that ''&mu;''<sub>2</sub>&nbsp;&minus;&nbsp;''&mu;''<sub>1</sub> is in this interval, given the data (ultimately the ''X''s) is the probability that a Behrens–Fisher-distributed random variable is between &minus;''A'' and&nbsp;''A''.


=== Fiducial intervals versus confidence intervals ===
=== Fiducial intervals versus confidence intervals ===


Bartlett{{fact}} showed that this "fiducial interval" is not a confidence interval because it does not have a constant coverage rate. Fisher did not consider that a cogent objection to the use of the fiducial interval.{{fact}}
Bartlett{{citation needed|date=August 2012}} showed that this "fiducial interval" is not a confidence interval because it does not have a constant coverage rate. Fisher did not consider that a cogent objection to the use of the fiducial interval.{{citation needed|date=August 2012}}

==See also==


<!-- ==See also==
* [[Fisher transformation]]
* [[Fisher transformation]]
* [[Fisher information]]
* [[Fisher information]] -->

== Further reading ==
*Kendall, Maurice G., Stuart, Alan (1973) ''The Advanced Theory of Statistics, Volume 2: Inference and Relationship, 3rd Edition'', Griffin. {{ISBN|0-85264-215-6}} (Chapter 21)

{{ProbDistributions|continuous-infinite}}

==References==
{{Reflist}}


{{DEFAULTSORT:Behrens-Fisher distribution}}
[[:Category:Probability distributions]]
[[:Category:Statistics]]
[[Category:Continuous distributions]]

Latest revision as of 05:52, 4 January 2023

In statistics, the Behrens–Fisher distribution, named after Ronald Fisher and Walter Behrens, is a parameterized family of probability distributions arising from the solution of the Behrens–Fisher problem proposed first by Behrens and several years later by Fisher. The Behrens–Fisher problem is that of statistical inference concerning the difference between the means of two normally distributed populations when the ratio of their variances is not known (and in particular, it is not known that their variances are equal).[1]

Definition

[edit]

The Behrens–Fisher distribution is the distribution of a random variable of the form

where T1 and T2 are independent random variables each with a Student's t-distribution, with respective degrees of freedom ν1 = n1 − 1 and ν2 = n2 − 1, and θ is a constant. Thus the family of Behrens–Fisher distributions is parametrized by ν1ν2, and θ.

Derivation

[edit]

Suppose it were known that the two population variances are equal, and samples of sizes n1 and n2 are taken from the two populations:

where "i.i.d" are independent and identically distributed random variables and N denotes the normal distribution. The two sample means are

The usual "pooled" unbiased estimate of the common variance σ2 is then

where S12 and S22 are the usual unbiased (Bessel-corrected) estimates of the two population variances.

Under these assumptions, the pivotal quantity

has a t-distribution with n1 + n2 − 2 degrees of freedom. Accordingly, one can find a confidence interval for μ2 − μ1 whose endpoints are

where A is an appropriate quantile of the t-distribution.

However, in the Behrens–Fisher problem, the two population variances are not known to be equal, nor is their ratio known. Fisher considered[citation needed] the pivotal quantity

This can be written as

where

are the usual one-sample t-statistics and

and one takes θ to be in the first quadrant. The algebraic details are as follows:

The fact that the sum of the squares of the expressions in parentheses above is 1 implies that they are the squared cosine and squared sine of some angle.

The Behren–Fisher distribution is actually the conditional distribution of the quantity (1) above, given the values of the quantities labeled cos θ and sin θ. In effect, Fisher conditions on ancillary information.

Fisher then found the "fiducial interval" whose endpoints are

where A is the appropriate percentage point of the Behrens–Fisher distribution. Fisher claimed[citation needed] that the probability that μ2 − μ1 is in this interval, given the data (ultimately the Xs) is the probability that a Behrens–Fisher-distributed random variable is between −A and A.

Fiducial intervals versus confidence intervals

[edit]

Bartlett[citation needed] showed that this "fiducial interval" is not a confidence interval because it does not have a constant coverage rate. Fisher did not consider that a cogent objection to the use of the fiducial interval.[citation needed]


Further reading

[edit]
  • Kendall, Maurice G., Stuart, Alan (1973) The Advanced Theory of Statistics, Volume 2: Inference and Relationship, 3rd Edition, Griffin. ISBN 0-85264-215-6 (Chapter 21)

References

[edit]
  1. ^ Kim, Seock-Ho; Cohen, Allan S. (December 1998). "On the Behrens-Fisher Problem: A Review". Journal of Educational and Behavioral Statistics. 23 (4): 356–377. doi:10.3102/10769986023004356. ISSN 1076-9986. S2CID 85462934.