Statistical Estimation
Basu's Theorem
A complete sufficient statistic is independent of every ancillary statistic. This provides the cleanest method for proving independence between statistics without computing joint distributions.
Prerequisites
Why This Matters
Proving that two statistics are independent usually requires computing their joint distribution and factoring it. This can be painful. Basu's theorem gives a shortcut: if one statistic is complete sufficient and the other is ancillary, they are independent. No joint distribution computation needed.
The classic application: in a normal sample, the sample mean is independent of the sample variance . This fact underpins the derivation of the t-test. Basu's theorem proves it in two lines.
Formal Setup
Ancillary Statistic
A statistic is ancillary for a parameter if and only if its distribution does not depend on . It carries no information about by itself, but may carry information in combination with other statistics.
Complete Statistic
A statistic is complete if and only if for every measurable function :
Completeness means there are no nontrivial unbiased estimators of zero based on . Informally, contains no "wasted" information.
Main Theorems
Basu's Theorem
Statement
If is a complete sufficient statistic for and is ancillary for , then and are independent (under every ).
Intuition
Sufficiency means that the conditional distribution of the data given does not depend on . Ancillarity means the marginal distribution of does not depend on . Completeness forces these two facts to combine into independence: the conditional distribution of given must equal the marginal distribution of .
Proof Sketch
Let be any measurable set. Define . By sufficiency, does not depend on . By ancillarity, does not depend on . So does not depend on , and for all (using ancillarity again). By completeness, a.s., meaning a.s. This is independence.
Why It Matters
Without this theorem, proving independence of and in normal sampling requires computing the joint density via a change of variables. With Basu's theorem, you only need three facts: (1) is complete sufficient for when is known, (2) is ancillary for , (3) apply the theorem. This pattern extends to many other settings.
Failure Mode
If the sufficient statistic is not complete, the theorem fails. For example, in a uniform distribution on , the order statistics are sufficient but not complete. The range is ancillary but not independent of the midrange .
Canonical Examples
Normal sampling: mean and variance independence
Let with known. The sample mean is complete sufficient for (this follows from the normal distribution being an exponential family). The statistic has a distribution that depends only on , not on . So is ancillary for . By Basu's theorem, and are independent. This is the fact that makes the t-statistic have a t-distribution.
Exponential distribution: mean and coefficient of variation
Let . The sample sum is complete sufficient for . The vector of ratios is ancillary (its distribution is uniform on the simplex, independent of ). By Basu's theorem, is independent of all the ratios .
Building the t-statistic pivot
Let . Treat as a fixed but arbitrary nuisance and apply Basu within the model where is the parameter of interest at held fixed: is complete sufficient for , and any function of the residuals has a distribution that does not depend on (shift invariance). In particular , and so , is ancillary for . By Basu, under for every . The studentized statistic is the ratio of and , which are independent by Basu and have distributions free of and . The ratio is therefore distributed for every , which is exactly what makes a valid confidence interval when is unknown — the cancellation of in the ratio is what removes the nuisance. Without the independence step, the distribution of the ratio does not follow from Cochran's theorem alone.
Common Confusions
Ancillary does not mean useless
An ancillary statistic carries no information about by itself. But conditionally, given the ancillary, the precision of estimation can change. This is the basis of conditional inference. Basu's theorem says: if you have a complete sufficient statistic, you cannot improve estimation by conditioning on the ancillary.
Completeness is doing the heavy lifting
Sufficiency alone does not imply independence from ancillary statistics. Completeness is the key condition. Think of completeness as saying the sufficient statistic has no redundancy: there is no function of that is itself ancillary.
Conditionality and Ancillary Statistics
The conditionality principle (Cox, 1958) says that if is ancillary for , inference about should be carried out in the conditional distribution of the data given , not the marginal distribution. The motivation is that indexes which experiment effectively took place, so reporting unconditional frequency properties averages over experiments the data did not come from.
A textbook case is Cox's two-instrument example. Flip a fair coin: heads uses a precise instrument, tails uses a noisy one. The outcome of the flip is ancillary for the unknown mean. The unconditional variance of the estimator averages both instruments. The conditional variance, given which instrument was used, is what the experimenter actually faces. Most statisticians find the conditional report more honest.
Basu's theorem interacts cleanly with this principle. If is complete sufficient and is ancillary, Basu gives , so conditioning on leaves the distribution of unchanged. In that case conditional and unconditional inference about based on agree. When multiple ancillaries exist or no complete sufficient statistic is available (for example, curved exponential families), conditioning on different ancillaries can yield different answers, and the choice becomes a genuine modeling decision. Fisher pushed for conditioning on a maximal ancillary; uniqueness is not guaranteed.
Conditionality is not universally accepted
The conditionality principle, the sufficiency principle, and the weak likelihood principle together imply the strong likelihood principle (Birnbaum, 1962), which frequentists who prize coverage guarantees often reject. The debate is real. Treat the principle as a defensible default when a natural ancillary exists, not as a theorem.
Summary
- Complete sufficient + ancillary implies independent
- The proof uses completeness to upgrade "same expectation" to "equal a.s."
- The main application is proving independence without computing joint distributions
- Fails without completeness: sufficiency alone is not enough
Exercises
Problem
Let . Identify a complete sufficient statistic and an ancillary statistic. State what Basu's theorem tells you.
Problem
Give an example where a sufficient statistic and an ancillary statistic are not independent. What condition of Basu's theorem fails?
References
Primary source:
- Basu, D. "On Statistics Independent of a Complete Sufficient Statistic," Sankhyā 15 (1955), 377-380. The original statement and proof.
Canonical textbooks:
- Casella & Berger, Statistical Inference (2nd ed., 2002), Section 6.2.4 "Sufficiency, ancillarity, and completeness" and Theorem 6.2.24 (Basu's theorem), pp. 287-289. Example 6.2.25 gives the normal mean/variance independence.
- Lehmann & Casella, Theory of Point Estimation (2nd ed., 1998), Section 1.6 "Complete sufficient statistics," Theorem 1.6.22 (Basu), and Example 1.6.23. Chapter 2.1 connects completeness to UMVUE via Lehmann-Scheffé.
- Lehmann & Romano, Testing Statistical Hypotheses (3rd ed., 2005), Sections 4.4-4.5. Canonical treatment of ancillarity, Basu's theorem, and applications to similar tests.
- Cox, D. R. "Some Problems Connected with Statistical Inference," Annals of Mathematical Statistics 29 (1958), 357-372. The conditionality principle and Cox's two-instrument example.
Additional references:
- Keener, Theoretical Statistics: Topics for a Core Course (2010), Chapter 3 (sufficiency, completeness, Basu).
- Schervish, Theory of Statistics (1995), Section 2.1.3 (completeness) and Section 2.2 (ancillarity and conditioning).
- Birnbaum, A. "On the Foundations of Statistical Inference," JASA 57 (1962), 269-306. Derives the likelihood principle from sufficiency and conditionality.
Next Topics
- Fisher information: the natural next step in estimation theory
Last reviewed: April 26, 2026
Canonical graph
Required before and derived from this topic
These links come from prerequisite edges in the curriculum graph. Editorial suggestions are shown here only when the target page also cites this page as a prerequisite.
Required prerequisites
1- Sufficient Statistics and Exponential Familieslayer 0B · tier 2
Derived topics
1- Fisher Information: Curvature, KL Geometry, and the Natural Gradientlayer 0B · tier 1
Graph-backed continuations