MATH 376 -- Probability and Statistics II 

A Comment on WMS, page 525. 

March 29, 2010 

 

I want to point out that there is a discussion in our text, Wackerly, Mendenhall,  

and Schaeffer, Mathematical Statistics with Applications, 7th edition, that is somewhat 

dated and that does not completely square with the current understanding in  

statistical science.  The statement in question is from the discussion of  

small sample tests for means on page 525 -- ``Such investigations [of 

the empirical distributions of the t-statistic] have shown that moderate 

departures from normality in the distribution of the population have 

little effect on the distribution of the test statistic''  -- a property 

called robustness.   This was certainly the consensus for a long time (say 

through the 1970's.)  However, more recent research has shown that 

this is not always the case  (or to be fair perhaps, that the precise meaning 

of ``moderate departures from normality'' needs to be more carefully stated!)   

 

How actually robust is the  t-test  in the case where the normality assumptions  

are violated? 

 

Let us study some small samples from a lognormal distribution.  (See the pdf 

plotted below.)  Note this is a skewed distribution with a relatively long right 

tail.   Many quantities observed in scientific experiments have distributions 

like this, so this is not just a theoretical example.   

 

with(Statistics); -1 

`assign`(Y, RandomVariable(LogNormal(.5, .9))); 1 

_R11 (1)
 

plot(PDF(Y, y), y = 0 .. 10); 1; `assign`(Means, []); -1; `assign`(Vars, []); -1 

Plot_2d
 

We generate 1000 random samples of size 8 from this distribution, and plot the means and 

variances of the samples in a scatter plot:
for i to 1000 do `assign`(YS, Sample(Y, 8)); `assign`(Means, [op(Means), Mean(YS)]); `assign`(Vars, [op(Vars), Variance(YS)]) end do; -1
for i to 1000 do `assign`(YS, Sample(Y, 8)); `assign`(Means, [op(Means), Mean(YS)]); `assign`(Vars, [op(Vars), Variance(YS)]) end do; -1
for i to 1000 do `assign`(YS, Sample(Y, 8)); `assign`(Means, [op(Means), Mean(YS)]); `assign`(Vars, [op(Vars), Variance(YS)]) end do; -1
for i to 1000 do `assign`(YS, Sample(Y, 8)); `assign`(Means, [op(Means), Mean(YS)]); `assign`(Vars, [op(Vars), Variance(YS)]) end do; -1
for i to 1000 do `assign`(YS, Sample(Y, 8)); `assign`(Means, [op(Means), Mean(YS)]); `assign`(Vars, [op(Vars), Variance(YS)]) end do; -1
for i to 1000 do `assign`(YS, Sample(Y, 8)); `assign`(Means, [op(Means), Mean(YS)]); `assign`(Vars, [op(Vars), Variance(YS)]) end do; -1
 

ScatterPlot(Means, Vars); 1 

Plot_2d
 



Recall that in the definition of  t-distributions, we said that when sampling from a
normal population,
 

             T = `/`(`*`(`+`(Ybar, `-`(mu)), `*`(sqrt(n))), `*`(S))
had a  t-distribution with n - 1 degrees of freedom,  because it could be 

rearranged to put it into the standard form  `/`(`*`(Z), `*`(sqrt(`/`(`*`(U), `*`(`+`(n, `-`(1))))))), where  Z  is standard 

normal,  U  has a distribution with  n - 1  degrees of freedom, and  

Z  and  U  are independent.   This is the same as saying  Ybar  and  S  are 

independent.   Does that match what we are seeing with these 

samples?? 

 


Conclusion:  
For these samples, it is not valid to assume that the sample mean and 

the sample variance (or standard deviation) are independent(!)   It is clearly true 

that the larger the sample mean is, the larger the sample variance is too. 

Does this make a difference??
`assign`(mu, ExpectedValue(Y)); 1
 

2.471931923 (2)
 


 

Now, let us compare the empirical distribution of the  t-statistics  from these samples 

with what we expect from the theoretical  t - density function:
 

`assign`(TListL, [seq(`/`(`*`(`+`(Means[i], `-`(mu))), `*`(sqrt(`+`(`*`(`/`(1, 8), `*`(Vars[i])))))), i = 1 .. 1000)]); -1; `assign`(empPDF, PlotEmpiricalPDF(TList, -10., 10.)); 1 

 

WARNING: There are
PLOT(CURVES([[-10.5000000000000000, 0.], [-9.50000000000000000, 0.401203610799999964e-2], [-8.50000000000000000, 0.100300902699999991e-2], [-7.50000000000000000, 0.601805416200000033e-2], [-6.50000000... (3)
 

`assign`(T, RandomVariable(StudentT(7))); 1; `assign`(tcurve, plot(PDF(T, t), t = -10 .. 10, color = blue)); -1; with(plots); -1
`assign`(T, RandomVariable(StudentT(7))); 1; `assign`(tcurve, plot(PDF(T, t), t = -10 .. 10, color = blue)); -1; with(plots); -1
`assign`(T, RandomVariable(StudentT(7))); 1; `assign`(tcurve, plot(PDF(T, t), t = -10 .. 10, color = blue)); -1; with(plots); -1
 

_R14 (4)
 

display(empPDF, tcurve); 1 

Plot_2d
 

Note that the empirical distribution of the  t-statistics  is 

quite a bit different from the theoretical  t-PDF  for  

8 - 1 = 7 degrees of freedom.  The red empirical distribution  

has significantly more probability mass in the lower tail than 

the t-curve, and correspondingly less in the center and upper 

tail.  This means that confidence intervals or hypothesis 

tests using these samples will lead to incorrect conclusions 

more often than we expect.  (That is, for instance, the α and β 

for hypothesis tests will not achieve the desired values.) 

 

This kind of thing is more common than people thought until relatively 

recently.  Hence a whole new branch of the subject devoted to developing 

robust statistics (that is, statistics that are not overly affected by the  

failure of an assumption such as normality) has developed in the last 20 or 

30 years.  Our book, being an introduction to mathematical statistics, 

unfortunately does not discuss these methods.