Wednesday, October 14, 2009

The cult of statistical significance

A few weeks ago, a friend asked me if I know the book "the cult of statistical significance". I did not. I replied the following, and today, he told me that he liked my summary of statistical inference:

Every year when I teach the confidence interval and t-test in the third and fourth weeks of the fall semester, I would have some new ideas about how to introduce these confusing concepts. This year, I said to the class that if I were to rewrite all statistics textbooks, I would not include hypothesis testing, because all the information included in a test is reflected in the confidence interval. The width of a confidence interval also tells us whether the data provided useful information. Should we use the confidence interval (and interpret it in the Bayesian way), we understand the problem of statistical inference better. This line of thinking is inspired by the "display()" function in R (Gelman's arm package), which summarizes a linear model coefficients only in terms of the estimated values and their standard error, no p-values are provided. Last year, I told my class to forget about the p-values and focus on the standard error, because we can quickly come up with an approximate 95% confidence interval of each coefficient (est +/- 2se). Many students told me that without p-values they think about the real meaning of the estimated
coefficients, and with p-values (using the standard R function summary()) they think about the significance.

Obviously, the development of hypothesis test is an important contribution to statistics. The way Fisher used hypothesis tests is more in line with a scientist evaluates his/her theory. A small p-value is evidence against the hypothesis, but not a support for any specific alternative. So, if we follow Fisher's approach, we start to think about new (specific) theory after seeing a small p-value. If we are interested in learning the speed of light by conducting a new experiment, we start from the existing estimate (mu0). Fisher would compare new data to the existing estimate and conclude whether mu0 might be an under- or over-estimate or about right. If the data show the speed is not the same as mu0, we may want to revise the estimate.
The data would say the likely true mean lies inside the confidence interval (when testing H0: mu=mu0, we would not reject the null if mu0 is inside the CI). The Neyman-Pearson paradigm of hypothesis test is aimed at decision making, not scientific inference. For them, statistical significance is of real significance, because it dictates the action we take. For example, Newman-Pearson approach would be very useful for a state agency when deciding whether to grant a permit for a wastewater treatment plant. A test H0: BOD<=30 versus BOD>30 (or the reverse) is all we need. We are not really interested in the true BOD concentration in the discharge. So, the misuse of significance is often a result of misunderstanding of a test. The way a typical statistics course treats this topic (mixing the two approaches into a unified test procedure) is a disservice to all. It simplifies a scientific problem into a yes/no dichotomy and prevents creative thinking.

Tuesday, October 13, 2009

The logical incoherence of modern statistical practice

The great Stuart Hurlbert came to Duke to give a talk on the logical incoherence of modern statistical practice. My immediate thought after the talk was that the title of his talk was all wrong. It should be: the logical incoherence of ecological applications of statistics. The most obvious "incoherence" in modern statistics is the violations of the likelihood principle of many concepts and practices that were not mentioned at all. In fact, he could have ended the talk in five minutes by citing the first paragraph of Berger and Wolpert (1988):

Among all prescriptions for statistical behavior, the Likelihood Principle (LP) stands out as the simplest and yet most far reaching. It essentially states that all evidence, which is obtained from an experiment, about an unknown quantity [;\theta;], is contained in the likelihood function of [;\theta;] for the given data. The implications of this are profound, since most non-Bayesian approaches to statistics and indeed most standard statistical measures of evidence (such as coverage probability, error probabilities, significance level, frequentist risk, etc.) are then contraindicated.

Here is a run-down of the "sins" of statistics discussed by Hurlbert:

1. a fixed type I error probability (alpha) --
2. the use of term "significant"
3. the concept of type II error (here he thinks that we should never accept the null hypothesis)
4. one-tailed hypothesis tests
5. multiple comparisons, and
6. repeated measures ANOVA.

Thursday, May 14, 2009

Environmental and Ecological Statistics with R

Finally, the work is done and the book is ready for the publisher. The title of this post will take you to my (yet-to-be-completed) homepage for the book.

Thursday, April 23, 2009

Ecological Threshold -- Quantitative Options

The term "threshold" appears as a key word in more and more papers. However, there are misunderstandings of the few statistical methods for quantifying a threshold. Often, people seek for a one-size-for-all type of models with an option of producing the magic p-value. With several papers documenting statistical models for detecting and quantifying environmental and ecological thresholds, the questions is how to select the right model.

A statistical model for detecting a threshold comes with assumptions, just like a simple linear regression model. These assumptions may or may not fit the data in question and residual plots are always the best vehicle for evaluating these assumptions.

Thursday, March 19, 2009

Embedding ALL fonts

After I turned in the PDF file of my final draft of "Environmental and Ecological Statistics with R", the publisher (Chapman and Hall/CRC) informed me that I did not embed all fonts in the PDF file. I had no idea what "embedding" really mean. The publisher directed me to their contractor at International Typesetting and Composition in India for help. ITC consultants were very helpful in explaining the process of using Adobe Distiller and illustrator:

1 In distiller, fonts location for font folder is shown.
2 Check/select the option embed all fonts in distiller.
3 All fonts used in figures must be embedded in figures itself. You can open the figures in illustrator, and save-as with the option "Embed all Fonts" selected

But I have hundreds of figures which means I have to spend weeks if not months to complete step 3. Besides, I cannot afford to buy Adobe just for this job. So, I Googled to see if there are better solutions. Grassbook.org has a piece on this topic for Linux users. After a few trials using the grassbook.org posting as a reference, I figured out a simple alternative for LaTeX users both for Mac and PC:

1. Compile the LaTeX file to create the .dvi file,
2. In DOS command console or Mac Ternimal:
$ dvips -f mybook.dvi > mybook.ps
3. Then, use ghostscript with option -dPDFSETTINGS=/prepress:
$ gs -dNOPAUSE -dBATCH -dPDFSETTINGS=/prepress -sDEVICE=pdfwrite -sOutputFile=mybook.pdf mybook.ps

The resulting PDF file has been checked and accepted by the publisher.

Log or not log

LOGorNOTLOG.html Log or not log, that is the question May 19, 2018 In 2014 I taught a special topics class on statistical i...