The Scholarly Kitchen

What’s Hot and Cooking In Scholarly Publishing

  • About
  • Archives
  • Collections
    Scholarly Publishing 101 -- The Basics
    Collections
    • Scholarly Publishing 101 -- The Basics
    • Academia
    • Business Models
    • Discovery and Access
    • Diversity, Equity, Inclusion, and Accessibility
    • Economics
    • Libraries
    • Marketing
    • Mental Health Awareness
    • Metrics and Analytics
    • Open Access
    • Organizational Management
    • Peer Review
    • Strategic Planning
    • Technology and Disruption
  • Translations
    topographic world map
    Translations
    • All Translations
    • Chinese
    • German
    • Japanese
    • Korean
    • Spanish
  • Chefs
  • Podcast
  • Follow

Desperately Seeking (Statistical) Significance

  • By Phil Davis
  • Jun 15, 2022
  • 4 Comments
  • Time To Read: 3 mins
  • Authors
  • Ethics
  • Metrics and Analytics
  • Peer Review
  • Research
Share
0 Shares

Mention the name Brian Wansink around Cornell and you’ll make a lot of people uncomfortable. Wansink, a food researcher, was known by his innovative experimental designs and headline-grabbing results. He was also exposed as someone who put good stories ahead of good results, investigated for “p-hacking” his data, and unceremoniously left Cornell in 2018. While his departure was more than four years ago, the damage he did to the reputation of his school and colleagues is still raw. For top-research institutions and researchers, reputation is everything. And reputation rubs off in both directions.

little boy in high grass searching for something through binoculars

p-hacking is a term used in the quantitative life and social sciences where a researcher analyzes data to ensure a statistically significant result. In practice, it can mean using inappropriate statistical tests, excluding data points, collecting lots of variables — in the hope that some associations will come out significant — among other inappropriate behaviors. Fear of p-hacking is the very reason why rigorous medical journals require authors to register their experiments in advance and follow a strict protocol for reporting their results.

In a previous post, I described how a recent study on the effects of Twitter on paper citations violated several of the journal’s rules, by failing to describe how sample sizes were calculated, omitting how readers could access the paper’s underlying data, and refusing to answer questions about the validity of the paper’s findings.

On 13 June 2022, more than a month after my initial request, the editorial office of the European Heart Journal provided me a link to the authors’ data (as of this writing, there is no public link from the published paper). The data were provided in .dta format — a proprietary format for users of a statistical software called Stata. Nevertheless, with the help of a Cornell data librarian (go librarians!), I was able to access the data in a generic format that allowed me to verify the results reported in the paper.

The statistics for analyzing and reporting the results of Randomized Controlled Trials (RCTs) are often very simple. Because treatment groups are made similar in all respects with the exception of the intervention, no fancy statistical analysis is normally required. This is why most RCTs are analyzed using simple comparisons of sample means or medians.

Deviating from this norm, the authors of the Twitter-citation paper used Poisson regression, a more complicated model that is very useful in some fields (e.g. economics) when analyzing data with lots of independent variables. However, Poisson regression is limited in its application because it comes with a big assumption — the mean value must equal the variance. When this assumption is violated, the researcher should use a more flexible model, like the Negative Binomial. Using an inappropriate model on data can result in unreliable results, which is exactly what I found with the Twitter data.

…the only test that provided a statistically significant finding was the one where the model was inappropriate for the data

Using Poisson regression on their data, I got the same results as reported (12% citation difference, 95% confidence interval 8% to 15%, p<0.0001), which appears to be a robust and statistically significant finding. However, the model fits the data very poorly. When I analyzed their dataset using a Negative Binomial model, the data were no longer significant (13%, 95% C.I. -5% to 33%, p=0.17). Yes, the estimate was close, but the confidence interval straddled zero. Using a common technique when dealing with highly-skewed data (normalizing the data with a log transformation) and employing a simple linear model also provided non-significant results (8%, 95% C.I. -7% to 25%, p=0.33). Similarly, a simple comparison of means (t-test) was non-significant (p=0.17), as was the non-parametric (signed-rank) equivalent (p=0.33).

In sum, the only test that provided a statistically significant finding was the one where the model was inappropriate for the data.

The authors didn’t register their protocols or even provide justification for a Poisson regression model in their preliminary paper. A description of how their sample size was determined was missing, as was a data availability statement — both are clear violations of the journal’s policy. The editorial office was kind enough to provide me with a personal link to the dataset, but it is still not public. They have continued to invite me to submit a formal letter to the Discussion Forum of their journal. They have not responded to questions about how this paper got through to publication without these required elements or what they plan on doing about it. No one is willing to admit fault, and the undeclared connection of several authors with current or past EHJ editorial board roles raises questions about special treatment.

I stopped using Twitter recently because it began serving me inappropriate ads and senseless personal tweets. Elon Musk’s pronouncements on how he would change the service, if he became CEO, added to my decision. Social media companies wax and wane, and their influence in the lives of researchers can be ephemeral. In five years, a study on the citation benefit of Twitter may be heaped into the same bin as other fill-in-the-blank-leads-to-citations papers. But a tarnished reputation is deep and long-lasting. I hope the editors of EHJ understand what they are sacrificing with this paper.

Share
0 Shares
Share
0 Shares
Phil Davis

Phil Davis

Phil Davis is a publishing consultant specializing in the statistical analysis of citation, readership, publication and survey data. He has a Ph.D. in science communication from Cornell University (2010), extensive experience as a science librarian (1995-2006) and was trained as a life scientist. https://phil-davis.com/

View All Posts by Phil Davis

Discussion

4 Thoughts on "Desperately Seeking (Statistical) Significance"

Thanks Phil (and the SK) for sharing this analysis. It’s really important that issues around data integrity get fresh air, and this sounds like an example where that was sorely lacking. It also helps educate those (‘most?) of us less familiar with statistical analysis about the critical role that journal policies around data play in ensuring transparency. It’s something I’ve read many times, but hard examples like this help drive it home.

  • By Tim Lloyd
  • Jun 15, 2022, 4:01 PM

Thanks Phil for such an insightful analysis. It reminds me of the “3 types of lies” and at the same time “devil is in the details”! As Tim said, journal policies around data is very important. On the other hand, I think the main issue of all these is the pressure of publishing more articles and fast turn-around time. Rigorous analysis and peer review might be compromised in come cases.

  • By Rick Lee
  • Jun 15, 2022, 9:10 PM

Thanks Phil for illuminating the importance of open data and reproducibility research. Rick rightly said that one of the causes is pressure to publish more. Policies of journals also effecting this trend. Another cause to surely produce statistically significant results may be that marginalized results are oftenly rejected by peers on reviewing stage although open data availibility may increase the acceptability of marginalized findings. It is also important to ensure the protocols and rigor in conduct of research.

  • By MUHAMMAD RAFIQ
  • Jun 16, 2022, 1:48 PM

It does seem that there’s a benefit to pulling main points from these posts calling out both the dodgy behavior of the authors and the lax behavior of the publisher and publishing it in EHJ as a more durable venue. Of course if they make difficult you can always rant about that here

  • By Chris Mebane
  • Jun 19, 2022, 8:27 AM

Comments are closed.

Official Blog of:

Society for Scholarly Publishing (SSP)

The Chefs

  • Rick Anderson
  • Todd A Carpenter
  • Angela Cochran
  • Lettie Y. Conrad
  • David Crotty
  • Joseph Esposito
  • Roohi Ghosh
  • Robert Harington
  • Haseeb Irfanullah
  • Lisa Janicke Hinchliffe
  • Phill Jones
  • Roy Kaufman
  • Scholarly Kitchen
  • Stephanie Lovegrove Hansen
  • Alice Meadows
  • Alison Mudditt
  • Jill O'Neill
  • Charlie Rapple
  • Dianndra Roberts
  • Maryam Sayab
  • Roger C. Schonfeld
  • Avi Staiman
  • Randy Townsend
  • Tim Vines
  • Hong Zhou

Interested in writing for The Scholarly Kitchen? Learn more.

Most Recent

  • Mental Health Awareness Mondays — Finding Balance While Navigating Career Uncertainty and Industry Changes
  • Guest Post — The SSP Annual Meeting: Keeping the Faith in Unsettled Times
  • Ask the Librarians: Recapping a Scholarly Kitchen Roundtable at the 2025 Charleston Library Conference 

SSP News

Scholarly Publishing Gets Its Awards Season Moment

Apr 9, 2026

Bring Your Creativity to Chula Vista: The 3rd Annual SSP Originals Auction

Apr 8, 2026

Annual Meeting Early Registration is Open—Download the Preliminary Program now!

Apr 8, 2026
Follow the Scholarly Kitchen Blog Follow Us

Related Articles:

  • phone screen showing social media apps Fill in the Blank Leads to More Citations
  • Intention to Tweet: Medical Study Reports Tweets Improve Citations
  • Twitter logo Tweets, and Our Obsession with Alt Metrics

Next Article:

dog sitting on top of packed suitcase Ask The Chefs: It's Travel Time Again!
Society for Scholarly Publishing (SSP)

The mission of the Society for Scholarly Publishing (SSP) is to advance scholarly publishing and communication, and the professional development of its members through education, collaboration, and networking. SSP established The Scholarly Kitchen blog in February 2008 to keep SSP members and interested parties aware of new developments in publishing.

The Scholarly Kitchen is a moderated and independent blog. Opinions on The Scholarly Kitchen are those of the authors. They are not necessarily those held by the Society for Scholarly Publishing nor by their respective employers.

  • About
  • Archives
  • Chefs
  • Podcast
  • Follow
  • Advertising
  • Privacy Policy
  • Terms of Use
  • Website Credits
ISSN 2690-8085