Everyone makes mistakes during data analysis. Literally everyone. The question is not what errors you make, it's what systems you put into place to prevent them from happening. Here are mine. [a thread because I'm sad to miss #SIPS2018]
External Tweet loading...
If nothing shows, it may have been deleted
by @siminevazire view original on Twitter
Since then, we've audited dozens of papers (I like this term much more than "data thugged" @jamesheathers). E.g. in @Tom_Hardwicke's new manuscript: osf.io/preprints/bits…. Summary: the error rate is very high. Most errors don't undermine papers, but most papers have errors.
So what do we do? 1) Literate programming. If you have to write code that others can read, you catch typos and errors much more quickly. And if you're not scripting your data analysis using code, it's time to start.
2) Standardize your workflow. If you do things consistently, you will be less likely to make new, ad-hoc errors that you don't recognize. For me this has meant learning tidyverse.org - an amazing ecosystem for R data analysis.
3) Version control. osf.io is a gateway. git and github.com take time to learn but are far better tools for collaborating. If you track what you did you will not lose files/erase work/accidentally modify data, blocking a whole set of possible errors!
4) Code review. I wish we did this more. But co-working (babieslearninglanguage.blogspot.com/2017/11/co-wor…) and pair-programming can help catch errors and promote clean, standardized analysis. And you can always ask a friend to "co-pilot" after you are done.
5) Openness by default. If you curate and analyze data as though someone who thinks you are wrong is watching you and criticizing, you will be much more careful. This is a very good thing. More thoughts in our transparency guide: psyarxiv.com/rtygm.
In sum: of course we should be afraid of errors! But don't *just* be afraid. Accept that you *have* made errors - and will make more. We all do. Act to put systems in place that help you catch and correct errors before they enter the literature. [end]
• • •
Missing some Tweet in this thread? You can try to
force a refresh
What is "the open science movement"? It's a set of beliefs, research practices, results, and policies that are organized around the central roles of transparency and verifiability in scientific practice. An introductory thread. /1
The core of this movement is the idea of "nullius in verba" - take no one's word for it. The distinguishing feature of science on this account is the ability to verify claims. Science is independent of the scientist and subject to skeptical inquiry. /2
These ideas about the importance of verification are supported by a rich and growing research literature suggesting that not all published science is verifiable. Some papers have typos, some numbers can't be reproduced, some experiments can't be replicated independently. /3
A thought on grad advising. When I was a second year, an announcement went out to our dept. with the abstract for a talk I was giving in the area talk series. A senior faculty member wrote back with a scathing critique (cc'd to my advisor, @LanguageMIT). /1
The part that made the biggest impression on me: they said that the first line of my abstract was *so embarrassing that they thought my graduate training had failed*! Actual quote: "You look naive at best, many other things at worst." And on from there. /2
My advisor wrote back immediately: "Hi [critic], I wrote that line." /3
Prosocial development throwdown at #icis18: presentations by Audun Dahl, Felix Warneken, and @JKileyHamlin. Three opinions on a fascinating topic! [livetweet thread]
Dahl up first. Puzzles of prosociality: there’s an amazing ability to help others prosocially from an early age, but some infants don’t! Why? Behaviors emerge via 1) social interest and 2) socialization.
Framework of co-action. Starting with early turn-taking and contingency, caregivers scaffold social interaction. They even encourage and facilitate helping behaviors.