Daniel Nüst�Institut für Geoinformatik�d.n@wwu.de | @nordholmen�0000-0002-0024-5046
Slides: http://bit.ly/hangout21-repro
1
Practical reproducibility and reproducibility vs. peer review
2
CC-BY 3.0, Sebastian Bertalan, Wikimedia Commons
An article about computational science in � a scientific publication is not the � scholarship itself, it is merely advertising
of the scholarship. The actual scholarship
is the complete software development � environment and the complete set of � instructions which generated the figures.
Claerbout’s claim:
Crisis? Crisis of what?
Credibility crisis?�Replicability crisis?�Reproducibility crisis?�Robustness crisis?�Generalisability crisis?
3
Practical Reproducibility in Geography and Geosciences
Daniel Nüst & Edzer Pebesma (2020) Practical Reproducibility in Geography and Geosciences, Annals of the American Association of Geographers, DOI:10.1080/24694452.2020.1806028�PDF: http://nuest.staff.ifgi.de/N%C3%BCst-and-Pebesma_2020_AAM_Practical-Reproducibility-in-Geography-and-Geosciences.pdf
Creating reproducible workflows
Computing environment: hardware + software, containers/virtualisation (Binder), freezing/pinning�Script-based workflows: no point-and-click GIS, notebooks (Jupyter, R Markdown)�> Research compendium
Challenges
Education, publishing practices, SDIs, GIS, proprietary software,�lack of rewards/pressure, sensitive data, time, ...�> all solvable
Reproducible Research &�research software
Peer Review
5
Reproducible research and peer review are cornerstones of science. But are they getting along?
CODECHECK
7
The inverse problem in reproducible research. Figure 1 of https://doi.org/10.12688/f1000research.51738.1
The left half of the diagram shows a diverse range of materials used within a laboratory. These materials are often then condensed for sharing with the outside world via the research paper, a static PDF document. Working backwards from the PDF to the underlying materials is impossible. This prohibits reuse and is not only non-transparent for a specific paper but is also ineffective for science as a whole. By sharing the materials on the left, others outside the lab can enhance this work.
8
The CODECHECK example process implementation. Figure 2 of https://doi.org/10.12688/f1000research.51738.1
The left half of the diagram shows a diverse range of materials used within a laboratory. These materials are often then condensed for sharing with the outside world via the research paper, a static PDF document. Working backwards from the PDF to the underlying materials is impossible. This prohibits reuse and is not only non-transparent for a specific paper but is also ineffective for science as a whole. By sharing the materials on the left, others outside the lab can enhance this work.
9
Independent execution of computations underlying research articles.
One re-execution by codechecker during peer review
👶
25 Certificates https://codecheck.org.uk/register/
📃
🔏
10
Nüst D and Eglen SJ. CODECHECK: an Open Science initiative for the independent execution of computations underlying research articles during peer review to improve reproducibility [version 1; peer review: awaiting peer review]. F1000Research 2021, 10:253 (https://doi.org/10.12688/f1000research.51738.1)
Reproducible AGILE
Created by AGILE Initiative in 2019, see report at https://osf.io/hupxr/
Transparency & Reproducibility�GIScience�https://osf.io/phmce/wiki/home/
Promotion�Acknowledge spectrum
12
The guidelines
Author guidelines�Data in Research Papers�Computational workflows �in Research Papers�Pre-submission checklist�Writing DASA section
Rationale/Motivation/Vision
Reviewer guidelines (what not to worry about)
Reproducibility reviewer guidelines
13
AGILE conference review process
Proceedings:�https://www.agile-giscience-series.net/review_process.html
Process documentation:�https://osf.io/7rjpe/
14
Reproducibility review after accept/reject decisions, triggered by regular reviewer
Reproducibility review & communication
Community conference
Badges on proceedings page
Presentation at conference
Read full report at https://osf.io/7rjpe/
Reproducibility review results
6 reproducibility reports published in 2020�🔥 9 more coming for 2021
16 (2020) not possible/not attempted (5 of which after communication with authors):
15
What can you do today?
16
Reproducible Research & Open Science
18
Quintana, D. S. (2020, November 28). Five things about open and reproducible science that every early career researcher should know. https://doi.org/10.17605/OSF.IO/DZTVQ
What can scientists do?
Take one step at a time.
Create and publish Research Compendia�(Your code is good enough!):�https://research-compendium.science/
Become a codechecker or reprohacker.
Strive to be an open science champion especially�if you’re junior in your field. We need to be the change, find communities.�[RIOT Science Club Talk by Gavin Buckinham; preprint by Sam Westwood]
19
Thanks!
Daniel Nüst�Institut für Geoinformatik�d.n@wwu.de | @nordholmen�0000-0002-0024-5046
Slides: http://bit.ly/hangout21-repro
Bonus slides for discussion
Reproducible AGILE and CODECHECK:�Highlights of Lessons learned
Spectrum or layers of reproducibility
Effect of guidelines at AGILE: improved reproducibility
Reproducibility reports/CODECHECK certificates full of recommendations for improvement, well received by authors, even included in revision before publication
Good practices spread slowly, establishing a process is tedious
Challenges for reproducibility reviewer: Inconsistencies and disconnects (figures), lack of documentation, unknown runtimes vs. no subsets of data, lack of repro guidance
Reproductions are rewarding and educational, matching expertises tricky
Safety net (👀), not security
22
Read full report at https://osf.io/7rjpe/
How does the future of reproducible research in peer review look like?
Reproducibility is possible, but disciplines/communities must agree what “peer review” entails and acknowledge the efforts (ECRs, RSEs) in a positive way.
Help each other! Move together as a community through disruptive changes.�Then reproducible research and peer review will get along just fine.
23
24
Independent execution of computations underlying research articles.
25
26
J. Leek’s tidypvals
“Notice�Anything�funny?”
The many problems of science
Publish or perish�Broken metrics (citations, JIF)�Structural change not considering � senior academics�Publication bias�Long-term funding for tools & infrastructure�HARKing�p-Hacking�Scholarly communication 1.0�Lack of reusability�Lack of transparency�Lack of reproducibility�Reinventing the wheel�Retraction practices�Not invented here syndrome�Fraud�Imposter syndrome�No “negative” citation�...
Open Science (OER, OA, OS, OPR)�Registered reports/preregistration�Altmetrics�Preprints�Leiden Manifesto�DORA�Vienna Principles�Citing data and software�Software papers�Data and software as products of research�RSEng & RSEs (software sustainability)�CRediT�Research Compendia�Ten Hot Topics Around Scholarly Publishing�Code review (PyOpenSci, ROpenSci, JOSS)�...�
27
28
Traditional and modern scientists
30
T
https://jakevdp.github.io/blog/2014/08/22/hacking-academia/
https://www.sciencemag.org/careers/2013/05/when-all-science-becomes-data-science
Π
Broad knowledge: across disciplines�collaborate with other experts, apply outside of own field
Deep knowledge: expertise and skills within a single field
Computer & method skills�statistics, reproducibility, programming, data science
Code Review
31
Boettiger, C., Chamberlain, S., Hart, E., & Ram, K. (2015). Building Software, Building Community: Lessons from the rOpenSci Project. Journal of Open Research Software, 3(1), e8. doi:10.5334/jors.bu
Reproducible computational research in journals & conferences
32
Findings
Overall
33
Read full report at https://osf.io/7rjpe/
Findings
Challenges for reproducibility reviewer:
�All efforts beyond mere workflow execution
34
Read full report at https://osf.io/7rjpe/
🙌
35
How to put your community on a path towards�more reproducibility in 5 easy hard steps
What can communities and institutions do?
Introduce reproducibility reviews - CODECHECK (or not)!
Workshops on RCR, ReproHacks
Provide support (R2S2, Anja)
Rewards and incentives
�Awareness > Change
36
Reproducibility review reports
37
Reproducibility review reports
38
Reproducibility review reports
39
The guidelines for
reproducibility reviewers (WIP)
Ideal vs. realistic
Role
Skills
Do’s & dont’s
40
🙌
41
How to put your community on a path towards�more reproducibility in 5 easy hard steps
The guidelines for data
“What if…” and Examples (not shown)
42
The guidelines�for workflows
Examples (not shown)
43
The guidelines for reproducibility reviewers (WIP)
Examples for “Do’s and Don’ts”:
44
Structural challenges
Metrics for acknowledging/measuring impact in science are broken (impact factor, ..) and they lead to publication bias, HARKing, p-Hacking, intransparency and lack of reproducibility
Leiden Manifesto: http://www.leidenmanifesto.org�DORA: https://sfdora.org �Vienna Principles: https://viennaprinciples.org
Acknowledging data and software as valuable products of research (instead of shoehorning software into papers)
Traditional and modern scientists
T
https://jakevdp.github.io/blog/2014/08/22/hacking-academia/
https://www.sciencemag.org/careers/2013/05/when-all-science-becomes-data-science
Π
Broad knowledge: across disciplines�collaborate with other experts, apply outside of own field
Deep knowledge: expertise and skills within a single field
Computer & method skills�statistics, reproducibility, programming, data science
≠
Images: https://pxhere.com/en/photo/477458 https://pxhere.com/en/photo/1087259 https://pxhere.com/en/photo/703106 https://pxhere.com/en/photo/103038
Professionalisation
Motivation
Back to 2010 The Software Sustainability Institute (SSI, UK) run a study (1000 randomly chosen researchers) …
“It's impossible to conduct�research without software,�say 7 out of 10 UK�researchers”
Motivation
A study of Nature papers from Jan-March 2016 reveals that
“32 of the 40 papers examined mention software, and the 32 papers contain 211 mentions of distinct pieces of software, for an average of 6.5 mentions per paper.”
[2] Nangia, Udit; Katz, Daniel S. (2017): Understanding Software in Research: Initial Results from Examining Nature and a Call for Collaboration. doi:10.1109/eScience.2017.78
Image: Event Horizon Telescope
DEVELOPERS
CONDITIONS FOR
LEAD TO
EDUCATION OF
RESEARCHERS USING
”Software is 95% human and only�5% code” *
* Eric Albers, CCC2019, https://media.ccc.de/v/thms-49-ber-die-nachhaltigkeit-von-software �Bilder © H. Seibold, S. Janosch, OSD2019
RSEng = create research software
RSEs = people behind research software
RSEs ≠ IT !!!
Researcher uses scripts for data analysis and needs working stable software for her work. She learns what is necessary to achieve her research goals.
Reproducibility guru dives deeply into manifold software and tools to make his research reproducible and develops his own software in a sustainable way.
Person for tough problems knows how to solve all kinds of computer-related issues; he was not hired for that, but enjoys to help and spends time to get to the bottom of other people’s challenges.
Geek writes software as part of her research project and would like to code more, but must keep an eye on her career in science and needs to write papers.
Software developer was hired to implement software for a research project and contributes to large collaborative software projects to realise the next generation of digital infrastructure for science.