A significant challenge across scientific fields is the reproducibility of research results, and third-party assessment of such reproducibility. Ensuring that results can be reliably reproduced is no small task: computational environments may vary drastically and can change over time, rendering code unable to run; specialized workflows might require specialized infrastructure not easily available; sensitive projects might involve data that cannot be directly shared; the robustness of algorithmic decisions and parameter selections varies widely; data collection methods may include crucial steps (e.g. wrangling, cleaning, missingness mitigation strategies, preprocessing) where choices are made but not well-documented. Yet a cornerstone of science remains the ability to verify and validate research findings, so it is important to find ways to overcome these challenges.
The first MIDAS Reproducibility Challenge was held in the first 8 months of 2020. Our goal was to highlight high-quality, reproducible work at the University of Michigan by collecting examples of best practices across diverse fields. Besides incentivizing reproducible workflows and enabling a deeper understanding of issues of reproducibility, the results of the challenge also provide templates that others can follow.
Judges
- Jake Carlson: Manager, Deep Blue Repositories and Research Data Services, U-M Libraries
- H.V. Jagadish: Director, MIDAS, and Professor, Computer Science and Engineering, CoE
- Matthew Kay: Assistant Professor, School of Information
- Jing Liu: Managing Director, MIDAS
- Josh Pasek: Assistant Professor, Communication and Media, LSA
- Brian Puchala: Assistant Research Scientist, Materials Science and Engineering, CoE
- Arvind Rao: Associate Professor, Computational Medicine and Bioinformatics, and Radiation Oncology, Med. School
MIDAS Reproducibility Challenge Winners
Category B
Exact reproducibility
Everyday Reproducibility: A multi-pronged approach to ensure analyses are fully reproducible, easy to access, and easy to use
Johann A. Gagnon-Bartsch
Statistics, University of Michigan
Yotam Shem-Tov
Economics, UCLA
Gregory J. Hunt
Mathematics, College of William & Mary
Mark A. Dane
Biomedical Engineering, Oregon Health & Science University
James E. Korkola
Biomedical Engineering, Oregon Health & Science University
Laura M. Heiser
Biomedical Engineering, Oregon Health & Science University
Saskia Freytag
Medical Biology, University of Melbourne
Melanie Bahlo
Medical Biology, University of Melbourne
Category B
Exact reproducibility
Statistical code sharing: a guide for clinical researchers
Thomas S. Valley
Internal Medicine, University of Michigan
Neil Kamdar
Institute for Healthcare Policy and Innovation, University of Michigan
Wyndy L. Wiitala
VA Center for Clinical Management Research
Andrew M. Ryan
Institute for Healthcare Policy and Innovation, University of Michigan
Sarah M. Seelye
VA Center for Clinical Management Research
Akbar K. Waljee
Internal Medicine, University of Michigan
Brahmajee K. Nallamothu
Internal Medicine, University of Michigan
Category C
Generalizable tools
Reproducible Materials Simulation and Analysis Workflows
Sharon C. Glotzer
Chemical Engineering, Biointerfaces Institute, University of Michigan
Karen Coulter
Chemical Engineering, Glotzer Group Lab, University of Michigan
Joshua Anderson
Chemical Engineering, Biointerfaces Institute, University of Michigan
Timothy Moore
Chemical Engineering, Biointerfaces Institute, University of Michigan
Allen LaCour
Chemical Engineering, Biointerfaces Institute, University of Michigan
Kelly Wang
Macromolecular Science and Engineering, Biointerfaces Institute, University of Michigan
Category D
Robustness
Translating Strategies for Promoting Engagement in Mobile Health: A Micro-randomized Feasibility Trial
Inbal Nahum-Shani
ISR, University of Michigan
Mashfiqui Rabbi
Statistics, Harvard University
Jamie Yap
ISR, University of Michigan
Meredith L. Philyaw-Kotov
Psychiatry and Addiction Center, University of Michigan
Predrag Klasnja
School of Information, University of Michigan
Erin E. Bonar
Psychiatry and Addiction Center, University of Michigan
Rebecca M. Cunningham
Vice President of Research, University of Michigan
Susan A. Murphy
Statistics & Computer Science, Harvard University
Maureen A. Walton
Psychiatry and Addiction Center, University of Michigan
MIDAS Reproducibility Challenge Honorable Mentions
Category A
Theory
INTRIGUE: Quantify and Control Reproducibility in High-throughput Experiments
Xiaoquan (William) Wen
Biostatistics, University of Michigan
Yi Zhao
Biostatistics, University of Michigan
Matthew Sampson
Pediatrics, Harvard Medical School
Category C
Generalizable tools
C2Metadata: Continuous Capture of Metadata for Statistical Data
Jie Song
Computer Science and Engineering, University of Michigan
George Alter
ICPSR, University of Michigan
Category D
Robustness
The eXtensible ontology development (XOD) principles and tool implementation to support ontology interoperability.
Yongqun “Oliver” He
Computational Medicine and Bioinformatics, University of Michigan Medical School
Edison Ong
Computational Medicine and Bioinformatics, University of Michigan Medical School
Matthias Kretzler
Computational Medicine and Bioinformatics, University of Michigan Medical School
Brian Athey
Chair of the Department of Computational Medicine and Bioinformatics in the University of Michigan Medical School
Category E
Assessments of Reproducibility
Replicate.Education: Lessons Learned Building a Platform for Educational Data Science Replications
Christopher Brooks
School of Information, University of Michigan
Josh Gardner
Computer Science & Engineering, University of Washington
Ryan S. Baker
Teaching, Learning, and Leadership Division, University of Pennsylvania