2020 Reproducibility Challenge

A significant challenge across scientific fields is the reproducibility of research results, and third-party assessment of such reproducibility. Ensuring that results can be reliably reproduced is no small task: computational environments may vary drastically and can change over time, rendering code unable to run; specialized workflows might require specialized infrastructure not easily available; sensitive projects might involve data that cannot be directly shared; the robustness of algorithmic decisions and parameter selections varies widely; data collection methods may include crucial steps (e.g. wrangling, cleaning, missingness mitigation strategies, preprocessing) where choices are made but not well-documented. Yet a cornerstone of science remains the ability to verify and validate research findings, so it is important to find ways to overcome these challenges.

The first MIDAS Reproducibility Challenge was held in the first 8 months of 2020. Our goal was to highlight high-quality, reproducible work at the University of Michigan by collecting examples of best practices across diverse fields. Besides incentivizing reproducible workflows and enabling a deeper understanding of issues of reproducibility, the results of the challenge also provide templates that others can follow.

Judges

  • Jake Carlson: Manager, Deep Blue Repositories and Research Data Services, U-M Libraries
  • H.V. Jagadish: Director, MIDAS, and Professor, Computer Science and Engineering, CoE
  • Matthew Kay: Assistant Professor, School of Information
  • Jing Liu: Managing Director, MIDAS
  • Josh Pasek: Assistant Professor, Communication and Media, LSA
  • Brian Puchala: Assistant Research Scientist, Materials Science and Engineering, CoE
  • Arvind Rao: Associate Professor, Computational Medicine and Bioinformatics, and Radiation Oncology, Med. School

MIDAS Reproducibility Challenge Winners

View Reproducibility Day Recording

Category B

Exact reproducibility

Everyday Reproducibility: A multi-pronged approach to ensure analyses are fully reproducible, easy to access, and easy to use

Johann A. Gagnon-Bartsch
Statistics, University of Michigan

Yotam Shem-Tov
Economics, UCLA

Gregory J. Hunt
Mathematics, College of William & Mary

Mark A. Dane
Biomedical Engineering, Oregon Health & Science University

James E. Korkola
Biomedical Engineering, Oregon Health & Science University

Laura M. Heiser
Biomedical Engineering, Oregon Health & Science University

Saskia Freytag
Medical Biology, University of Melbourne

Melanie Bahlo
Medical Biology, University of Melbourne

Category B

Exact reproducibility

Statistical code sharing: a guide for clinical researchers

Thomas S. Valley
Internal Medicine, University of Michigan

Neil Kamdar
Institute for Healthcare Policy and Innovation, University of Michigan

Wyndy L. Wiitala
VA Center for Clinical Management Research

Andrew M. Ryan
Institute for Healthcare Policy and Innovation, University of Michigan

Sarah M. Seelye
VA Center for Clinical Management Research

Akbar K. Waljee
Internal Medicine, University of Michigan

Brahmajee K. Nallamothu
Internal Medicine, University of Michigan

Category C

Generalizable tools

Reproducible Materials Simulation and Analysis Workflows

Sharon C. Glotzer
Chemical Engineering, Biointerfaces Institute, University of Michigan

Karen Coulter
Chemical Engineering, Glotzer Group Lab, University of Michigan

Joshua Anderson
Chemical Engineering, Biointerfaces Institute, University of Michigan

Timothy Moore
Chemical Engineering, Biointerfaces Institute, University of Michigan

Allen LaCour
Chemical Engineering, Biointerfaces Institute, University of Michigan

Kelly Wang
Macromolecular Science and Engineering, Biointerfaces Institute, University of Michigan

Category D

Robustness

Translating Strategies for Promoting Engagement in Mobile Health: A Micro-randomized Feasibility Trial

Inbal Nahum-Shani
ISR, University of Michigan

Mashfiqui Rabbi
Statistics, Harvard University

Jamie Yap
ISR, University of Michigan

Meredith L. Philyaw-Kotov
Psychiatry and Addiction Center, University of Michigan

Predrag Klasnja
School of Information, University of Michigan

Erin E. Bonar
Psychiatry and Addiction Center, University of Michigan

Rebecca M. Cunningham
Vice President of Research, University of Michigan

Susan A. Murphy
Statistics & Computer Science, Harvard University

Maureen A. Walton
Psychiatry and Addiction Center, University of Michigan

MIDAS Reproducibility Challenge Honorable Mentions

Category A

Theory

INTRIGUE: Quantify and Control Reproducibility in High-throughput Experiments

Xiaoquan (William) Wen 

Biostatistics, University of Michigan

Yi Zhao 

Biostatistics, University of Michigan

Matthew Sampson 

Pediatrics, Harvard Medical School

Category C

Generalizable tools

C2Metadata: Continuous Capture of Metadata for Statistical Data

Jie Song
Computer Science and Engineering, University of Michigan

George Alter
ICPSR, University of Michigan

Category D

Robustness

The eXtensible ontology development (XOD) principles and tool implementation to support ontology interoperability.

Yongqun “Oliver” He
Computational Medicine and Bioinformatics, University of Michigan Medical School

Edison Ong
Computational Medicine and Bioinformatics, University of Michigan Medical School

Matthias Kretzler
Computational Medicine and Bioinformatics, University of Michigan Medical School

Brian Athey
Chair of the Department of Computational Medicine and Bioinformatics in the University of Michigan Medical School

Category E

Assessments of Reproducibility

Replicate.Education: Lessons Learned Building a Platform for Educational Data Science Replications

Christopher Brooks
School of Information, University of Michigan

Josh Gardner
Computer Science & Engineering, University of Washington

Ryan S. Baker
Teaching, Learning, and Leadership Division, University of Pennsylvania