Enabling Reproducibility in Computational and Data-enabled Science

Similar documents
Computational Reproducibility in Medical Research:

Advancing Data Science through a Lifecycle Approach

Scientific Transparency, Integrity, and Reproducibility

Reproducibility Interest Group

The Importance of Scientific Reproducibility in Evidence-based Rulemaking

Elements of Scholarly Discourse in a Digital World

Reproducibility in Computationally-Enabled Research: Integrating Tools and Skills

The Value of Computational Transparency

Enhancing Reproducibility for Computational Methods

Law & Ethics of Big Data Research Dissemination

Reproducibility in Computational Science: A Computable Scholarly Record

A CyberInfrastructure Wish List for Statistical and Data Driven Discovery

When Should We Trust the Results of Data Science?

Reproducibility in Computational Science: Opportunities and Challenges

How Science is Different: Digitizing for Discovery

Document Downloaded: Wednesday September 16, June 2013 COGR Meeting Afternoon Presentation - Victoria Stodden. Author: Victoria Stodden

The Reproducible Research Movement in Statistics

Open Licensing and Science Policy

Disseminating Numerically Reproducible Research

Tools for Academic Research: Resolving the Credibility Crisis in Computational Science

Open Methodology and Reproducibility in Computational Science

Applying the Creative Commons Philosophy to Scientific Innovation

Thoughts on Reimagining The University. Rajiv Ramnath. Program Director, Software Cluster, NSF/OAC. Version: 03/09/17 00:15

Two Ideas for Open Science (forget Open Data!)

The Impact of Computational Science on the Scientific Method

Scientific Reproducibility and Software

Opening Science & Scholarship

RECOMMENDATIONS. COMMISSION RECOMMENDATION (EU) 2018/790 of 25 April 2018 on access to and preservation of scientific information

Software Patents as a Barrier to Scientific Transparency: An Unexpected Consequence of Bayh-Dole

APEC Internet and Digital Economy Roadmap

Open Science for the 21 st century. A declaration of ALL European Academies

December 10, Why HPC? Daniel Lucio.

Software Patents as a Barrier to Scientific Transparency: An Unexpected Consequence of Bayh-Dole

g~:~: P Holdren ~\k, rjj/1~

Trends in. Archives. Practice MODULE 8. Steve Marks. with an Introduction by Bruce Ambacher. Edited by Michael Shallcross

New forms of scholarly communication Lunch e-research methods and case studies

President Barack Obama The White House Washington, DC June 19, Dear Mr. President,

ICSU World Data System Strategic Plan Trusted Data Services for Global Science

Executive Summary Industry s Responsibility in Promoting Responsible Development and Use:

University of Massachusetts Amherst Libraries. Digital Preservation Policy, Version 1.3

Reproducible Research for Scientific Computing: Tools and Strategies for Changing the Culture

Enabling FAIR Data in the Earth, Space, and Environmental Sciences

National Medical Device Evaluation System: CDRH s Vision, Challenges, and Needs

Building an Infrastructure for Data Science Data and the Librarians Role. IAMSLIC, Anchorage August, 2012 Linda Pikula, NOAA and IODE GEMIM

Prepared in a cooperative effort by: Elsevier IEEE The IET

Digitisation Plan

STRATEGIC FRAMEWORK Updated August 2017

High Performance Computing Systems and Scalable Networks for. Information Technology. Joint White Paper from the

COMMISSION RECOMMENDATION. of on access to and preservation of scientific information. {SWD(2012) 221 final} {SWD(2012) 222 final}

Library Special Collections Mission, Principles, and Directions. Introduction

Working Paper Series of the German Data Forum (RatSWD)

What is a collection in digital libraries?

Liquid Benchmarks. Sherif Sakr 1 and Fabio Casati September and

HTA Position Paper. The International Network of Agencies for Health Technology Assessment (INAHTA) defines HTA as:

14 th Berlin Open Access Conference Publisher Colloquy session

Digital Preservation Strategy Implementation roadmaps

Testimony of Dr. Victoria Stodden Columbia University. Before the House Committee on Science, Space and Technology Subcommittee on Research

December Eucomed HTA Position Paper UK support from ABHI

FDA Centers of Excellence in Regulatory and Information Sciences

Evolution of Data Creation, Management, Publication, and Curation in the Research Process

TECHNICAL AND OPERATIONAL NOTE ON CHANGE MANAGEMENT OF GAMBLING TECHNICAL SYSTEMS AND APPROVAL OF THE SUBSTANTIAL CHANGES TO CRITICAL COMPONENTS.

FROM BRAIN RESEARCH TO FUTURE TECHNOLOGIES. Dirk Pleiter Post-H2020 Vision for HPC Workshop, Frankfurt

NCRIS Capability 5.7: Population Health and Clinical Data Linkage

Expectations around Impact in Horizon 2020

Data and Knowledge as Infrastructure. Chaitan Baru Senior Advisor for Data Science CISE Directorate National Science Foundation

Open Data, Open Science, Open Access

Over the 10-year span of this strategy, priorities will be identified under each area of focus through successive annual planning cycles.

The Blockchain Ethical Design Framework

Open Science policy and infrastructure support in the European Commission. Joint COAR-SPARC Conference. Porto, 15 April 2015

Strategic Plan Public engagement with research

Reproducible Research in Computational Science

How do you teach AI the value of trust?

Senate Bill (SB) 488 definition of comparative energy usage

A POLICY in REGARDS to INTELLECTUAL PROPERTY. OCTOBER UNIVERSITY for MODERN SCIENCES and ARTS (MSA)

ADVANCING KNOWLEDGE. FOR CANADA S FUTURE Enabling excellence, building partnerships, connecting research to canadians SSHRC S STRATEGIC PLAN TO 2020

NEES CYBERINFRASTRUCTURE: A FOUNDATION FOR INNOVATIVE RESEARCH AND EDUCATION

Technology and Innovation in the NHS Scottish Health Innovations Ltd

UN-GGIM Future Trends in Geospatial Information Management 1

Why? A Documentation Consortium Ted Habermann, NOAA. Documentation: It s not just discovery... in global average

EarthCube Conceptual Design: Enterprise Architecture for Transformative Research and Collaboration Across the Geosciences

Framework Programme 7

The European Approach

EXPLORATION DEVELOPMENT OPERATION CLOSURE

Advanced Cyberinfrastructure for Science, Engineering, and Public Policy 1

Department of Arts and Culture NATIONAL POLICY ON THE DIGITISATION OF HERITAGE RESOURCES

preface Motivation Figure 1. Reality-virtuality continuum (Milgram & Kishino, 1994) Mixed.Reality Augmented. Virtuality Real...

A New Path for Science?

Deep Learning Overview

CSTA K- 12 Computer Science Standards: Mapped to STEM, Common Core, and Partnership for the 21 st Century Standards

Office of Science and Technology Policy th Street Washington, DC 20502

GENEVA COMMITTEE ON DEVELOPMENT AND INTELLECTUAL PROPERTY (CDIP) Fifth Session Geneva, April 26 to 30, 2010

University of Southern California Guidelines for Assigning Authorship and for Attributing Contributions to Research Products and Creative Works

BI TRENDS FOR Data De-silofication: The Secret to Success in the Analytics Economy

Science as an Open Enterprise

Methodology for Agent-Oriented Software

Cisco Live Healthcare Innovation Roundtable Discussion. Brendan Lovelock: Cisco Brad Davies: Vector Consulting

Climate Change Innovation and Technology Framework 2017

Information Communication Technology

Open Science. challenge and chance for medical librarians in Europe.

International Symposium on Knowledge Communities 2012

Transcription:

Enabling Reproducibility in Computational and Data-enabled Science Victoria Stodden School of Information Sciences University of Illinois at Urbana-Champaign EPFl Seminar October 25, 2018

Agenda 1. Framing Reproducibility in the Computational Sciences 2. How Much of a Problem is Computational Reproducibility? 3. Infrastructure for Computational and Data-enabled Experiments 4. Thoughts on Data Science as a Scientific Field

Skepticism and Boyle s Idea for Scientific Communication Skepticism interpreted to mean claims can be independently verified, which requires transparency of the research process in publications. Standards established by Transactions of the Royal Society in the 1660 s (Robert Boyle).

Now: Technology Impacts Transparency Big Data / Data Driven Discovery: e.g. high dimensional data. International Data Corportation estimates that data generated from connected devices will exceed 40 trillion gigabytes by 2025. Computational Power: simulation of the complete evolution of a physical system, systematically varying parameters, Software as a first class scholarly object: Deep intellectual contributions now encoded only in software. The software contains ideas that enable biology CSHL Keynote; Dr. Lior Pachter, Caltech Stories from the Supplement from the Genome Informatics meeting 11/1/2013 https://youtu.be/5nifibnbe8o

Querying the Scholarly Record Show a table of effect sizes and p-values in all phase-3 clinical trials for Melanoma published after 1994; Name all of the image denoising algorithms ever used to remove white noise from the famous Barbara image, with citations; List all of the classifiers applied to the famous acute lymphoblastic leukemia dataset, along with their type-1 and type-2 error rates; Create a unified dataset containing all published whole-genome sequences identified with mutation in the gene BRCA1; Randomly reassign treatment and control labels to cases in published clinical trial X and calculate effect size. Repeat many times and create a histogram of the effect sizes. Perform this for every clinical trial published in the year 2003 and list the trial name and histogram side by side. Courtesy of Donoho and Gavish 2012

Parsing Reproducibility Empirical Reproducibility Statistical Reproducibility Computational Reproducibility V. Stodden, IMS Bulletin (2013)

Empirical Reproducibility

Statistical Reproducibility False discovery, p-hacking (Simonsohn 2012), file drawer problem, overuse and mis-use of p-values, lack of multiple testing adjustments. Low power, poor experimental design, nonrandom sampling, Data preparation, treatment of outliers, re-combination of datasets, insufficient reporting/tracking practices, inappropriate tests or models, model misspecification, Model robustness to parameter changes and data perturbations,

It is common now to consider computation as a third branch of science, besides theory and experiment. This book is about a new, fourth paradigm for science based on data-intensive computing.

Computational Reproducibility Traditionally two branches to the scientific method: Branch 1 (deductive): mathematics, formal logic, Branch 2 (empirical): statistical analysis of controlled experiments. Now, new branches due to technological changes? Branch 3,4? (computational): large scale simulations / data driven computational science.

The Ubiquity of Error The central motivation for the scientific method is to root out error: Deductive branch: the well-defined concept of the proof, Empirical branch: the machinery of hypothesis testing, appropriate statistical methods, structured communication of methods and protocols. Claim: Computation presents only a potential third/fourth branch of the scientific method (Donoho, Stodden, et al. 2009), until the development of comparable standards.

Really Reproducible Research Inspired by Stanford Professor Jon Claerbout, from 1992: The idea is: An article about computational science in a scientific publication is not the scholarship itself, it is merely advertising of the scholarship. The actual scholarship is the complete... set of instructions [and data] which generated the figures. David Donoho, 1998 Note the difference between: reproducing the computational steps and, replicating the experiments independently including data collection and software implementation. (Both required)

Stodden and Krafczyk 2018, submitted

The digital age in science Claim 1: Virtually all published discoveries today have a computational component. Claim 2: Dissemination of research results generally follows the same tradition established for noncomputational reserach, leading to reproducibility concerns.

INSIGHTS POLICY FORUM REPRODUCIBILITY Enhancing reproducibility for computational methods Data, code, and workflows should be available and cited By Victoria Stodden, 1 Marcia McNutt, 2 David H. Bailey, 3 Ewa Deelman, 4 Yolanda Gil, 4 Brooks Hanson, 5 Michael A. Heroux, 6 John P.A. Ioannidis, 7 Michela Taufer 8 Over the past two decades, computational methods have radically changed the ability of researchers from all areas of scholarship to process and analyze data and to simulate complex systems. But with these advances come challenges that are contributing to broader concerns over irreproducibility in the scholarly literature, among them the lack of transparency in disclosure of computational methods. Current reporting methods are often uneven, incomplete, and still evolving. We present a novel set of Reproducibility Enhancement Principles (REP) targeting disclosure challenges involving computation. These recommendations, which build upon more general proposals from the Transparency and Openness Promotion (TOP) guidelines (1) and recommendations for field data (2), emerged from workshop discussions among funding agencies, publishers and journal editors, industry participants, and researchers repreto understanding how computational results were derived and to reconciling any differences that might arise between independent replications (4). We thus focus on the ability to rerun the same computational steps on the same data the original authors used as a minimum dissemination standard (5, 6), which includes workflow information that explains what raw data and intermediate results are input to which computations (7). Access to the data and code that underlie discoveries can also enable downstream scientific contributions, such as meta-analyses, reuse, and other efforts that include results from multiple studies. RECOMMENDATIONS Share data, software, workflows, and details of the computational environment that generate published findings in open trusted repositories. The minimal components that enable independent regeneration of computational results are the data, the computational steps that produced the findings, and the workflow describing how to generate the results using the data and code, including parameter settings, random number seeds, make files, or Sufficient metadata should be provided for someone in the field to use the shared digital scholarly objects without resorting to contacting the original authors (i.e., http:// bit.ly/2fvwjph). Software metadata should include, at a minimum, the title, authors, version, language, license, Uniform Resource Identifier/DOI, software description (including purpose, inputs, outputs, dependencies), and execution requirements. To enable credit for shared digital scholarly objects, citation should be standard practice. All data, code, and workflows, including software written by the authors, should be cited in the references section (10). We suggest that software citation include software version information and its unique identifier in addi- Access to the computational steps taken to process data and generate findings is as important as access to data themselves. Stodden, Victoria, et al. Enhancing reproducibility for computational methods. Science 354(6317) (2016)

7: Funding agencies should instigate new research programs and pilot studies. Reproducibility Enhancement Principles 1: To facilitate reproducibility, share the data, software, workflows, and details of the computational environment in open repositories. 2: To enable discoverability, persistent links should appear in the published article and include a permanent identifier for data, code, and digital artifacts upon which the results depend. 3: To enable credit for shared digital scholarly objects, citation should be standard practice. 4: To facilitate reuse, adequately document digital scholarly artifacts. 5: Journals should conduct a Reproducibility Check as part of the publication process and enact the TOP Standards at level 2 or 3. 6: Use Open Licensing when publishing digital scholarly objects.

Fostering Integrity in Research RECOMMENDATION SIX: Through their policies and through the development of supporting infrastructure, research sponsors and science, engineering, technology, and medical journal and book publishers should ensure that information sufficient for a person knowledgeable about the field and its techniques to reproduce reported results is made available at the time of publication or as soon as possible after publication. RECOMMENDATION SEVEN: Federal funding agencies and other research sponsors should allocate sufficient funds to enable the longterm storage, archiving, and access of datasets and code necessary for the replication of published findings. Fostering Integrity in Research, National Academies of Sciences, Engineering, and Medicine, 2017

Testing the Claims: How Much of a Problem is Computational Reproducibility?

Study 1: Effectiveness of Artifact Access February 11, 2011: on Demand All data necessary to understand, assess, and extend the conclusions of the manuscript must be available to any reader of Science. All computer codes involved in the creation or analysis of data must also be available to any reader of Science. After publication, all reasonable requests for data and materials must be fulfilled... Survey of publications in Science Magazine from Feb 11, 2011 to June 29, 2012 inclusive. Obtained a random sample of 204 scientific articles with computational findings. Asked for the data and code! Stodden et al., Journal Policy for Computational Reproducibility, PNAS, March 2018

Responses to Artifact Requests (n=204) No response Contact to another person Asks for reasons Refusal to share Directed back to Supplemental Materials Unfulfilled promise to follow up Email bounced Impossible to share Shared data and code 26% 11% 11% 7% 3% 3% 2% 2% 36% Total 100% 12% of the articles provided direct access to code/data

Computational Replication Rates We were able to obtain data and code from the authors of 89 articles in our sample of 204, overall artifact recovery rate estimate: 44%, 95% confidence interval [0.36, 0.50] Of the 56 articles we deemed potentially reproducible, we randomly choose 22 to attempt replication, and all but one provided enough information to do so. overall computational reproducibility estimate: 26%, 95% confidence interval [0.20, 0.32]

Study 2: Reproducibility in Computational Physics Examined 306 articles in the Journal of Computational Physics published between Oct 1 2016 and Feb 15 2017. Are artifacts available (can we obtain them)? Do they replicate the published results? Artifact Access via Information in the Article (n=306) No discussion in the article and no artifacts made available 58.8% Some discussion of artifacts none made available 35.6% Some artifacts made available 5.6% Stodden, Krafczyk, and Bhaskar, Enabling the Verification of Computational Results: An Empirical Evaluation of Computational Reproducibility, Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems, 2018

ICERM Article Information Evaluation Criteria Implementation (n=55) A precise statement of assertions to be made in the paper 100% Full statement (or valid summary) of experimental results 100% Salient details of data reduction & statistical analysis methods 73% Necessary run parameters were given 86% A statement of the computational approach and why it tests the proposed hypotheses 100% Complete statements of, or references to, algorithms and salient software details 63% Discussion of the adequacy of parameters such as precision level and grid resolution 76% Proper citation of all code and data used, including that generated by the authors 4% Availability of computer code, input and output data, with reasonable level of documentation 4% Avenues of exploration examined throughout development, including negative findings 0% Instructions for repeating computational experiments described in the article 79% Precise functions were given, with settings 11% Salient test environment details: hardware, system software, and number of processors used 24%

Attempts to Replicate Results (n=55) Computational Reproducibility Evaluation (n=55) Straightforward to reproduce with minimal effort 0% Minor difficulty in reproducing 0% Reproducible after some tweaking 9.1% Could reproduce with fairly substantial skill and knowledge 16.4% Reproducible with substantial intellectual effort 12.7% Reproducible with substantial tedious effort 3.6% Difficult to reproduce because of unavoidable inherent complexity 3.6% Nearly impossible to reproduce 3.6% Impossible to reproduce 50.9%

Infrastructure for Computational Research

Example: AIM: An Abstraction for Improving Machine learning We developed infrastructure for comparative Machine Learning. Our goal: List all of the classifiers applied to the famous acute lymphoblastic leukemia dataset, along with their misclassification rates. See Stodden, Wu, and Sochat, AIM: An Abstraction For Improving Machine Learning Prediction," IEEE Data Science Workshop, June 2018

Our (Naive) Expectation We hoped to apply the machine learning algorithms from the literature to the Golub dataset, in the 5 cases we identified. However, we found that the articles implemented (at least) three steps, each varying from one article to the next: 1. data preprocessing, 2. feature selection, 3. application of machine learning algorithm.

Computational Steps in the 5 Articles

https: //github.com/aim-project/aim-manuscript AIM: Using Structured Containers We compared models via classification rates: We then designed a container image to run the preprocessing/feature selection (PPFS) separately from the model fitting/prediction (P) step.

Query Conclusions Lengthy to obtain comparable estimates (200+ student hours) Many points of variability: starting dataset; preprocessing steps; feature selection methods; algorithm choice; parameter tuning... Details not well-captured in the traditional article, making comparisons difficult or impossible. Would be easier if: there was prior agreement on the dataset, prior agreement on hold-out data for testing, full disclosure of preprocessing and feature selection steps, full disclosure of algorithm application and parameter tuning.

Abstraction for Improving Machine learning (AIM) Agreement on datasets prior to analysis, conferences around those datasets, Hold-out data held by a neutral third party (e.g. NIST), not seen by researchers, Researchers distinguish and specify feature selection and preprocessing vs learning algorithm application, Send code to the third party who returns your misclassification rate on the test data. Side effect: training data and code/algorithm shared.

Infrastructure Solutions Research Environments and Document Enhancement Tools StatTag.org SHARE Code Ocean Jupyter Verifiable Computational Research Sweave Cyverse NanoHUB knitr SOLE Open Science Framework Vistrails Collage Authoring Environment GenePattern IPOL Popper Workflow Systems Sumatra torch.ch Whole Tale flywheel.io Taverna Wings Pegasus CDE binder.org Kurator Kepler Everware Reprozip Galaxy Dissemination Platforms ResearchCompendia.org DataCenterHub RunMyCode.org ChameleonCloud Occam RCloud TheDataHub.org Madagascar Wavelab Sparselab

Quantitative Programming Environments Define and create Quantitative Programming Environments to (easily) manage the conduct of massive computational experiments and expose the resulting data for analysis and structure the subsequent data analysis Better transparency will allow people to run much more ambitious computational experiments. And better computational experiment infrastructure will allow researchers to be more transparent. See Donoho and Stodden, Reproducible Research in the Mathematical Sciences Princeton Companion to Applied Mathematics, 2015

Three Principles for Cyberinfrastructure 1. Supporting scientific norms enable new discoveries AND permit others to reproduce the computational findings, reuse and combine digital outputs.. 2. Supporting best practices in science CI in support of science should embed and encourage best practices in scientific research and discovery. 3. Taking a holistic approach to CI the complete end-to-end research pipeline should be considered for interoperability and the effective implementation of 1 and 2. See Stodden, Miguez, Seiler, ResearchCompendia.org: Cyberinfrastructure for Reproducibility and Collaboration in Computational Science CiSE 2015

Whole Tale Project The Whole Tale project seeks to leverage & contribute to existing cyberinfrastructure and tools to support the whole research story, and provide access to data and computing power. Integrate tools to simplify usage and promote best practices B. Ludaescher, K. Chard, N. Gaffney, M. B. Jones, J. Nabrzyski, V. Stodden, M. Turk NSF CC*DNI DIBBS awarded 2016: 5 Institutions for 5 Years ($5M total)

Whole Tale Project Goals Expose existing digital resources to researchers through popular frontends (Jupyter, RStudio,..) Develop necessary software glue for seamless access to different CI-backend capabilities Enhance conceptualization-to-publication lifecycle by empowering scientists to create computational narratives in their usual programming environments Embed reproducibility and best/better practices in the digital research environment

Whole Tale: What s in a Name? (1) Whole Tale Whole Story: Support (computational & data) scientists along the complete research lifecycle from experiment to publication and back! (2) Whole Tale Long Tail of Science: Engage researchers of all project scales image from Ferguson et al. 2014 doi:10.1038/nn.3838

Tales Tales are the final research output from a project, capturing the complete provenance of a particular activity/analysis within the system: easily sharable with others, publishable in repositories, associated with persistent identifiers, linked to publications, execute in the same state as it was when first published, acts as a starting point for research.

Try it! We released a public version of the Whole Tale platform! http://wholetale.readthedocs.io/users_guide/index.html Feedback is very welcome at feedback@wholetale.org and/ or at https://github.com/whole-tale/whole-tale/issues

ezdmp NSF funded project to provide structured guidance for a second generation data management plan. EAGER: Collaborative Proposal: Supporting Public Access to Supplemental Scholarly Products Generated from Grant Funded Research (2016). Helen M. Berman, Rutgers Kerstin Lehnert, Columbia Victoria Stodden, UIUC Maggie Gabanyi, Rutgers Vicki Ferrini, Columbia

ezdmp Progress Examined selected data management plans to understand gaps, successes, and patterns of use in IEDA DMP Tool. Reviewed the patterns exhibited by DMP creators using the IEDA DMP Tool Implement into IEDA ( ezdmp ) Try our prototype! http://dev.ezdmp.org and we have a feedback rubric here https://goo.gl/forms/ CaEB3ddJ3iuUmpxS2

The Future of Data and Computationally-enabled Research The future: a major effort to develop infrastructure that supports the entire Lifecycle of Data Science, from the hardware through applications, to ethics. Infrastructure promotes good scientific practice downstream like transparency and reproducibility. People will use such infrastructure not out of ethics or hygiene, but because this is a corollary of managing massive amounts of computational work, and used because it enables efficiency and productivity, and discovery.

Progress on computational reproducibility is enabled through coordination by a variety of stakeholders. Scientific Societies Funders (policy) Publishers (TOP guidelines) Regulatory Bodies (OSTP Memos) Researchers (processes) The Public/Press Universities/institutions (hiring/promotion) Universities/libraries (empowering w/tools, support)

The LifeCycle of Data Science as a Framework

Lifecycle of Data Berman et al., Realizing the Potential of Data Science, CACM, April 2018

Lifecycle of Data Science Framework to incorporate data science contributions from different fields, Explicit emphasis on re-use and reproducibility, Explicit emphasis on computational tools (e.g. Kubernetes), hardware (e.g. Google Edge TPUs) and software (e.g. Jupyter Notebooks) Surfaces ethics (human subjects, privacy), social context (interpretations of bias ), scholarly communication and reproducible research.

Lifecycle of Data Science: An Abstraction the study of data science ethics, documentation and metadata creation, best practices, policy; the science of data science application level experimental design data generation and collection data exploration and hypothesis generation data cleaning and organization feature selection and data preparation model building and statistical inference simulation and cross-validation visualization publication and artifact preservation / archiving infrastructure level notebooks and workflow software database structures workflow software and preregistration tools data management tools notebooks, workflow software; containerization tools notebooks, inference languages notebooks notebooks, visualization software workflow software, artifact linking tools system level hardware, cloud computing infrastructure, systems and system management, data structures, storage

Lifecycle of Data Science: An Abstraction the study of data science ethics, documentation and metadata creation, best practices, policy; the science of data science application level experimental design data generation and collection data exploration and hypothesis generation data cleaning and organization feature selection and data preparation model building and statistical inference simulation and cross-validation visualization publication and artifact preservation / archiving infrastructure level notebooks and workflow software database structures workflow software and preregistration tools data management tools notebooks, workflow software; containerization tools notebooks, inference languages notebooks notebooks, visualization software workflow software, artifact linking tools system level hardware, cloud computing infrastructure, systems and system management, data structures, storage

Challenges for the Research Community Funders are now funding cyberinfrastructure more expansively in addition to traditional foundational research; More and more fields (e.g. cybersecurity (LASER2014), networks (SIGCOMM2017)) are becoming empirical, not just transformed by opportunities due to data; Leveraging cyberinfrastructure and methods across fields (e.g. Computational Photo-Scatterography); how to reward, promote, fund; New research areas: Datasets as discovery drivers (ImageNet; Wiki* text datasets); Scientific software resilience and data preserve/destroy decisions; Technology transfer beyond the university. managing massive computational projects requires better, more transparent tools; and such tools will enable much more ambitious computational experiments.

Statistical Reproducibility In January 2014 Science enacted new manuscript submission requirements: a data-handling plan i.e. how outliers will be dealt with, sample size estimation for effect size, whether samples are treated randomly, whether experimenter blind to the conduct of the experiment. Also added statisticians to the Board of Reviewing Editors.

National Strategic Computing Initiative 2015

NSCI Sec. 2. Objectives. 1. Accelerating delivery of a capable exascale computing system that integrates hardware and software capability to deliver approximately 100 times the performance of current 10 petaflop systems across a range of applications representing government needs. 2. Increasing coherence between the technology base used for modeling and simulation and that used for data analytic computing. 3. Establishing, over the next 15 years, a viable path forward for future HPC systems even after the limits of current semiconductor technology are reached (the "post- Moore's Law era"). 4. Increasing the capacity and capability of an enduring national HPC ecosystem by employing a holistic approach that addresses relevant factors such as networking technology, workflow, downward scaling, foundational algorithms and software, accessibility, and workforce development. 5. Developing an enduring public-private collaboration to ensure that the benefits of the research and development advances are, to the greatest extent, shared between the United States Government and industrial and academic sectors.

From a technical requirements perspective, infrastructure for data- intensive science needs to consider data acquisition, storage and archiving, search and retrieval, analytics, and collaboration (including publish/sub- scribe services). Recent NSF requirements to submit data management plans as part of proposals signal recognition that access to data is increasingly important for interdisciplinary science and for research reproducibility. Although the focus is sometimes on the hardware infrastructure (amount of storage, bandwidth, etc.), the human and software infrastructure is also important. Understanding the software frameworks that are enabled within the various cloud services and then mapping scientific workflows onto them requires a high level of both technical and scientific insight. Moreover, these new services enable a deeper level of collaboration and software reuse that are critical for data-intensive science. changing scientific workflows extend to the human side of scientific computing as well. Especially in regards to data-intensive science, reproducibility will be challenging. These requirements will often be as important as the traditional technical requirements of CPU performance, latency, storage, and bandwidth. deciding how much data to save is a trade-off between the cost of saving and the cost of reproducing, and this is potentially more significant than the trade-off between disks and processors.

Community Infrastructure Research Environments Innovations Verifiable Computational Research SHARE Code Ocean Jupyter knitr Sweave Cyverse NanoHUB Collage Authoring Environment SOLE Open Science Framework Vistrails Workflow Systems Sumatra GenePattern IPOL Popper Galaxy torch.ch Whole Tale flywheel.io Taverna Wings Pegasus CDE binder.org Kurator Kepler Everware Reprozip Dissemination Platforms ResearchCompendia.org DataCenterHub RunMyCode.org ChameleonCloud Occam RCloud TheDataHub.org Madagascar Wavelab Sparselab

Lifecycle of Data Science: An Abstraction the study of data science ethics, documentation and metadata creation, best practices, policy; the science of data science application level experimental design data generation and collection data exploration and hypothesis generation data cleaning and organization feature selection and data preparation model building and statistical inference simulation and cross-validation visualization publication and artifact preservation / archiving infrastructure level notebooks and workflow software database structures workflow software and preregistration tools data management tools notebooks, workflow software; containerization tools notebooks, inference languages notebooks notebooks, visualization software workflow software, artifact linking tools system level hardware, cloud computing infrastructure, systems and system management, data structures, storage

A (Very) Brief History..

Yale 2009 Inspired by the Bermuda Principles, Data and Code Sharing Roundtable on November 21, 2009. See http://stodden.net/roundtablenov212009 We collectively produced the Data and Code Sharing Declaration including a description of the problem, proposed solutions, and dream goals we d like to see.

ICERM 2012

ICERM Workshop Report

Issues from ICERM The need to carefully document the full context of computational experiments including system environment, input data, code used, computed results, etc. The need to save the code and data in a permanent repository, with version control and appropriate meta-data. The need for reviewers, research institutions, and funding agencies to recognize the importance of computing and computing professionals, and to allocate funding for after-the-grant support and repositories. The increasing importance of numerical reproducibility, and the need for tools to ensure and enhance numerical reliability. The need to encourage publication of negative results as other researchers can often learn from them. The re-emergence of the need to ensure responsible reporting of performance.

Supercomputing Efforts by SIGHPC, SIGMOD, SIGCOMM