Reproducible Neuroimaging: Principles and Actions

Reproducible Neuroimaging: Principles and Actions

ReproNim’s principles of reproducible neuroimaging

  1. Study planning
    1. Implement good science basics (power analysis, statistical consults, etc.).
    1. Use pre-existing data for planning and/or analysis.
    1. Create an NIH-compliant data management and sharing plan.
    1. Adopt open consent to allow broad sharing of data.
    1. Pre-register your study.
  1. Data and metadata management
    1. Use standard data formats and extend them to meet your needs.
    1. Use version control from start to finish.
    1. Annotate data using standard, reproducible procedures.
  1. Software management
    1. Use released versions of open source software.
    1. Use version control from start to finish.
    1. Automate the installation of your code and its dependencies.
    1. Automate the execution of your data analysis.
    1. Annotate your code and workflows using standard, reproducible procedures.
    1. Use containers where reasonable.
  1. Publishing everything (publishing re-executable publications)
    1. Share plans (pre-registration).
    1. Share software.
    1. Share data.
    1. Make all research objects FAIR.

ReproNim’s four core actions

As indicated by the blue highlights in the figure below, four core actions are key to implementing the above principles.

  1. Use of standards

    Using standard data formats and extending them to meet specific research needs is important for data and metadata management (Principle 2) in reproducible neuroimaging.

  2. Annotation and provenance

    Annotating data using standard, reproducible procedures ensures clarity and transparency in data management (Principle 2). Provenance refers to the origin and history of data and processes, enabling researchers to track how data was generated, modified, and analyzed (Principles 2, 3, and 4). This is essential for understanding the context of data and ensuring reproducibility.

  3. Implementation of version control

    Version control is crucial for both data and software management. It allows researchers to track changes over time, revert to previous versions if necessary, and collaborate effectively.

    For data, version control helps manage different versions of datasets and track modifications made during processing and analysis (Principle 2).

    For software, version control helps track code changes, manage different versions of analysis scripts, and ensure that the correct version of the code is used for each analysis (Principle 3).

    And even publications can be versioned (Principle 4).

  4. Use of containers

    Containers provide a portable and self-contained environment for running software, ensuring that the analysis can be executed consistently across different computing environments (Principle 3). Containers encapsulate all of the software dependencies needed to run an analysis, making it easier to share software (Principle 4) and reproduce results.

A diagram showing how ReproNim’s principles and core actions work together to reproducible neuroimaging.