Skip to content

DCAN-Labs/processing_audit

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

48 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

processing audit

This audit is used to check on the status of jobs running either the NiBabies (infant-fMRIprep), fMRIprep, or XCP-D pipeline. It takes in a BIDS valid input folder, and an output folder where the pipeline outputs are stored. It searches through the BIDS input folder to find which subject and session to check for, then searches the output folder for if an executive summary was created and if a crash file exists. PLEASE NOTE THAT THE S3 FUNCTIONALITY FOR CHECKING CRASH LOGS IS NOT YET BUILT, IT WILL ONLY CHECK IF THERE IS AN EXECUTIVE SUMMARY. It will produce a csv and html file with each subject/session from the input folder and a column for the overall status, filled with 'success?', 'fail?' and columns for the executive summary and crash log, marking them found, not found, or if no outputs were found. Input and output folders can be either s3 buckets or on Tier 1 storage. If you are using data on the s3, you need to provide the s3 access and secret keys.

Please not that not all of these flags have been fully implemented in this audit version yet, as this was directly forked from the abcd-hcp-pipline_audit.

Usage

usage: run.py [-h] --report_output_dir REPORT_OUTPUT_DIR
              [--participant_label PARTICIPANT_LABEL [PARTICIPANT_LABEL ...]]
              [--n_cpus N_CPUS] [--s3_access_key S3_ACCESS_KEY]
              [--s3_hostname S3_HOSTNAME] [--s3_secret_key S3_SECRET_KEY]
              [--skip_bids_validator]
              [--session_label SESSION_LABEL [SESSION_LABEL ...]] [-v]
              bids_dir output_dir {participant,group}

abcd-hcp-pipeline_audit entrypoint script.

positional arguments:
  bids_dir              The directory with the input dataset formatted
                        according to the BIDS standard. In the case that the
                        BIDS dataset is within s3 provide the path to the
                        folder along with
                        "s3://BUCKET_NAME/path_to_BIDS_folder".
  output_dir            The directory where the output files are stored. If
                        you are running group level analysis this folder
                        should be prepopulated with the results of
                        theparticipant level analysis.In the case that this
                        folderis within s3 provide the path to the folder
                        along with
                        "s3://BUCKET_NAME/path_to_derivatives_folder".
  {participant,group}   Level of the analysis that will be performed. Unless
                        checking on status of one participants processing, use
                        "group".

optional arguments:
  -h, --help            show this help message and exit
  --report_output_dir REPORT_OUTPUT_DIR, --report-output-dir REPORT_OUTPUT_DIR
                        The directory where the CSV and HTML files will be
                        outputted once the report finishes.
  --participant_label PARTICIPANT_LABEL [PARTICIPANT_LABEL ...], --participant-label PARTICIPANT_LABEL [PARTICIPANT_LABEL ...]
                        The label(s) of the participant(s) that should be
                        analyzed. The label corresponds to
                        sub-<participant_label> from the BIDS spec (so it does
                        not include "sub-"). If this parameter is not provided
                        all subjects should be analyzed. Multiple participants
                        can be specified with a space separated list.
  --n_cpus N_CPUS       Number of CPUs to use for parallel download.
  --s3_access_key S3_ACCESS_KEY
                        Your S3 access key, if data is within S3. If using
                        MSI, this can be found at:
                        https://www.msi.umn.edu/content/s3-credentials
  --s3_hostname S3_HOSTNAME
                        URL for S3 storage hostname, if data is within S3
                        bucket. Defaults to s3.msi.umn.edu for MSIs tier 2
                        CEPH storage.
  --s3_secret_key S3_SECRET_KEY
                        Your S3 secret key. If using MSI, this can be found
                        at: https://www.msi.umn.edu/content/s3-credentials
  --skip_bids_validator
                        Whether or not to perform BIDS dataset validation
  --session_label SESSION_LABEL [SESSION_LABEL ...]
                        The label(s) of the session(s) that should be
                        analyzed. The label corresponds to ses-<session_label>
                        from the BIDS spec (so it does not include "sub-"). If
                        this parameter is not provided all subjects should be
                        analyzed. Multiple participants can be specified with
                        a space separated list.
  -p, --pipeline 
                        Which processing pipeline you're wanting to audit. 
                        Currently supports nibabies, fmriprep, and xcpd. 
                        Please enter the pipelines in all lowercase.

About

Processing audit designed to work with NiBabies and fMRIprep

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages

  • Python 70.5%
  • JavaScript 27.9%
  • CSS 1.3%
  • Shell 0.3%