dbGaPCheckup

1 Overview

I want to make the data sing, but it is torturing me instead. Will the real data please stand up?

The goal of dbGaPCheckup is to make your National Library of Medicine database of Genotypes and Phenotypes (dbGaP) data set submission a tiny bit easier. Specifically, our package implements several check, awareness, utility, and reporting functions designed to help you ensure that your Subject Phenotype data set and data dictionary meet a variety of dbGaP specific formatting requirements. A list of the functions available can be found below.

The software announcement for our package has been published in BMC Bioinformatics and is available at https://bmcbioinformatics.biomedcentral.com/articles/10.1186/s12859-023-05200-8.

Heinsberg, L.W., Weeks, D.E. dbGaPCheckup: pre-submission checks of dbGaP-formatted subject phenotype files. BMC Bioinformatics 24, 77 (2023). https://doi.org/10.1186/s12859-023-05200-8

Function_Name Function_Type Function_Description
field_check check Checks for dbGaP required fields: variable name (VARNAME), variable description (VARDESC), units (UNITS), and variable value and meaning (VALUES).
pkg_field_check check Checks for package-level required fields: variable type (TYPE), minimum value (MIN), and maximum value (MAX).
dimension_check check Checks that the number of variables match between the data set and data dictionary.
name_check check Checks that variable names match between the data set and data dictionary.
id_check check Checks that the first column of the data set is the primary ID for each participant labeled as SUBJECT_ID, that values contain no illegal characters or padded zeros, and that each participant has an ID.
row_check check Checks for empty or duplicate rows in the data set and data dictionary.
NA_check check Checks for NA values in the data set and, if NA values are present, also checks for an encoded NA value=meaning description.
type_check check If a TYPE field exists, this function checks for any TYPE entries that aren’t allowable per dbGaP instructions.
values_check check Checks for potential errors in the VALUES columns by ensuring (1) required format of VALUE=MEANING (e.g., 0=No or 1=Yes); (2) no leading/trailing spaces near the equals sign (e.g., 0=No vs. 0 = No); (3) all variables of TYPE encoded have VALUES entries; and (4) all variables with VALUES entries are listed as TYPE encoded.
integer_check check Checks for variables that appear to be incorrectly listed as TYPE integer.
decimal_check check Checks for variables that appear to be incorrectly listed as TYPE decimal.
misc_format_check check Checks miscellaneous dbGaP formatting requirements to ensure (1) no empty variable names; (2) no duplicate variable names; (3) variable names do not contain “dbgap”; (4) there are no duplicate column names in the dictionary; and (5) column names falling after VALUES column are unnamed.
description_check check Checks for unique and non-missing descriptions (VARDESC) for every variable in the data dictionary.
minmax_check check Checks for variables that have values exceeding the listed MIN or MAX.
missing_value_check check Checks for variables that have non-encoded missing value codes.
complete_check bulk check Runs the entire workflow (field_check, pkg_field_check, dimension_check, name_check, id_check, row_check, NA_check, type_check, values_check, integer_check, decimal_check, misc_format_check, description_check, minmax_check, and missing_value_check).
add_missing_fields utility Adds additional fields required by this package including variable type (‘TYPE’), minimum value (‘MIN’), and maximum value (‘MAX’).
name_correct utility Updates the data set so variable names match those listed in the data dictionary.
reorder_dictionary utility Reorders the data dictionary to match the data set.
reorder_data utility Reorders the data set to match the data dictionary.
id_first_data utility Reorders the data set so that SUBJECT_ID comes first.
id_first_dict utility Reorders the data dictionary so that SUBJECT_ID comes first.
label_data utility, awareness Adds non-missing information from the data dictionary as attributes to the data.
value_meaning_table utility, awareness Generates a value-meaning table by parsing the VALUES fields.
missingness_summary awareness Summarizes the amount of missingness in the data set.
value_missing_table awareness Checks for consistent usage of encoded values and missing value codes between the data dictionary and the data set.
dictionary_search awareness Facilitates searches of the data dictionary.
check_report bulk check, reporting Generates a user-readable report of the checks run by the complete_check function.
create_report reporting, awareness Generates a textual and graphical report of the selected variables in HTML format.
create_awareness_report reporting, awareness Generates an awareness report, calling missingness_summary and value_missing_table functions.

List of function names and types.

2 Copyright information

Copyright 2022, University of Pittsburgh. All Rights Reserved. License: GPL-2

3 Installation

You can install dbGaPCheckup from CRAN using:

install.packages("dbGaPCheckup")

You can install the development version of dbGaPCheckup from GitHub using:

# install.packages("devtools")
devtools::install_github("lwheinsberg/dbGaPCheckup")

4 Example

An introductory example is provided below. For more details see https://lwheinsberg.github.io/dbGaPCheckup/ or the dbGaPCheckup_vignette (expanded instructions) or dbGaPCheckup (Quick Start) vignette documents.

A special note: As you will see below, this package requires several fields beyond those required by the dbGaP formatting requirements. Specifically, while dbGaP requires that only the VARNAME, VARDESC, UNITS, and VALUES columns be present in the data dictionary, this package requires that MIN, MAX, and TYPE fields are also included. If your data dictionary does not include these fields already, you can use the add_missing_fields function to auto fill them (see below).

library(dbGaPCheckup)

Read in your Subject Phenotype data into DS.data.

DS.path <- system.file("extdata", "DS_Example.txt",
   package = "dbGaPCheckup", mustWork=TRUE)
DS.data <- read.table(DS.path, header=TRUE, sep="\t", quote="", as.is = TRUE)

Read in your Subject Phenotype data dictionary into DD.dict.

DD.path <- system.file("extdata", "DD_Example2f.xlsx",
   package = "dbGaPCheckup", mustWork=TRUE)
DD.dict <- readxl::read_xlsx(DD.path)
#> New names:
#> • `` -> `...15`
#> • `` -> `...16`
#> • `` -> `...17`
#> • `` -> `...18`
#> • `` -> `...19`

Run the function check_report. Note that, for many functions, specification of missing value codes are important for accurate results.

report <- check_report(DD.dict = DD.dict, DS.data = DS.data, non.NA.missing.codes=c(-4444, -9999))
#> # A tibble: 15 × 3
#>    Function            Status        Message                                    
#>    <chr>               <chr>         <chr>                                      
#>  1 field_check         Passed        Passed: required fields VARNAME, VARDESC, …
#>  2 pkg_field_check     Failed        ERROR: not all package-level required fiel…
#>  3 dimension_check     Passed        Passed: the variable count matches between…
#>  4 name_check          Passed        Passed: the variable names match between t…
#>  5 id_check            Passed        Passed: All ID variable checks passed.     
#>  6 row_check           Passed        Passed: no blank or duplicate rows detecte…
#>  7 NA_check            Not attempted ERROR: Required pre-check pkg_field_check …
#>  8 type_check          Failed        ERROR: TYPE column not found. Consider usi…
#>  9 values_check        Not attempted ERROR: Required pre-check type_check faile…
#> 10 integer_check       Not attempted ERROR: Required pre-check pkg_field_check …
#> 11 decimal_check       Not attempted ERROR: Required pre-check pkg_field_check …
#> 12 misc_format_check   Passed        Passed: no check-specific formatting issue…
#> 13 description_check   Failed        ERROR: missing and duplicate descriptions …
#> 14 minmax_check        Not attempted ERROR: Required pre-check pkg_field_check …
#> 15 missing_value_check Not attempted ERROR: Required pre-check pkg_field_check …
#> --------------------
#> pkg_field_check: Failed 
#> ERROR: not all package-level required fields are present in the data dictionary. Consider using the add_missing_fields function to auto fill these fields. 
#> $pkg_field_check.Info
#>  TYPE   MIN   MAX 
#> FALSE FALSE FALSE 
#> 
#> --------------------
#> type_check: Failed 
#> ERROR: TYPE column not found. Consider using the add_missing_fields function to autofill TYPE. 
#> $type_check.Info
#> [1] "ERROR: TYPE column not found."
#> 
#> --------------------
#> description_check: Failed 
#> ERROR: missing and duplicate descriptions found in data dictionary. 
#> $description_check.Info
#> # A tibble: 4 × 2
#>   VARNAME  VARDESC              
#>   <chr>    <chr>                
#> 1 PREGNANT <NA>                 
#> 2 REACT    <NA>                 
#> 3 HEIGHT   Height of participant
#> 4 WEIGHT   Height of participant
#> 
#> --------------------

As described in more detail in the dbGaPCheckup_vignette vignette, some checks contain embedded “pre-checks” that must be passed before the check can be run. For example, as mentioned above, this package requires MIN, MAX, and TYPE fields in the data dictionary. We can see above that several of the checks (e.g., NA_check, integer_check, decimal_check) were not run because these additional required fields were missing. Never fear though! We have created a function to auto fill these fields that can be used to get further along in the checks!

DD.dict.updated <- add_missing_fields(DD.dict, DS.data)
#> $Message
#> [1] "CORRECTED ERROR: not all package-level required fields were present in the data dictionary. The missing fields have now been added! TYPE was inferred from the data, and MIN/MAX have been added as empty fields."
#> 
#> $Missing
#> [1] "TYPE" "MIN"  "MAX"

Once the fields are added, you can simply return to run your checks. Don’t forget to use the updated dictionary though!

report.v2 <- check_report(DD.dict = DD.dict.updated , DS.data = DS.data, non.NA.missing.codes=c(-4444, -9999))
#> # A tibble: 15 × 3
#>    Function            Status Message                                           
#>    <chr>               <chr>  <chr>                                             
#>  1 field_check         Passed Passed: required fields VARNAME, VARDESC, UNITS, …
#>  2 pkg_field_check     Passed Passed: package-level required fields TYPE, MIN, …
#>  3 dimension_check     Passed Passed: the variable count matches between the da…
#>  4 name_check          Passed Passed: the variable names match between the data…
#>  5 id_check            Passed Passed: All ID variable checks passed.            
#>  6 row_check           Passed Passed: no blank or duplicate rows detected in da…
#>  7 NA_check            Passed Passed: no NA values detected in data set.        
#>  8 type_check          Passed Passed: All TYPE entries found are accepted by db…
#>  9 values_check        Passed Passed: all four VALUES checks look good.         
#> 10 integer_check       Passed Passed: all variables listed as TYPE integer appe…
#> 11 decimal_check       Passed Passed: all variables listed as TYPE decimal appe…
#> 12 misc_format_check   Passed Passed: no check-specific formatting issues ident…
#> 13 description_check   Failed ERROR: missing and duplicate descriptions found i…
#> 14 minmax_check        Passed Passed: when provided, all variables are within t…
#> 15 missing_value_check Failed ERROR: some variables have non-encoded missing va…
#> --------------------
#> description_check: Failed 
#> ERROR: missing and duplicate descriptions found in data dictionary. 
#> $description_check.Info
#> # A tibble: 4 × 2
#>   VARNAME  VARDESC              
#>   <chr>    <chr>                
#> 1 PREGNANT <NA>                 
#> 2 REACT    <NA>                 
#> 3 HEIGHT   Height of participant
#> 4 WEIGHT   Height of participant
#> 
#> --------------------
#> missing_value_check: Failed 
#> ERROR: some variables have non-encoded missing value codes. 
#> $missing_value_check.Info
#>     VARNAME VALUE MEANING  PASS
#> 16 CUFFSIZE -9999    <NA> FALSE
#> 
#> --------------------

Now we see that 13 out of 15 checks pass, but the workflow fails at description_check and missing_value_check. Specifically, in description_check we see that variables PREGNANT and REACT were identified as having missing variable descriptions (VARDESC), and variables HEIGHT and WEIGHT incorrectly have identical descriptions. In missing_value_check, we see that the variable CUFFSIZE contains a -9999 encoded value that is not specified in a VALUES column. While we have included functions that support “simple fixes”, the issues identified here would need to be corrected manually in your data dictionary before moving on.

Note that our package also includes some helpful reporting functions (create_report and create_awareness_report) that can help you more fully interrogate your data and catch potential errors prior to dbGaP data submission.

Finally, after your data dictionary is fully consistent with your data, you can use the label_data function to convert your data to labelled data, essentially embedding the data dictionary into the data.

For more information about these reporting functions, and all of the checks available in our package, see the Quick Start (dbGaPCheckup) or expanded (dbGaPCheckup_vignette) vignettes for more information!

5 Contact information

If you have any questions or comments, please feel free to contact us!

Lacey W. Heinsberg, PhD, RN: law145@pitt.edu
Daniel E. Weeks, PhD: weeks@pitt.edu

Bug reports: https://github.com/lwheinsberg/dbGaPCheckup/issues

6 Acknowledgments

This package was developed with partial support from the National Institutes of Health under award numbers R01HL093093, R01HL133040, and K99HD107030. The eval_function and dat_function functions that form the backbone of the awareness reports were inspired by an elegant 2016 homework answer submitted by Tanbin Rahman in our HUGEN 2070 course ‘Bioinformatics for Human Genetics’. We would also like to thank Nick Moshgat for testing and providing feedback on our package during development.