Have a personal or library account? Click to login
Recommendations for Discipline-Specific FAIRness Evaluation Derived from Applying an Ensemble of Evaluation Tools Cover

Recommendations for Discipline-Specific FAIRness Evaluation Derived from Applying an Ensemble of Evaluation Tools

Open Access
|Mar 2022

Figures & Tables

Table 1

Summary of the FAIRness evaluation tools which we assessed but decided not to use in the context of this study. The evaluation approaches were assessed in April 2021; a reassessment took place for some tools in February 2022 (see references).

TOOLNOT USED BECAUSEREFERENCE
ANDS-Nectar-RDS FAIR data self-assessment toolnot accessibleANDS (2021)
DANS-Fairdatpilot version meant for internal testing at DANSThomas (2017)
SATIFYDnot maintained anymore (L. Cepinskas (DANS), pers. comm. 24 March 21)Fankhauser et al. (2019)
The CSIRO 5-star Data Rating toolnot accessible as online toolYu & Cox (2017)
The Scientific Data Stewardship Maturity Assessment Modelnon-automated capture of evaluation results; proprietary document formatPeng et al. (2015)
Data Stewardship Wizardassistance for FAIR data management planning, not for evaluation of archived dataPergl et al. (2019)
RDA-SHARC Evaluationno fillable form readily providedDavid et al. (2018)
WMO Stewardship Maturity Matrix for Climate Data (SMM-CD)non-automated capture of evaluation results; proprietary document formatPeng et al. (2020)
Data Use and Services Maturity Matrixunclear application conceptThe MM-Serv Working Group (2018)
ARDC FAIR Self-Assessment Tooltest results not saveable; no quantitative FAIR measureSchweitzer et al. (2021)
Table 2

Summary of the five FAIRness evaluation tools used in this study. The hybrid method of FAIRshake combines automated and manual evaluation. The covered FAIR ((F)indable, (A)ccessible, (I)nteroperable, (R)eusable) dimensions refer to the number of metrics the tool tests, such as FMES checks for Findability using 8 different tests.

TOOLACRONYMMETHODCOVERED FAIR DIMENSIONSREFERENCE
Checklist for Evaluation of Dataset Fitness for UseCFUmanualn/aAustin et al. (2019)
FAIR Maturity Evaluation ServiceFMESautomatedF: 8, A: 5, I: 7, R: 2Wilkinson et al. (2019)
FAIRshaken/ahybridF: 3, A: 1, I: 0, R: 5Clarke et al. (2019)
F-UJIn/aautomatedF: 7, A: 3, I: 4, R: 10Devaraju et al. (2021)
Self Assessmentn/amanualF: 13, A: 12, I: 10, R: 10Bahim et al. (2020)
Table 3

WDCC projects selected for evaluation. The project acronyms can be directly used to search and find the evaluated projects using the WDCC GUI. The project volume in TB (third column) refers to the total volume of the entire project named in the first column. See Peters-von Gehlen, & Höck (2021) for details of evaluated resources.

PROJECT ACRONYMDATA SUMMARYPROJECT VOLUME [TB]DOI ASSIGNEDCREATION DATECOMMENTS
IPCC-AR5_CMIP5Coupled Climate Model Output, prepared following CMIP5 guidelines and basis of the IPCC 5th Assessment Report (2 AICs evaluated)1655yes and no2012-05-31 and 2011-10-10
CliSAPObservational data products from satellite remote sensing (2 AICs evaluated)163yes and no2015-09-15 and 2009-11-12one collection with no data access
WASCALDynamically downscaled climate data for West Africa73yes2017-02-23
CMIP6_RCM_forcing_MPI-ESM1-2Coupled Climate Model output prepared as boundary conditions for regional climate models, prepared following CMIP6 experiment guidelines51yes2020-02-27
MILLENNIUM_COSMOSCoupled Climate Model of ensemble simulations covering the last millennium (800-2000AD)47no2009-05-12
IPCC_TAR_ECHAM4/OPYCCoupled Climate Model Output, prepared to support the IPCCs 3rd Assessment Report2.6yes2003-01-26Experiment and dataset with DOI; First ever DOI assigned to data (Stendel et al. 2004)
Storm_Tide_1906_German_BightNumerical simulation of the 1906 storm tide in the German Bight0.3yes2020-10-27
COPSObservational data obtained from radar remote sensing during the COPS (Convective and Orographically-Induced Precipitation Study) campaign0.2yes2008-01-28
HDCP2-OBSObservations collected during the HDCP2 (High Definition Clouds and Precipitation for Climate Prediction) project0.06yes2018-09-18
OceanRAINIn-situ, along-track shipboard observations of routinely measured atmospheric and oceanic state parameters over global oceans0.01yes2017-12-13 7
CARIBICObservations of atmospheric parameters obtained from commercial aircraft equipped with an instrumentation container7.7E-5no2002-04-27
Table 4

Results of FAIR assessments of WDCC data holdings using the ensemble of FAIRness evaluation tools detailed in Section 2.1. The scores per test are calculated as unweighted mean over all tested FAIR maturity indicators. The mean (∅), standard deviation (σ) and relative standard deviation (σ) on a project basis (three rightmost columns) are calculated across the scores of the five FAIR assessment tools. The mean value representative for the WDCC (∅ (WDCC), last row) is calculated for all values in the respective column of the table. See main text for more details. Results at finer granularity are provided in the supporting data (Peters-von Gehlen et al., 2021).

PROJECT ACRONYMSELF-ASSESSMENTCFUFMESF-UJIFAIRSHAKE∅ PER PROJECTσ PER PROJECTσ PER PROJECT
IPCC-AR5_CMIP50.840.720.440.580.950.710.200.29
IPCC-AR5_CMIP5, no DOI0.650.670.440.540.930.650.190.29
CliSAP0.860.780.480.580.970.730.200.28
CliSAP, no data accessible0.270.300.430.520.640.430.150.36
WASCAL0.900.800.500.580.910.740.180.25
CMIP6_RCM_forcing_MPI-ESM1-20.860.850.570.620.920.760.160.21
MILLENNIUM_COSMOS0.630.530.450.510.820.590.140.24
IPCC_TAR_ECHAM4/OPYC0.820.630.500.640.890.700.160.23
Storm_Tide_1906_German_Bight0.900.680.550.620.830.710.150.21
COPS0.860.470.530.550.870.660.190.29
HDCP2-OBS0.900.480.530.590.860.670.190.29
OceanRAIN0.900.750.570.600.970.760.180.23
CARIBIC0.620.700.500.540.820.640.130.20
∅(WDCC)0.770.640.500.580.880.670.150.22
Table 5

Cross-correlations between the scores per project obtained with the five FAIRness evaluation tools (Table 4).

SELF-ASSESSMENTCFUFMESF-UJIFAIRSHAKE
Self-Assessmentn/a0.610.650.730.79
CFUn/a0.360.500.78
FMESn/a0.650.30
F-UJIn/a0.49
FAIRshaken/a
Table 6

Summary of the experiences gained from applying the ensemble of different FAIRness evaluation approaches in this study.

AUTOMATEDMANUALHYBRID
applied toolsFMES (Wilkinson et al., 2019)
F-UJI (Devaraju & Huber, 2020)
CFU
self-assessment (Bahim et al., 2020)
FAIRshake (Clarke et al., 2019)
application/use of the toolthe tools take PID/DOI of the resource to be evaluated
if available, selection of appropriate metric sets is critical and requires prior review
completing questionnaires is time intensive and depends on the extent of metrics
expert knowledge is essential
the tools take PID/DOI of the resource to be evaluated
selection of appropriate metric sets is critical and requires prior review
expert knowledge required to evaluate contextual reusability time intensive
preservation of resultsresults are saved in an online database or are exported (printed) as PDF
local installations store results locally
date of the evaluation has to be manually noted (in the tools evaluated here)
results are saved locally as spreadsheets
date of the evaluation has to be manually noted
results are saved in an online database
date of the evaluation has to be manually noted (using the tool evaluated here)
interpretation of resultsdetailed information on the applied metrics is available as documentation
if tests fail, the tools provide technical output interpretable by experts results are provided as quantitative measure
the form is filled by a knowledgeable expert, interpretation is thus performed during the evaluation itself
quantification of results depends on evaluator perception
detailed information on the applied automated metrics is available as documentation
manual parts filled by a knowledgeable expert, interpretation is thus performed during the evaluation itself
quantification of results partly depends on evaluator perception
reproducibilityresults are reproducible as long as the same code version is usedhuman evaluation is subjective, reproducibility depends on manual documentation of each evaluationreproducibility of atomated parts is given as long as the same code version is used
human evaluation is subjective, reproducibility depends on manual documentation of each evaluation
evaluation of technical reusability/machine actionabilitygood
tests fail if code specifications are not exactly met
limited
machine actionability cannot be specifically tested
assessment only based on implemented methods/protocols, not their functionality
very good
failed automated tests can be manually amended given that an implementation is present but does not exactly match the test implementation
evaluation of con-textual reusabilitylimited
domain-specific and agreed standardised FAIR metrics are needed
good to excellent
depends on the domain-expertise of the evaluator and the time and effort put into the evaluation
good to excellent
depends on the domain-expertise of the evaluator and the time and effort put into the evaluation
ACRONYMDEFINITION
AICArchival Information Collection
AIPArchival Information Package
AIUArchival Information Unit
ANDSAustralian National Data Service
AR55th Assessment Report
ARDCAustralian Research Data Commons
CFUChecklist for Evaluation of Dataset Fitness for Use
CliSAPIntegrated Climate System Analysis and Prediction
CMIP5/6Coupled Model Intercomparison Project 5/6
COPSConvective and Orographically Induced Precipitation Study
CORDEXCoordinated Regional Downscaling Experiment
CSIROCommonwealth Scientific and Industrial Research Organisation
DANSData Archiving and Networked Services
DKRZGerman Climate Computing Center
DOIDigital Object Identifier
DSJData Science Journal
FMESFAIR Maturity Evaluation Service
GUIGraphical User Interface
HDCP2High Definition Clouds and Precipitation for Climate Prediction
IPCCIntergovernmental Panel on Climate Change
JSON-LDJavaScript Object Notation for Linked Data
NetCDFNetwork Common Data Form
OAISOpen Archival Information System
ORCiDOpen Researcher and Contributor Identifier
PBPetabyte
PIDPersistent Identifier
RCMRegional Climate Model
RDAResearch Data Alliance
URLUniform Resource Locator
WASCALWest African Science Service Centre on Climate Change and Adapted Land Use
WDCCWorld Data Center for Climate
WDSWorld Data System
WGWorking Group
WMOWorld Meteorological Organization
Language: English
Submitted on: Sep 3, 2021
|
Accepted on: Feb 11, 2022
|
Published on: Mar 24, 2022
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2022 Karsten Peters-von Gehlen, Heinke Höck, Andrej Fast, Daniel Heydebreck, Andrea Lammert, Hannes Thiemann, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.