GVAL is an interoperable, scalable, and efficient Python framework to validate gridded datasets for a variety of applications. GVAL compares modeled output maps with either observation or alternative modeled output maps, which can be of raster or vector formats, producing agreement maps and metrics. Maps undergo necessary homogenization to ensure spatial, data format, and numerical alignment prior to comparison. Comparisons are handled for two-class categorical, multi-class categorical, continuous, and probabilistic statistical data types. Libraries of standard metrics for each statistical data type are included with the ability to register custom, user-defined statistical metrics. Functionality also includes the visualization of agreement maps, subsampling for regions of interest or for use in masking, catalog comparisons including cloud-native catalogs, and attribute tracking methods to facilitate readable metadata, statistical analysis, and hypothesis testing. Leveraging the Pangeo stack, GVAL takes advantage of libraries which provide options for both local serial and parallel processing as well as distributed processing capabilities. The GVAL package is portable, open-source, and supports a variety of analysis domains with geospatial modeling output.
This work will demonstrate the utility of GVAL showing examples of evaluations including several statistical data types and geospatial variables of interest including fluvial inundation extent, storm surge, and total precipitation. We will also demonstrate GVAL’s scalability in workflows by batch processing evaluations of cloud datasets via cataloging as well as the interpretability and reusability of its evaluation workflows via attribute tracking with the intention to adhere to the Findable Accessible Interoperable Reusable (FAIR) data standards and best practices. Each example will also include subsampling the dataset with regions of interest representing vulnerable infrastructure, political boundaries, and wildlife refuges respectively. Through these examples we will articulate how the advancement of geospatial evaluation workflows provide detailed insight into the model performance of gridded datasets in cross-cutting contexts.

