Abstract
Numerical climate model simulations run at high spatial and temporal resolutions generate massive quantities of data. As our computing capabilities continue to increase, storing all of the data is not sustainable, and thus it is important to develop methods for representing the full datasets by smaller compressed versions. We propose a statistical compression and decompression algorithm based on storing a set of summary statistics as well as a statistical model describing the conditional distribution of the full dataset given the summary statistics. We decompress the data by computing conditional expectations and conditional simulations from the model given the summary statistics. Conditional expectations represent our best estimate of the original data but are subject to oversmoothing in space and time. Conditional simulations introduce realistic small-scale noise so that the decompressed fields are neither too smooth nor too rough compared with the original data. Considerable attention is paid to accurately modeling the original dataset—1 year of daily mean temperature data—particularly with regard to the inherent spatial nonstationarity in global fields, and to determining the statistics to be stored, so that the variation in the original data can be closely captured, while allowing for fast decompression and conditional emulation on modest computers. Supplementary materials for this article are available online.
| Original language | English |
|---|---|
| Pages (from-to) | 56-67 |
| Number of pages | 12 |
| Journal | Journal of the American Statistical Association |
| Volume | 113 |
| Issue number | 521 |
| DOIs | |
| State | Published - Jan 2 2018 |
Keywords
- Gaussian process
- Half-spectral
- Nonstationary
- Spatial-temporal data
- SPDE
Fingerprint
Dive into the research topics of 'Compression and Conditional Emulation of Climate Model Output'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver