TY - JOUR
T1 - Multistream STGAN
T2 - A Spatiotemporal Image Fusion Model With Improved Temporal Transferability
AU - Lyu, Fangzheng
AU - Yang, Zijun
AU - Diao, Chunyuan
AU - Wang, Shaowen
N1 - Our computational work used ROGER, which is a geospatial supercomputer supported by the CyberGIS Center for Advanced Digital and Spatial Studies and the School of Earth, Society, and Environment at University of Illinois Urbana-Champaign.
This work is supported in part by the National Science Foundation (NSF) under grant numbers: 1833225, 2112356, 2118329, and 1951657, in part by the National Aeronautics and Space Administration (NASA) under grant number: 80NSSC21K0946, and in part by the United States Department of Agriculture (USDA) under grant number: 2021-67021-33446. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the NSF, NASA, and USDA. Corresponding author: Chunyuan Diao and Shaowen Wang. Fangzheng Lyu, and Zijun Yang contributed equally and are co-first authors.
PY - 2025
Y1 - 2025
N2 - Spatiotemporal satellite image fusion aims to generate remote sensing images satisfying both high spatial and temporal resolution by integrating different satellite imagery datasets with distinct spatial and temporal resolutions. Such fusion technique is crucial for numerous applications that require frequent monitoring at fine spatial and temporal scales spanning agriculture, environment, natural resources, and disaster management. However, existing fusion models have difficulty accommodating abrupt spatial changes in land cover among images and dealing with temporally distant image data. This article proposes a novel multistream spatiotemporal fusion generative adversarial network (STGAN) model for spatiotemporal satellite image fusion that can produce accurate fused images and accommodate substantial temporal differences between the input images. The STGAN employs a conditional generative adversarial network architecture with a multistream input design to better learn temporal features. The generator of STGAN comprises convolutional blocks, a spatial transformer module, a channel attention network, and a U-net module designed to better capture spatial and temporal features from the multiresolution input images. Comprehensive evaluations of the proposed STGAN model have been performed on the Coleambally Irrigation Area and Lower Gwydir Catchment datasets, using both visual inspection and spatial and spectral metrics, including root mean square error, relative dimensionless global error synthesis, spectral angle mapping, structural similarity index measure, and local binary pattern. The experiments show that the proposed STGAN model consistently outperforms existing benchmark models and is capable of generating high-quality fused remote sensing data product of high spatial and temporal resolution.
AB - Spatiotemporal satellite image fusion aims to generate remote sensing images satisfying both high spatial and temporal resolution by integrating different satellite imagery datasets with distinct spatial and temporal resolutions. Such fusion technique is crucial for numerous applications that require frequent monitoring at fine spatial and temporal scales spanning agriculture, environment, natural resources, and disaster management. However, existing fusion models have difficulty accommodating abrupt spatial changes in land cover among images and dealing with temporally distant image data. This article proposes a novel multistream spatiotemporal fusion generative adversarial network (STGAN) model for spatiotemporal satellite image fusion that can produce accurate fused images and accommodate substantial temporal differences between the input images. The STGAN employs a conditional generative adversarial network architecture with a multistream input design to better learn temporal features. The generator of STGAN comprises convolutional blocks, a spatial transformer module, a channel attention network, and a U-net module designed to better capture spatial and temporal features from the multiresolution input images. Comprehensive evaluations of the proposed STGAN model have been performed on the Coleambally Irrigation Area and Lower Gwydir Catchment datasets, using both visual inspection and spatial and spectral metrics, including root mean square error, relative dimensionless global error synthesis, spectral angle mapping, structural similarity index measure, and local binary pattern. The experiments show that the proposed STGAN model consistently outperforms existing benchmark models and is capable of generating high-quality fused remote sensing data product of high spatial and temporal resolution.
KW - Deep Learning
KW - GAN
KW - Remote Sensing
KW - Spatiotemporal Image Fusion
KW - Deep learning
KW - generative adversarial network (GAN)
KW - spatiotemporal image fusion
KW - remote sensing
UR - http://www.scopus.com/inward/record.url?scp=85210939735&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85210939735&partnerID=8YFLogxK
U2 - 10.1109/JSTARS.2024.3506879
DO - 10.1109/JSTARS.2024.3506879
M3 - Article
AN - SCOPUS:85210939735
SN - 1939-1404
VL - 18
SP - 1562
EP - 1576
JO - IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
JF - IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
ER -