Browse > Article
http://dx.doi.org/10.7780/kjrs.2020.36.5.1.13

Evaluation of Spatio-temporal Fusion Models of Multi-sensor High-resolution Satellite Images for Crop Monitoring: An Experiment on the Fusion of Sentinel-2 and RapidEye Images  

Park, Soyeon (Department of Geoinformatic Engineering, Inha University)
Kim, Yeseul (Department of Geoinformatic Engineering, Inha University)
Na, Sang-Il (National Institute of Agricultural Sciences, Rural Development Administration)
Park, No-Wook (Department of Geoinformatic Engineering, Inha University)
Publication Information
Korean Journal of Remote Sensing / v.36, no.5_1, 2020 , pp. 807-821 More about this Journal
Abstract
The objective of this study is to evaluate the applicability of representative spatio-temporal fusion models developed for the fusion of mid- and low-resolution satellite images in order to construct a set of time-series high-resolution images for crop monitoring. Particularly, the effects of the characteristics of input image pairs on the prediction performance are investigated by considering the principle of spatio-temporal fusion. An experiment on the fusion of multi-temporal Sentinel-2 and RapidEye images in agricultural fields was conducted to evaluate the prediction performance. Three representative fusion models, including Spatial and Temporal Adaptive Reflectance Fusion Model (STARFM), SParse-representation-based SpatioTemporal reflectance Fusion Model (SPSTFM), and Flexible Spatiotemporal DAta Fusion (FSDAF), were applied to this comparative experiment. The three spatio-temporal fusion models exhibited different prediction performance in terms of prediction errors and spatial similarity. However, regardless of the model types, the correlation between coarse resolution images acquired on the pair dates and the prediction date was more significant than the difference between the pair dates and the prediction date to improve the prediction performance. In addition, using vegetation index as input for spatio-temporal fusion showed better prediction performance by alleviating error propagation problems, compared with using fused reflectance values in the calculation of vegetation index. These experimental results can be used as basic information for both the selection of optimal image pairs and input types, and the development of an advanced model in spatio-temporal fusion for crop monitoring.
Keywords
Spatio-temporal fusion; Crop monitoring; Multi-sensor image; Resolution;
Citations & Related Records
Times Cited By KSCI : 4  (Citation Analysis)
연도 인용수 순위
1 Chen, Y., R. Cao, J. Chen, X. Zhu, J. Zhou, G. Wang, M. Shen, X. Chen, and W. Yang, 2020. A new cross-fusion method to automatically determine the optimal input image pairs for NDVI spatiotemporal data fusion, IEEE Transactions on Geoscience and Remote Sensing, 58(7): 5179-5194.   DOI
2 Gao, F., J. Masek, M. Schwaller, and F. Hall, 2006. On the blending of the Landsat and MODIS surface reflectance: Predicting daily Landsat surface reflectance, IEEE Transactions on Geoscience and Remote Sensing, 44(8): 2207-2218.   DOI
3 Gevaert, C.M. and F.J. García-Haro, 2015. A comparison of STARFM and an unmixing-based algorithm for Landsat and MODIS data fusion, Remote Sensing of Environment, 156: 34-44.   DOI
4 Huang, B. and H. Song, 2012. Spatiotemporal reflectance fusion via sparse representation, IEEE Transactions on Geoscience and Remote Sensing, 50(10): 3707-3716.   DOI
5 Jarihani, A.A., T.R. McVicar, T.G. Van Niel, I.V. Emelyanova, J.N. Callow, and K. Johansen, 2014. Blending Landsat and MODIS data to generate multispectral indices: A comparison of "index-then-blend" and "blend-then-index" approaches, Remote Sensing, 6(10): 9213-9238.   DOI
6 Jia, D., C. Song, C. Cheng, S. Shen, L. Ning, and C. Hui, 2020. A novel deep learning-based spatiotemporal fusion method for combining satellite images with different resolutions using a two-stream convolutional neural network, Remote Sensing, 12(4): 698.   DOI
7 Kim, B., H. Kim, K. Song, S. Hong, and W. Lee, 2015. Analysis on technical specification and application for the medium-satellite payload in agriculture and forestry, Journal of Satellite, Information and Communications, 10(4): 117-127 (in Korean with English abstract).
8 Kim, Y. and N.-W. Park, 2019. Comparison of spatiotemporal fusion models of multiple satellite images for vegetation monitoring, Korean Journal of Remote Sensing, 35(6-3): 1209-1219 (in Korean with English abstract).   DOI
9 Lee, H.-S. and K.-S. Lee, 2017. Effect of red-edge band to estimate leaf area index in close canopy forest, Korean Journal of Remote Sensing, 33(5-1): 571-585 (in Korean with English abstract).   DOI
10 Kim, Y., P.C. Kyriakidis, and N.-W. Park, 2020. A cross-resolution, spatiotemporal geostatistical fusion model for combining satellite image time-series of different spatial and temporal resolutions, Remote Sensing, 12(10): 1553.   DOI
11 Liu, M., Y. Ke, Q. Yin, X. Chen, and J. Im, 2019. Comparison of five spatio-temporal satellite image fusion models over landscapes with various spatial heterogeneity and temporal variation, Remote Sensing, 11(22): 2612.   DOI
12 Song, H. and B. Huang, 2013. Spatiotemporal satellite image fusion through one-pair image learning, IEEE Transactions on Geoscience and Remote Sensing, 51(4): 1883-1896.   DOI
13 Song, H., Q. Liu, G. Wang, R. Hang, and B. Huang, 2018. Spatiotemporal satellite image fusion using deep convolutional neural networks, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 11(3): 821-829.   DOI
14 Xue, J. and B. Su, 2017. Significant remote sensing vegetation indices: A review of developments and applications, Journal of Sensors, 2017: 1353691.
15 Wang, Z., A.C. Bovik, H.R. Sheikh, and E.P. Simoncelli, 2004. Image quality assessment: From error visibility to structural similarity, IEEE Transactions on Image Processing, 13(4): 600-612.   DOI
16 Xie, D., J. Zhang, X. Zhu, Y. Pan, H. Liu, Z. Yuan, and Y. Yun, 2016. An improved STARFM with help of an unmixing-based method to generate high spatial and temporal resolution remote sensing data in complex heterogeneous regions, Sensors, 16(2): 207.   DOI
17 Xie, D., F. Gao, L. Sun, and M. Anderson, 2018. Improving spatial-temporal data fusion by choosing optimal input image pairs, Remote Sensing, 10(7): 1142.   DOI
18 Zhu, X., E.H. Helmer, F. Gao, D. Liu, J. Chen, and M.A. Lefsky, 2016. A flexible spatiotemporal method for fusing satellite images with different resolutions, Remote Sensing of Environment, 172: 165-177.   DOI
19 Zhang, W., A. Li, H. Jin, J. Bian, Z. Zhang, G. Lei, Z. Qin, and C. Huang, 2013. An enhanced spatial and temporal data fusion model for fusing Landsat and MODIS surface reflectance to generate high temporal Landsat-like data, Remote Sensing, 5(10): 5346-5368.   DOI
20 Zhu, X., F. Cai, J. Tian, and T.K.A. Williams, 2018. Spatiotemporal fusion of multisource remote sensing data: Literature survey, taxonomy, principles, applications, and future directions, Remote Sensing, 10(4): 527.   DOI
21 Zhu, X., J. Chen, F. Gao, X. Chen, and J.G. Masek, 2010. An enhanced spatial and temporal adaptive reflectance fusion model for complex heterogeneous regions, Remote Sensing of Environment, 114(11): 2610-2623.   DOI
22 Zurita-Milla, R., J.G. Clevers, and M.E. Schaepman, 2008. Unmixing-based Landsat TM and MERIS FR data fusion, IEEE Geoscience and Remote Sensing Letters, 5(3): 453-457.   DOI
23 Yang, J., J. Wright, T.S. Huang, and Y. Ma, 2010. Image super-resolution via sparse representation, IEEE Transactions on Image Processing, 19(11): 2861-2873.   DOI