Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
DBFNet: A Dual-Branch Fusion Network for Underwater Image Enhancement
Next Article in Special Issue
Analysis and Verification of Building Changes Based on Point Clouds from Different Sources and Time Periods
Previous Article in Journal
Visualization of Environmental Sensing Data in the Lake-Oriented Digital Twin World: Poyang Lake as an Example
Previous Article in Special Issue
Using Machine-Learning for the Damage Detection of Harbour Structures
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Global and Local Graph-Based Difference Image Enhancement for Change Detection

High-Tech Institute of Xi’an, Xi’an 710025, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(5), 1194; https://doi.org/10.3390/rs15051194
Submission received: 20 December 2022 / Revised: 18 January 2023 / Accepted: 23 January 2023 / Published: 21 February 2023
Figure 1
<p>Framework of the proposed GLGDE.</p> ">
Figure 2
<p>Datasets #1 (top row) and #2 (bottom row). From left to right are: (<b>a</b>) pre-event image; (<b>b</b>) post-event image; and (<b>c</b>) ground truth.</p> ">
Figure 3
<p>Initial and enhanced DIs of Datasets #1 and #2. From top to bottom, they correspond to initial DIs of Dataset #1, enhancement DIs of Dataset #1, initial DIs of Dataset #2, and enhancement DIs of Dataset #2. From left to right are DIs generated using: (<b>a1</b>–<b>a4</b>) Diff/E-Diff; (<b>b1</b>–<b>b4</b>) LR/E-LR; (<b>c1</b>–<b>c4</b>) MR/E-MR; (<b>d1</b>–<b>d4</b>) NR/E-NR; (<b>e1</b>–<b>e4</b>) SDCD/E-SDCD; and (<b>f1</b>–<b>f4</b>) INLPG/E-INLPG.</p> ">
Figure 4
<p>ROC and PR curves of Datasets #1 and #2. From left to right are: (<b>a</b>) ROC curves on Dataset #1; (<b>b</b>) PR curves on Dataset #1; (<b>c</b>) ROC curves on Dataset #2; and (<b>d</b>) PR curves on Dataset #2.</p> ">
Figure 5
<p>CMs computed from initial and enhanced DIs of Datasets #1 and #2. From top to bottom, they correspond to initial CMs of Dataset #1, enhancement CMs of Dataset #1, initial CMs of Dataset #2, and enhancement CMs of Dataset #2. From left to right are CMs generated by: (<b>a1</b>–<b>a4</b>) Diff/E-Diff; (<b>b1</b>–<b>b4</b>) LR/E-LR; (<b>c1</b>–<b>c4</b>) MR/E-MR; (<b>d1</b>–<b>d4</b>) NR/E-NR; (<b>e1</b>–<b>e4</b>) SDCD/E-SDCD; and (<b>f1</b>–<b>f4</b>) INLPG/E-INLPG. In the CM, White: true positives (TP); Red: false positives (FP); Black: true negatives (TN); and Cyan: false negatives (FN).</p> ">
Figure 6
<p>Datasets #3 (top row) and #4 (bottom row). From left to right are: (<b>a</b>) pre-event image; (<b>b</b>) post-event image; and (<b>c</b>) ground truth.</p> ">
Figure 7
<p>Initial and enhanced DIs of Datasets #3 and #4. From top to bottom, they correspond to initial DIs of Dataset #3, enhancement DIs of Dataset #3, initial DIs of Dataset #4, and enhancement DIs of Dataset #4. From left to right are DIs generated using: (<b>a1</b>–<b>a4</b>) CVA/E-CVA; (<b>b1</b>–<b>b4</b>) MAD/E-MAD; (<b>c1</b>–<b>c4</b>) IRMAD/E-IRMAD; (<b>d1</b>–<b>d4</b>) DSFA/E-DSFA; (<b>e1</b>–<b>e4</b>) DCVA/E-DCVA; and (<b>f1</b>–<b>f4</b>) INLPG/E-INLPG.</p> ">
Figure 8
<p>ROC and PR curves of Datasets #3 and #4. From left to right are: (<b>a</b>) ROC curves on Dataset #3; (<b>b</b>) PR curves on Dataset #3; (<b>c</b>) ROC curves on Dataset #4; (<b>d</b>) PR curves on Dataset #4.</p> ">
Figure 9
<p>CMs computed from initial and enhanced DIs of Datasets #3 and #4. From top to bottom, they correspond to initial CMs of Dataset #3, enhancement CMs of Dataset #3, initial CMs of Dataset #4, enhancement CMs of Dataset #4. From left to right are CMs generated by: (<b>a1</b>–<b>a4</b>) CVA/E-CVA; (<b>b1</b>–<b>b4</b>) MAD/E-MAD; (<b>c1</b>–<b>c4</b>) IRMAD/E-IRMAD; (<b>d1</b>–<b>d4</b>) DSFA/E-DSFA; (<b>e1</b>–<b>e4</b>) DCVA/E-DCVA; and (<b>f1</b>–<b>f4</b>) INLPG/E-INLPG.</p> ">
Figure 10
<p>Datasets #5 (top row) and #6 (bottom row). From left to right are: (<b>a</b>) pre-event image; (<b>b</b>) post-event image; and (<b>c</b>) ground truth.</p> ">
Figure 11
<p>Initial and enhanced DIs of Datasets #5 and #6. From top to bottom, they correspond to initial DIs of Dataset #5, enhancement DIs of Dataset #5, initial DIs of Dataset #6, and enhancement DIs of Dataset #6. From left to right are DIs generated using: (<b>a1</b>–<b>a4</b>) CVA/E-CVA; (<b>b1</b>–<b>b4</b>) MAD/E-MAD; (<b>c1</b>–<b>c4</b>) IRMAD/E-IRMAD; (<b>d1</b>–<b>d4</b>) DSFA/E-DSFA; (<b>e1</b>–<b>e4</b>) DCVA/E-DCVA; and (<b>f1</b>–<b>f4</b>) INLPG/E-INLPG.</p> ">
Figure 12
<p>ROC and PR curves of Datasets #5 and #6. From left to right are: (<b>a</b>) ROC curves on Dataset #5; (<b>b</b>) PR curves on Dataset #5; (<b>c</b>) ROC curves on Dataset #6; and (<b>d</b>) PR curves on Dataset #6.</p> ">
Figure 13
<p>CMs computed from initial and enhanced DIs of Datasets #5 and #6. From top to bottom, they correspond to initial CMs of Dataset #5, enhancement CMs of Dataset #5, initial CMs of Dataset #6, and enhancement CMs of Dataset #6. From left to right are CMs generated by: (<b>a1</b>–<b>a4</b>) CVA/E-CVA; (<b>b1</b>–<b>b4</b>) MAD/E-MAD; (<b>c1</b>–<b>c4</b>) IRMAD/E-IRMAD; (<b>d1</b>–<b>d4</b>) DSFA/E-DSFA; (<b>e1</b>–<b>e4</b>) DCVA/E-DCVA; and (<b>f1</b>–<b>f4</b>) INLPG/E-INLPG.</p> ">
Figure 14
<p>Sensitivity analysis of parameter <math display="inline"><semantics> <mi>α</mi> </semantics></math> in GLGDE.</p> ">
Versions Notes

Abstract

:
Change detection (CD) is an important research topic in remote sensing, which has been applied in many fields. In the paper, we focus on the post-processing of difference images (DIs), i.e., how to further improve the quality of a DI after the initial DI is obtained. The importance of DIs for CD problems cannot be overstated, however few methods have been investigated so far for re-processing DIs after their acquisition. In order to improve the DI quality, we propose a global and local graph-based DI-enhancement method (GLGDE) specifically for CD problems; this is a plug-and-play method that can be applied to both homogeneous and heterogeneous CD. GLGDE first segments the multi-temporal images and DIs into superpixels with the same boundaries and then constructs two graphs for the DI with superpixels as vertices: one is the global feature graph that characterizes the association between the similarity relationships of connected vertices in the multi-temporal images and their changing states in a DI, the other is the local spatial graph that exploits the change information and contextual information of the DI. Based on these two graphs, a DI-enhancement model is built, which constrains the enhanced DI to be smooth on both graphs. Therefore, the proposed GLGDE can not only smooth the DI but also correct the it. By solving the minimization model, we can obtain an improved DI. The experimental results and comparisons on different CD tasks with six real datasets demonstrate the effectiveness of the proposed method.

1. Introduction

1.1. Background

The change detection (CD) of remote sensing images is a technique to extract change information by comparing multi-temporal images acquired over the same geographical area but at different times [1]. Earth observation technology can provide long-term, wide-area, periodic observations of the Earth’s surface; CD is also one of the earliest and most widely used research areas of remote sensing [2], which has also achieved very successful applications in many practical tasks, such as environmental monitoring [3], agricultural surveys [4], urban studies [5], and disaster assessment [6].
Up to now, numerous CD algorithms have been investigated and one can refer to the latest review articles [7,8,9,10]. Generally, there are different ways to classify CD algorithms. For example, (1) according to the detection and analysis granularity, it can be divided into pixel-level-, object-level-, and scene-level-based CD; (2) according to the type of output result, it can be divided into binary and multiple CD; (3) depending on the usage of label data, it can be classified as supervised, semi-supervised, and unsupervised CD; (4) according to whether deep neural networks are used, it can be divided into traditional methods and deep learning-based methods; (5) depending on the source of the input images, it can be divided into homogeneous and heterogeneous CD; (6) and, according to the technology used, it can be also divided into spectral change identification methods, post-classification comparison methods, direct multi-temporal classification methods, and hybrid methods. In this paper, we focus on the homogeneous CD of optical images, homogeneous CD of synthetic aperture radar (SAR) images, and heterogeneous CD.
The CD process can usually be divided into three main sequential steps [11,12]: image preprocessing, differential image (DI) generation, and change extraction by analyzing the DI. In the preprocessing, radiometric correction and geometric coregistration are usually required, which enables pairwise images to correspond to the same geographic area. Then, it compares the two aligned images to obtain the DI in the second process, which is aimed at enhancing the contrast between the changed and unchanged regions. Finally, it classifies the DI to extract the changed region and obtain the change map (CM) in the third process.
DI generation has a significant impact in the CD process for the following reasons. (1) DIs have a direct influence on the accuracy of CM. A high-quality DI is highly discriminative between the changed and unchanged classes and, thus, only a simple threshold segmentation or clustering is needed to classify the DIs for obtaining CM [13,14,15]. (2) DIs provide richer change information. Compared to the binary CM, which can only provide certain change/unchange information, DIs are able to provide the probability of change. For some difficult problems or problems that must be handled with caution, such DIs that provide uncertainty results may be more valuable for assisting experts [11]. (3) DIs could be used to support other unsupervised CD methods. For example, DIs could assist the training process or build pseudo-training labels for some deep learning-based methods [16,17]. Therefore, we focus on how to improve the quality of DIs in the paper, while, for the final CM, we only use a simple thresholding method on DIs to obtain it.

1.2. Related Work

Here, we review some methods for obtaining DIs in different CD tasks. Define  X  and  Y  to be the multi-temporal images to be compared and let x and y be the data extracted from the same positions of  X  and  Y , respectively, which can be individual pixels, square patches or superpixels, depending on the granularity of the detection analysis. According to the sources of  X  and  Y , we review some relevant algorithms for calculating DIs.

1.2.1. DI of Homogeneous Optical Images

For the homogeneous CD of optical images, the image differencing (e.g.,  x y ) is the simplest way to calculate DIs, which is based on the assumption that the noise in optical images is usually additive. Change vector analysis (CVA) [18] calculates the change vectors by comparing the spectra of two images, which can provide both change magnitude and change direction. Based on the canonical correlations analysis, the multivariate alteration detection method (MAD) [19] and iteratively reweighted MAD method (IRMAD) [20] have been proposed, which is invariant to separate linear (affine) transformations of the spectra. Therefore, they can significantly reduce false alarms due to differences in device gain or linear radiation and atmospheric correction schemes. Lv et al. [15] have proposed an adaptive spatial–contextual extraction method (ASE) for CD of very high resolution optical images, which first adaptively selects a suitable local area for each pixel to exploit the contextual information and then uses a defined band-to-band (B2B) distance metric to calculate the change magnitude. By exploiting the convolutional neural network (CNN), Saha et al. [21] have combined CVA with deep neural networks and proposed deep change vector analysis (DCVA), which calculates the deep change vectors and extracts richer contextual information by utilizing the features from different layers of CNN. Finally, the binary and multiple CM are computed by analyzing the deep change vector. Du et al. have extended the slow feature analysis (SFA) [22] to the deep network and proposed a deep slow feature analysis (DSFA) for unsupervised CD [23]. DSFA first uses the CVA to find unchanged pixels to construct a pseudo-training set and then utilizes two symmetric deep networks to project the multi-temporal images into the latent feature space, where SFA is employed to identify changed and unchanged regions, and, finally, the DIs are computed using the chi-square distance metric and the CM is obtained using the thresholding method.

1.2.2. DIs of Homogeneous SAR Images

For the homogeneous CD of SAR images, it is difficult to detect the changes using an image differencing operator due to the inherent multiplicative speckle noise of SAR images. Alternatively, the ratio [24], log-ratio [25], and mean-ratio [26] operators are the commonly used DI calculation methods. Nar et al. [27] have proposed a sparsity-driven change detection (SDCD) method, which can reduce the influence of speckle noise on the DI by using an  1 -norm-based total variation regularization. Sun et al. [28] have proposed a nonlocal low-rank- and a two-level clustering-based method, which jointly employs the nonlocal despeckling method for computing the DI and the cascade clustering strategy for calculating the binary CM. In [29], an adaptive Contourlet fusion clustering (CFC)-based CD method has been proposed, which combines the log-ratio and mean-ratio images to generate a new DI by using the adaptive Contourlet fusion and then segments the fused DI by using a fast non-local clustering algorithm to reduce the impact of the noise. A hierarchical heterogeneous graph-driven CD method has been proposed in [30], which combines two inter-connected pixel-based and superpixel-based graph layers to fully exploit the structure information of images and then uses the graph cuts to separate the DI to obtain CM. A log cumulants and stacked autoencoder-based method has been proposed for detecting changes caused by fires using SAR images acquired by Sentinel-1 in [31], which extracts the features by using a tunable Q wavelet transformation with higher order log-cumulants statistics and uses the stacked autoencoder to classify the changed and unchanged regions. Zhang et al. [32] have proposed a spatial–temporal gray-level co-occurrence aware network (STGCNet) to suppress the influence of speckles on detecting changes, which first uses the log-ratio operation to generate an initial DI and uses the fuzzy c-means (FCM) clustering to select the reliable changed and unchanged training samples and then trains the two-stream STGCNet to mine the spatial–temporal information. A dynamic graph-level neural network (DGLNN) is proposed in [33], which builds a dynamic graph on the three-channel pixel neighborhood block constructed from the initial log-ratio-based DI and multi-temporal SAR images and learns discriminative representations for each block by using feature propagation and node aggregation in the graph.

1.2.3. DI of Heterogeneous CD

For the heterogeneous CD, the compared multi-temporal images are acquired by different sensors and characterize different physical quantities, so it is impossible to obtain the DI by directly comparing the heterogeneous images. Therefore, it is essential to transform heterogenous images into the same metric space for comparison. Liu et al. [13] have proposed the homogeneous pixel transformation (HPT) method, which uses the labeled unchanged pixel pairs to transform one image (e.g.,  X ) to the domain of the other images ( Y ) with kernel regression functions and then compare the regression image ( Y ) and original image ( Y ) in the same domain using image differencing ( Y Y ). To overcome the dependence on labeled samples, some unsupervised regression methods have also been proposed for heterogeneous CD. In [14], an affinity matrix distance (AMD) is used to pick samples that have a high probability of being unchanged and then four traditional regression functions are trained based on the pseudo-training set. Furthermore, the AMD has also been used to train deep regression networks as the prior change not only to construct training samples [17] but also to assist in the training process [16]. Mignotte proposed a fractal projection and Markovian segmentation-based algorithm (FPMS) for heterogeneous CD [34], which consists of a fractal encoding step that encodes the pre-event image and a fractal decoding step that projects the pre-event image to the domain of a post-event image and then computes and segments the DI using Markovian segmentation algorithms. Based on the structure consistency, Sun et al. proposed several graph-based heterogeneous CD methods. For example, they proposed the structure regression-based methods that constrain the structures of original images and regression images that are the same and causes the changed image to be sparse [35,36,37]; they also proposed the structure comparison-based methods that first construct two graphs of heterogeneous images to capture the structure information and then compare the structures using graph projection [11,38,39]. Recently, they have also proposed a heterogeneous CD framework based on graph signal processing [40] and analyzed the heterogeneous CD problem from the vertex and spectral domains of the graph, respectively.
According to the introduction of the above-related methods, we have the following findings.
  • Different CD problems face different challenges. (1) For the CD of homogeneous optical images, its difficulty lies in that, when the image resolution is very high, the great intraclass variation and low interclass variance as well as the influence of illuminations and seasons can lead to a lot of salt-and-pepper noise [7]. (2) For the CD of homogeneous SAR images, its difficulty lies in the inherent speckle noise and high intensity variation that can lead to difficult trade-offs between noise removal and geometrical detail preservation in the DI. (3) For the heterogeneous CD, the key lies in how to construct relationships between heterogeneous images so that incomparable images can be compared; it also faces the challenges of both the homogeneous CD of optical images and SAR images.
  • How to obtain high-quality DI is one of the keys to the CD problem. After generating the difference maps, most of these methods treat them as conventional image segmentation problems to obtain the final CM, such as [13,14,15,17,21,27,39].

1.3. Motivations

Notwithstanding that the DI quality is very important for the performance of CD, few methods have been investigated so far for re-processing DI after its acquisition. The benefits of post-processing the DI are self-evident, as it can further enhance the quality of the DI, either for the next step of computing CM with segmentation or for assisting other methods.
In order to reduce the influence of speckle noise on DIs of homogeneous SAR images, Zheng et al. [41] have combined the mean filter and the median filter to obtain a better DI, which uses the mean filter on the DI computed using the image differencing operator for smoothing and uses the median filter on the DI computed using the log-ratio operator for preserving edge information. In [42], the DIs generated using a Gauss-log ratio, the log-ratio operators are fused using a discrete wavelet transform, and then the fused DI is filtered using the nonsubsampled contourlet transform model, which can reduce the noise of the DI and keep the edge information of the changed regions. In [14,17], after obtaining the DI, the authors have removed the outliers in DI, i.e., clipping the pixels whose values are beyond a few standard deviations from the mean value of DI, and then used the fully connected conditional random field model (CRF) proposed in [43] to filter the DI, which exploits the spatial context information to improve the quality of the DI. By drawing on the convolution and pooling operations in CNN, Zhang et al. [44] have proposed a weighted average filter for the DI generated using a log-ratio operator, which can suppress the speckle influence and enhance the edges of the DI. A graph signal smoothness representation method has been proposed in [45], which uses the smoothing property of the changed signal on the fused graph to smooth the DI.
Although the above methods can improve the quality of the DI to some extent, especially in reducing the noise influence, they still have the following two shortcomings.
  • Most of these methods are for the conventional denoising and smoothing of DI and they only exploit the information of the DI itself, such as the change information (pixel value) and spatial context information, while ignoring the specificity of the change detection task and neglecting the information in the original multi-temporal images, which limits their performance.
  • Most of the methods only serve as “icing on the cake” for smoothing the DI, but cannot further correct the DI. For example, when there is an overall error in the local area in the DI, i.e., when the pixel values of the entire local area that really changed are all 0 in the DI or when the pixel values of the entire local area that really unchanged are all 1, it is difficult to correct this error based on the spatial smoothing or filtering operations.
In order to address the above challenges, in this paper, we propose a global and local graph-based DI-enhancement method (GLGDE) for CD problems. It is a plug-and-play approach for the post-processing of DI, which can be applied to the homogeneous CD of optical images and SAR images and heterogeneous CD. Specifically, once the initial DI is obtained, e.g., a coarse DI is obtained by using the methods introduced in the related work, we smooth and correct the DI by the following steps. Firstly, we co-segment the DI and multi-temporal images into superpixels with the same shape and boundaries and then extract the features of the images. Second, we construct two graphs for the DI with superpixels as vertices; one connects each vertex and its similar vertices in the multi-temporal images to exploit the association between the similarity relationships of the multi-temporal images and the changing states of the DI and the other connects each vertex and its neighboring vertices in the local spatial neighborhood to capture the local structure information of the DI. In the constructed graphs, the edge weights not only exploit the change information and contextual information of the DI but also the correlation information of the original multi-temporal images, which is crucial for correcting and smoothing DIs. Third, based on the constructed graphs, we propose a DI-enhancement model that contains three constraint terms: a global feature graph-smoothing regularization term, a local spatial graph-smoothing regularization term, and a change data-regularization term. By solving the minimization model, we can obtain the improved DI. Finally, the binary CM can be easily obtained using threshold segmentation.

1.4. Contributions

The main contributions of this paper can be summarized as follows.
  • First, we have designed a DI-enhancement algorithm specifically for the change detection task, which is a plug-and-play approach for DI post-processing. This is a rarely found work specifically designed for smoothing and correcting DIs in CD problems.
  • Second, the proposed DI-enhancement algorithm, named GLGDE for short, not only can smooth the DI but also correct it by using the constructed global feature graph and local spatial graph, which can fully fuse and utilize the change and contextual information in the DI and correlation information in the multi-temporal images.
  • Third, due to using superpixels as vertices, the scale of the model is small. The algorithm achieves DI improvement with low computational complexity, which would be of great practical value. Extensive experiments in different CD scenarios, i.e., homogeneous CD of SAR and optical images and heterogeneous CD, demonstrate the effectiveness of the proposed method.

1.5. Outline

The rest of this paper is structured as follows: Section 2 describes the proposed global and local graph-based DI smoothing and correction method, and Section 3 provides the experimental results. Finally, Section 4 concludes this paper.

2. Global and Local Graph-Based DI Enhancement

With two co-registered multi-temporal images acquired using the same (homogeneous) or different (heterogeneous) sensors, denoted as  X R M × N × C x  and  Y R M × N × C y , and the initial coarse DI obtained from other methods, denoted as  D R M × N , the goal of this paper is to improve the quality of the DI to obtain an enhanced DI, which has a great discrimination between changed and unchanged classes. We denote the pixels of the images as  x m , n , c y m , n , c  and  d m , n , respectively, and suppose that the DI is normalized, i.e.,  0 d m , n 1 , and that a larger value of  d m , n  indicates a higher probability of change in the region represented by  m , n .
Next, we describes the proposed DI-enhancement method in detail, which contains three main steps: pre-processing, constructing the global and local graphs, and solving the GLGDE model. Figure 1 shows the framework of the proposed DI-enhancement method.

2.1. Pre-Processing

As aforementioned, we use the graphs to capture the structure information of images. We choose to use superpixels instead of individual pixels or square patches as graph vertices, which provides two benefits. First, the size of the graph is greatly reduced, e.g., for a  1000 × 1000  image with individual pixels as vertices, the size of the graph is  10 6 , which results in a heavy computational burden for the algorithm. On the contrary, when using superpixels as vertices, the size of the graph is only related to the number of superpixels, which will greatly reduce the complexity of the algorithm. Second, the superpixel contains rich contextual information and is able to preserve the shape and edges of the object, which is very important for the CD problem with very-high-resolution images.
In order to segment multiple images (i.e.,  X Y , and  D ) into superpixels with the same boundaries, we first construct a false RGB image  Z R M × N × 3 . The first and second channels of  Z  are the normalized grayscale images of  X  and  Y , respectively, and the third channel of  Z  is the initial DI of  D . Note that if there is an SAR image in the multi-temporal images, e.g.,  X , the corresponding channel of  Z  is the normalized log-transformed SAR image, which can convert the multiplicative noise of the SAR image into additive noise, thus facilitating the distance calculation in the subsequent superpixel segmentation. Then, the Gaussian mixture model-based superpixel segmentation method (GMMSP) [46] is employed to segment  Z  into  N S  regions, in which each superpixel is associated with a weighted sum of Gaussian functions. GMMSP can efficiently generate superpixels that adhere to the boundaries of an object. We define the superpixel co-segmentation map as  Λ , then we can obtain the segmented superpixels of  X i Y i , and  D i  by projecting the map  Λ  to the multiple images (i.e.,  X Y , and  D ) as:
X i = x m , n , c | m , n Λ i , c = 1 , , C x Y i = y m , n , c | m , n Λ i , c = 1 , , C y D i = d m , n | m , n Λ i
Then,  X i Y i , and  D i  represent the same region. More importantly, the pixels inside each superpixel obtained through co-segmentation are mostly internally homogeneous, i.e., they represent the same kind of objects, so the pixels inside each superpixel are likely to belong to the same class of labels, i.e., all changed or unchanged. In this way, the computational complexity can be reduced by using superpixels and the interference caused by inconsistent internal pixel labels can be avoided by using contextual information.
Once the co-segmentation superpixels are obtained, different features can be extracted to capture the different information of superpixel, such as the intensity (spectra) and textual and structure information. In this paper, we simply use the mean, median and variance from each band of  X i  and  Y i  as the superpixel feature. Naturally, some other features are also available. Then, we can obtain the feature matrices of the images  X  and  Y , denoted as  X ˜  and  Y ˜ , respectively, where each column corresponds to the superpixel feature vector. We define the index set as  I = 1 , , N S , the feature distance between  X i  and  X j  as  d i s t i , j x = X ˜ i X ˜ j 2 2 , and the feature distance between  Y i  and  Y j  as  d i s t i , j y = Y ˜ i Y ˜ j 2 2 , we denote the i-th distance vectors as  d i s t i x = d i s t i , j x | j I  and  d i s t i y = d i s t i , j y | j I , and we define the labels of the i-th superpixel as  L i , with  L i = 1  indicating changed and  L i = 0  indicating unchanged.

2.2. Global Feature Graph

We construct two graphs for the DI, which can capture the change information and contextual information in the DI and utilize the relationships between the multi-temporal images to enhance the DI.
We define the global feature graph as  G f = V f , E f , w , which sets each superpixel of DI as the vertex, i.e.,  V f = D i | i = 1 , , N S , and connect each superpixel with the superpixel corresponding to its K-nearest neighbor (KNN) in the original multi-temporal images, that is, for the i-th and j-th superpixels of  D i  and  D j ; if  X j  is the KNN of  X i  or  Y j  is the KNN of  Y i , then  D i  and  D j  are connected by the edge  E f , defined as:
E f = i , j | i I , j N i x N i y
where  N i x  and  N i y  denote the position sets of KNN of  X i  and  Y i , respectively. Furthermore, the KNN sets of  N i x  and  N i y  are computed as: if and only if  d i s t i , j x  belongs to one of the K-smallest elements of the distance vectors of  d i s t i x  or  d i s t j x , then  j N i x ; for the  N i y , we have a similar definition. Next, we investigate the association between the similarity relationships of connected vertices in the multi-temporal images and their changing states in DI.
First, we consider the i-th vertex and j-th vertex using  j N i x . Since  X j  is the KNN of  X i , we have that  X i  and  X j  have a high probability of belonging to the same kind of object, e.g., road. However, for the corresponding  Y i  and  Y j  in the post-event image connected by this edge  i , j j N i x , we have two different states.
State #1, if  Y i  and  Y j  also belong to the same kind of object, then we have that the labels of the i-th and j-th superpixels should be same. That is,  L i = L j = 1  when  Y i  and  Y j  also belong to the same kind of object as  X i , e.g., road, and  L i = L j = 0 , when  Y i  and  Y j  belong to the same kind of object that is different from  X i , e.g., water. In state #1, we have  L i = L j , that is, the pixel values of  D i  and  D j  in the DI should be very close (all large or all small).
State #2, if  Y i  and  Y j  belong to the different kinds of objects, then we have that the labels of the i-th and j-th superpixels should be different. That is, (1)  L i = 1 L j = 0  when  Y i  belongs to the object of the same kind as  X i  while  Y j  belongs to the object of the different kind as  X i  (e.g., water); (2)  L i = 1 L j = 0  when  Y i  belongs to the object of the different kinds as  X i  (e.g., water) while  Y j  belongs to the object of the same kind as  X i . In state #2, we have  L i L j , that is, the pixel values of  D i  and  D j  in the DI should be very different (one is large, the other is small).
We define  p i  as the probability that the i-th superpixel belongs to the changed class. In these two states, we can find that the value of  p i p j 2 j N i x  is determined by the probability of whether  Y i  and  Y j j N i x  belong to the same kind of object. The latter can be measured using the following Gaussian kernel function as:
f i , j y = exp 2 d i s t i , j y + min m N i y d i s t i , m y + min m N j y d i s t j , m y , j N i x
Then, we have that the larger the value of  f i , j y , the more likely it is that  Y i  and  Y j  belong to the same kind of object, the smaller the distance of  p i p j 2 . On the contrary, the smaller the value of  f i , j y , the more likely it is that  Y i  and  Y j  represent different kinds of objects, the larger the distance of  p i p j 2 .
Second, we consider the i-th vertex and the j-th vertex with  j N i y . Since  Y j  is the KNN of  Y i , we have that  Y i  and  Y j  have a high probability of belonging to the same kind of object. However, for the corresponding  X i  and  X j  in the pre-event image connected by this this edge  i , j j N i y , we also have two different states, similarly to state #1 and state #2.
State #3, if  X i  and  X j  also belong to the same kind of object, then we have that the labels of the i-th and j-th superpixels should be same, i.e.,  L i = L j .
State #4, if  X i  and  X j  belong to the different kinds of objects, then we have that the labels of the i-th and j-th superpixels should be different, i.e.,  L i L j .
Similarly, the value of  p i p j 2  is determined by the probability of whether  X i  and  X j  belong to the same kind of object, which can be measured using the following function:
f i , j x = exp 2 d i s t i , j x + min m N i x d i s t i , m x + min m N j x d i s t j , m x , j N i y
We also have that the larger the value of  f i , j x , the smaller the distance of  p i p j 2  and the smaller the value of  f i , j x , the larger the distance of  p i p j 2 .
By using Equations (3) and (4), we can set the edge weight of the graph  G f  as:
W i , j f = f i , j x δ j N i y + f i , j y δ j N i x , i , j E f
where  δ ·  is the discriminant function: when the condition in parentheses holds, it adopts the value of 1; otherwise, it adopts the value of 0.
In the graph  G f , the i-th and j-th superpixels of  D i  and  D j  are connected by the edge of  i , j E f  with the weight  W i , j f  of (5). We can find that the edges are constructed in the global feature space of the multi-temporal images and the weights are determined based on the relationships between the multi-temporal images.
Based on the above analysis, we have the following regularization:
min p R N S i , j E f W i , j f p i p j 2
By defining the Laplacian matrix of graph  G f  as  L f , the regularization (6) can be rewritten as:
min p R N S 2 p T L f p
Then, we can find that it requires the DI to be smooth on the graph  G f . This global feature graph-induced change-smoothness-based regularization has two advantages: first, it characterizes the association between the similarity relationships of the connected vertices in the multi-temporal images and their changing states in DI, which can be used to smooth the DI globally, i.e., to correct the DI, as it exploits information in the global feature space rather than the local spatial information. Second, this change smoothness is widespread because we make no other qualifying assumptions on the CD problem in the derivation, so it can be applied to the both homogeneous and heterogeneous CD problems and it can also be applied to other methods as a constraint on the change, such as the image regression-based CD methods [35,36,47].

2.3. Local Spatial Graph

We define the local spatial graph as  G s = V s , E s , w , which sets each superpixel of the DI as the vertex, i.e.,  V s = D i | i = 1 , , N S , and connected each superpixel with its spatially close neighbors, defined as R-adjacent neighbors. That is, two superpixels of  D i  and  D j  are connected by the graph  G s  as long as their boundaries intersect or the distance between their center points is less than R, denote as  i N j R  or  j N i R . Since the average pixel number of the co-segmented superpixels is  M N / N S , we set  R = 2 M N / N S  for simplicity. Then, we have:
E s = i , j | i I , j N i R
Inspired by the n-links proposed in [38], we set the weight of  W i , j s  for the graph  G s  as:
W i , j s = 1 d i s t i , j Λ × exp d i s t i , j y 2 σ 1 2 d i s t i , j x 2 σ 2 2 , if d i s t i , j y σ 1 2 , d i s t i , j x σ 2 2 exp d i s t i , j y 2 σ 1 2 d i s t i , j x 2 σ 2 2 1 , if d i s t i , j y σ 1 2 , d i s t i , j x > σ 2 2 exp d i s t i , j y 2 σ 1 2 + d i s t i , j x 2 σ 2 2 1 , if d i s t i , j y > σ 1 2 , d i s t i , j x σ 2 2 exp 1 , if d i s t i , j y > σ 1 2 , d i s t i , j x > σ 2 2
where  d i s t i , j Λ  denote the spatial distance between the center points of  Λ i  and  Λ j σ 1 2 = d i s t i , j y  and  σ 2 2 = d i s t i , j x  are the two normalized parameters using  ·  to calculate the mean value of the feature distance throughout the entire image. From (9), there are four different cases of edge weight assignment: a large weight when  X i  and  X j  and  Y i  and  Y j  are both similar; a small weight when  X i  and  X j  are not similar but  Y i  and  Y j  are similar; a small weight when  X i  and  X j  are similar but  Y i  and  Y j  are not similar; and a median weight when  X i  and  X j  and  Y i  and  Y j  are both not similar.
With this defined edge weight  W i , j s , the constructed local spatial graph  G s  not only exploits the contextual information of the DI but also the similarity information of the multi-temporal image. Based on spatial continuity, we have the following regularization:
min p R N S i , j E s W i , j s p i p j 2
From (10), we can find that for the spatially adjacent  D i  and  D j , the regularization (10) requires that their values are close, especially for those that originally have the same similarity relationship in the two multi-temporal images. By defining the Laplacian matrix of graph  G s  as  L s , the regularization (10) can be rewritten as:
min p R N S 2 p T L s p
It can be regarded as a local spatial smoothness constraint for the DI.

2.4. GLGDE Model

For the DI-enhancement model, we have a change data regularization term, which constrains the difference between the enhanced DI and the original DI, defined as:
min p R N S p d ¯ 2 2
where  d ¯  is the mean vector of DI with  d ¯ i  being the mean value of the superpixel  D i  as  d ¯ i = m , n Λ i d m , n Λ i .
By combining the global feature graph  G f -induced regularization (GRGR) of (7), the local spatial graph  G s -induced regularization (LSGR) of (11) and the change data regularization (CDR) of (12), we have the final GLGDE model as:
min p R N S p d ¯ 2 2 + α p T L f p + β p T L s p
where  α , β > 0  are the balance parameters that control the weights of the GFGR and LSGR, respectively. Here, the larger the  α  and  β , the smoother the DI on the graphs of  G f  and  G s , respectively. The closed-form solution of the minimization problem (13) is:
p * = I N S + α L f + β L s 1 d ¯
where  I N S  denotes the  N S × N S  identity matrix. Because  p i *  represents the change level of the i-th superpixel, we can obtain the enhanced DI as  D * :
d * m , n = p i * , if m , n Λ i
From the solution of  p *  and  D * , we can find that it is a process of improving the DI using the graph model, i.e., requiring the DI to be smooth on two different graphs of  G f  and  G s .
Once the enhanced DI is computed, the binary CM can be calculated by using some image-segmentation methods, for example, the thresholding methods such as the OTSU threshold [48] or the kittler and Illingworth (KI) threshold [49], the clustering methods such as k-means clustering [50] or the Fuzzy c-means clustering [51], or the random field-based methods such as the Markov random field [37] or conditional random field [52]. Since the focus of this paper is on how to improve DI quality and not on how to segment DI, we directly use the OTSU method to classify the DI into changed and unchanged classes to obtain the final CM. The framework of the global and local graph-based DI-enhancement for CD is summarized in Algorithm 1.
Algorithm 1: GLGDE-based CD.
Input: Images of  X  and  Y , initial DI of  D .
      Parameters of  N S α , and  β .
Pre-processing:
   Segment  X Y , and  D  into superpixels with GMMSP.
   Extract the features to obtain  X ˜  and  Y ˜ .
Graph construction:
   Find the KNN sets of  N i x  and  N i y .
   Find the R-adjacent neighbors of  N i R .
   Construct the graphs of  G f  and  G s .
Model solving:
   Compute the  p *  by using (14).
   Compute the  D *  by using (15).
   Compute final CM by using OTSU thresholding method.

3. Experimental Results and Discussions

In this section, we demonstrate the performance of the proposed GLGDE through experiments, which are conducted on different CD tasks (homogeneous CD of SAR images and optical images and heterogeneous CD) with six datasets, as listed in Table 1.

3.1. Experimental Settings

The main parameters of the GLGDE are the number of superpixels  N S , the balance parameters of  α  and  β . For all the experimental results, we set  N S = 5000 α = 0.5 , and  β = i , j E f W i , j f i , j E s W i , j s × α  for GLGDE. The impact of these parameters will be discussed in Section 3.3.
To measure the effect of DI enhancement, we use two types of evaluation metrics. First, we evaluate the quality of the DI directly by using: (1) the precision-recall (PR) curve along with the areas under the PR curve (AUP); (2) the empirical receiver operating characteristics (ROC) curve along with the areas under the ROC curve (AUR). Second, we evaluate the quality of the DI indirectly by assessing the CM obtained using the OTSU thresholding, which can be measured using the False alarm rate (Fa), Miss rate (Mr), overall accuracy (Oa), and the Kappa coefficient (Kc) computed using  Fa = FP / ( FP + TN ) Mr = FN / ( TP + FN ) Oa = ( TP + TN ) / ( TP + TN + FP + FN ) , and  Kc = ( Oa PRE ) / ( 1 PRE )  with:
PRE = ( TP + FN ) ( TP + FP ) + ( TN + FP ) ( TN + FN ) ( TP + TN + FP + FN ) 2 ,
where TN, TP, FN, and FP represent the true negatives, true positives, false negatives, and false positives, respectively. At the same time, in the CM, we mark the TN, TP, FN, and FP with different colors.

3.2. Experimental Results

3.2.1. Homogeneous CD of SAR Images

Two pairs of SAR images are used in this task. Datasets #1 and #2 are both collected using a Radarsat-2 SAR sensor over the Yellow River Estuary, China, as shown in Figure 2. The pre-event images and post-event images are acquired in June 2008 and June 2009, respectively, the spatial resolution of the images is 8 m/pixel, and the ground change maps indicate the newly irrigated areas over Yellow River Estuary.
To obtain the initial DI, we choose the difference operator (Diff), log-ratio operator (LR), mean-ratio operator (MR) [26], neighborhood-ratio operator (NR) [53], sparsity-driven change-detection (SDCD) method [27], and the improved nonlocal patch-based graph (INLPG) model [11] as the baselines. In the MR and NR, the patch size is set to  3 × 3 ; in the SDCD, the regularization parameter  λ  manually selects the best examples by adjusting it from  10 4  to  10 1  with 20 logarithmic intervals; in the INLPG, the default parameters of the official code are used directly. After the initial DI is computed, the proposed GLGDE is applied to each DI to obtain the enhanced DI, denoted as the E-Diff, E-LR, E-MR, E-NR, E-SDCD, and E-INLPG, respectively.
Figure 3 shows the initial DIs and enhanced DIs of Datasets #1 and #2, Figure 4 shows the ROC and PR curves of these DIs, and Table 2 lists the corresponding AUR and AUP. From these results, we have three findings: first, the qualities of the enhanced DIs generated using the proposed GLGDE are much higher than that of the initial DIs, which means that the GLGDE can increase the contrast between changed and unchanged in the DI. For example, by comparing the DIs generated using Diff and E-Diff in Figure 3a, it can be seen that, with the former, it is difficult to detect changes, while the latter can clearly highlight the changes. Second, GLGDE is not only a local spatial smoothing of DI but, more importantly, it can correct the DI, that is, it can modify the DI by introducing the correlation information of the original multi-temporal images, as illustrated by Figure 3e. Third, the poorer the initial DI performance, the more significant the improvement of the GLGDE. It can also be found that INLPG can achieve better performance in these initial DI-generation methods.
Figure 5 shows the binary CMs generated from the initial and enhanced DIs by using the OTSU thresholding on Datasets #1 and #2, where we mark the false positives (FP) and false negatives (FN) with different colors (red and cyan) for easy observation and comparison. Table 2 lists the corresponding Fa, Mr, Oa, and Kc. We can see that there are a large number of errors in the original CMs and, in the CMs obtained after using GLGDE enhancement, these errors are corrected and the resulting CMs are much more accurate. For example, as reported in Table 2, the average improvements (Avg.ipv) of GLGDE in Dataset #1 on the AUR, AUP, Oa, and Kc metrics are 0.123, 0.278, 0.153, and 0.359, respectively.

3.2.2. Homogeneous CD of Optical Images

Two pairs of optical images are used in this task. Datasets #3 and #4 are both collected from Google Earth over Beijing, China, as shown in Figure 6. The pre-event images and post-event images were acquired in September 2012 and March 2013, respectively, the spatial resolution of the images is 1m/pixel, and the ground change maps indicate the newly constructed buildings.
To obtain the initial DI, we choose the change vector analysis (CVA) [18], multivariate alteration detection method (MAD) [19], iteratively reweighted MAD method (IRMAD) [20], deep slow feature analysis network (DSFA) [23], deep CVA network (DCVA) [21], and the INLPG [11] as the baselines. After the initial DI is computed, the proposed GLGDE is applied to each DI to obtain the enhanced DI, denoted as the E-CVA, E-MAR, E-IRMAD, E-DSFA, E-DCVA, and E-INLPG, respectively.
Figure 7 shows the initial DIs and enhanced DIs of Datasets #3 and #4, Figure 8 shows the ROC and PR curves of these DIs, and Table 3 lists the corresponding AUR and AUP. As can be seen from Figure 6 and Figure 7, the seasonal differences in the pre-event and post-event images lead to the poor qualities of the initial DIs obtained by some methods, such as the CVA and DCVA on Dataset #3 and MAD and IRMAD on Dataset #4. By using the proposed GLGDE on these initial DIs, the distinguishability between changed and unchanged is greatly improved, as shown in Figure 7 and Figure 8 and especially illustrated by the PR curves of Figure 8. The average improvements of GLGDE in Datasets #3 and Datasets #4 on the AUP metric are 0.421 and 0.450, respectively.
Figure 9 shows the binary CMs of Datasets #3 and #4 generated by using the OTSU thresholding on the DIs of Figure 7 and Table 3 lists the corresponding Fa, Mr, Oa, and Kc. For Dataset #3, there are lots of false alarms in the initial CMs of CVA, MAD, and DCVA and lots of miss detection in the initial CMs of IRMAD and DSFA, though most of these errors are corrected after the enhancement of GLGDE. Therefore, it can be seen that GLGDE is not only a smoothing of DI but also a correction of DI. It enables the entire error area to be corrected, which is not possible in the common spatial smoothing methods that only use contextual information of a DI, as shown in Figure 7 and Figure 9. The average improvements of GLGDE in Dataset #3 and Dataset #4 on the Kc metric are 0.414 and 0.148, respectively.

3.2.3. Heterogeneous CD

Two pairs of heterogeneous images are used in this task, as shown in Figure 10. In Dataset #5, the pre-event image is collected using Landsat-5 with the near-infrared band in September 1995, the post-event image is obtained from Google Earth using R, G, and B bands in July 1996, the spatial resolution of the images is 30 m/pixel, and the ground change map indicates the Lake expansion in Sardinia, Italy. In Dataset #6, the pre-event image is collected using a Radarsat-2 SAR sensor in June 2008, the post-event image is obtained from Google Earth with R, G, and B bands in September 2012, the spatial resolution of the images is 8 m/pixel, and the ground change map indicates the building construction in Shuguang Village, China.
To obtain the initial DI, we choose the homogeneous pixel transformation (HPT) method [13], affinity matrix distance-based image regression (AMDIR) [14], adaptive local structure consistency-based method (ALSC) [54], INLPG [11], fractal projection and Markovian segmentation-based algorithm (FPMS) [34], and sparse-constrained adaptive structure consistency-based method (SCASC) [37] as the baselines. For the HPT, we use 40% of the unchanged pixels as the training samples; for other methods, we directly use the default parameters of their official codes. After the initial DI is computed, the proposed GLGDE is applied to each DI to obtain the enhanced DI, denoted as the E-HPT, E-AMDIR, E-ALSC, E-INLPG, E-FPMS, and E-SCASC, respectively.
Figure 11 shows the initial DIs and enhanced DIs of Datasets #5 and #6, Figure 12 shows the ROC and PR curves of these DIs, and Table 4 lists the corresponding AUR and AUP. From these results, we can see that the initial DIs all show some change information and, among these DIs, FPMS and INLPG perform relatively better. By comparing the initial DIs and enhanced DIs of Figure 11, two findings can be noted: first, the enhanced DIs generated using GLGDE can cause the change area to be more continuous than the corresponding initial DIs, such as the DIs obtained using SCASC and E-SCASC in Figure 11(f1–f4); second, the GLGDE can effectively suppress the unchanged regions with high change levels in the initial DI, thus reducing the interference of the background area on change detection, such as the DIs obtained using AMDIR and E-AMDIR in Figure 11(b1–b4). The ROC and PR curves in Figure 12 also verify the enhancement of GLGDE on DI and the average improvements of GLGDE in Datasets #5 and Datasets #6 on the AUP metric are 0.276 and 0.236, respectively.
Figure 13 shows the CMs of Datasets #5 and #6 using different DIs with OTSU thresholding and Table 4 lists the corresponding quantitative measures. It can be seen that the GLGDE can effectively improve the detection performance. For example, GLGDE can significantly reduce false alarms in HPT and AMDIR in Figure 13(a1–a4) and Figure 13(b1–b4), respectively, and preserve the edges of the detected area, such as the results on Dataset #6. As reported in Table 4, the average improvements of GLGDE in Datasets #5 and #6 on the Kc metric are 0.196 and 0.252, respectively.

3.3. Parameter Analysis

The main parameters of the proposed GLGDE are the number of superpixels  N S  and the balance parameters of  α  and  β .
Generally, the  N S  should be selected according to the image resolution and granularity requirement of the CD task. A larger  N S  cause the segmented superpixel to be smaller, which improves the detection granularity but also increases the computational complexity. In addition, we have the following notes for the influence of the image resolution on the proposed method. First, high-resolution images usually contain more detailed structure and textural information, so when performing superpixel co-segmentation in the pre-processing, the segmented superpixels are required to be finer, i.e.,  N S  is larger. Second, when the spatial resolution of the compared image is very high, the great intraclass variation and low interclass variance usually lead to a lot of salt-and-pepper noise in the original DI. At the same time, it is difficult to accurately characterize the structure of high-resolution images with the KNN graphs. Therefore, when facing the very high-resolution images, the following two issues need to be considered in the proposed method: first, the number of superpixels  N S  should be increased appropriately; second, the KNN graphs used to capture the structures of multi-temporal images and the DI may need to be replaced with the more advanced graph neural networks (GNN).
The balance parameters of  α  and  β  are used to control the weights of the GFGR and LSGR in the DI-enhancement model (13), respectively. First, in order to cause the GFGR and LSGR to be equivalent (the penalty is balanced), we set  β = i , j E f W i , j f i , j E s W i , j s × α , which makes causes the values of  α p T L f p  and  β p T L s p  to be approximately at the same level. Second, to measure the impact of  α , we adjust the  α  from  2 4  to  2 2  with ratio of 2 and plot the average improvement of the AUP generated using GLGDE with different  α  in Figure 14. It can be found that too large and too small  α  are both not suitable: first, if  α  is too large, the GLGDE tends to over-smooth the DI, causing all the superpixels connected by the constructed graphs ( G f  and  G s ) to converge to the same; second, if  α  is too small, then GFGR and LSGR cannot play a noticeable effect in the DI-enhancement model, which limits the performance of GLGDE. Based on Figure 14, we fix  α = 0.5  in this paper for simplicity.

4. Conclusions

In this paper, we focus on the post-processing of the DI that is rarely noticed by other methods and propose a DI-enhancement method, named GLGDE. Once the initial DI is obtained, GLGDE first segments the DI and multi-temporal images into superpixels and then constructs a global feature graph and a local spatial graph with superpixels as vertices for the DI, which can exploit the change information and contextual information in the DI and the correlation information in the multi-temporal images. With the constructed graphs, the DI-enhancement model is built using three terms: a global-feature-graph-induced regularization term, a local spatial-graph-induced regularization term, and a change-data regularization term. By solving the minimization model, we can obtain the improved DI. Different from the previous DI-smoothing algorithm that only uses the contextual information of the DI itself, the proposed GLGDE can also exploit the association information of the multi-temporal images, so it can not only smooth the DI but also correct the DI. Therefore, it is a DI-enhancement method specifically designed for a CD problem, which takes into account the characteristics of the CD problem. Finally, extensive experiments in different CD scenarios demonstrate the effectiveness of the proposed method. In our future study, we will exploit the global feature graph and local spatial graph in the DI segmentation process of a CD problem to improve the detection performance of the final change map.

Author Contributions

Methodology, X.Z.; software, X.Z. and D.G.; validation, X.Z., B.L. and Z.C. Original draft preparation, X.Z. and D.G.; writing—review and editing, L.P. and D.G.; supervision, B.L. and L.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 12171481 and in part by the Natural Science Basic Research Plan in Shaanxi Province of China under Grant 2022JQ-694.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Singh, A. Review Article Digital change detection techniques using remotely-sensed data. Int. J. Remote Sens. 1989, 10, 989–1003. [Google Scholar] [CrossRef] [Green Version]
  2. Bovolo, F.; Bruzzone, L. The time variable in data fusion: A change detection perspective. IEEE Geosci. Remote Sens. Mag. 2015, 3, 8–26. [Google Scholar] [CrossRef]
  3. Kennedy, R.E.; Townsend, P.A.; Gross, J.E.; Cohen, W.B.; Bolstad, P.; Wang, Y.; Adams, P. Remote sensing change detection tools for natural resource managers: Understanding concepts and tradeoffs in the design of landscape monitoring projects. Remote Sens. Environ. 2009, 113, 1382–1396. [Google Scholar] [CrossRef]
  4. Gil-Yepes, J.L.; Ruiz, L.A.; Recio, J.A.; Balaguer-Beser, Á.; Hermosilla, T. Description and validation of a new set of object-based temporal geostatistical features for land-use/land-cover change detection. ISPRS J. Photogramm. Remote Sens. 2016, 121, 77–91. [Google Scholar] [CrossRef]
  5. Taubenböck, H.; Esch, T.; Felbier, A.; Wiesner, M.; Roth, A.; Dech, S. Monitoring urbanization in mega cities from space. Remote Sens. Environ. 2012, 117, 162–176. [Google Scholar] [CrossRef]
  6. Zhang, P.; Ban, Y.; Nascetti, A. Learning U-Net without forgetting for near real-time wildfire monitoring by the fusion of SAR and optical time series. Remote Sens. Environ. 2021, 261, 112467. [Google Scholar] [CrossRef]
  7. Lv, Z.; Liu, T.; Benediktsson, J.A.; Falco, N. Land Cover Change Detection Techniques: Very-high-resolution optical images: A review. IEEE Geosci. Remote Sens. Mag. 2022, 10, 44–63. [Google Scholar] [CrossRef]
  8. You, Y.; Cao, J.; Zhou, W. A survey of change detection methods based on remote sensing images for multi-source and multi-objective scenarios. Remote Sens. 2020, 12, 2460. [Google Scholar] [CrossRef]
  9. Shafique, A.; Cao, G.; Khan, Z.; Asad, M.; Aslam, M. Deep learning-based change detection in remote sensing images: A review. Remote Sens. 2022, 14, 871. [Google Scholar] [CrossRef]
  10. Shi, W.; Zhang, M.; Zhang, R.; Chen, S.; Zhan, Z. Change detection based on artificial intelligence: State-of-the-art and challenges. Remote Sens. 2020, 12, 1688. [Google Scholar] [CrossRef]
  11. Sun, Y.; Lei, L.; Li, X.; Tan, X.; Kuang, G. Structure Consistency-Based Graph for Unsupervised Change Detection with Homogeneous and Heterogeneous Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4700221. [Google Scholar] [CrossRef]
  12. Bruzzone, L.; Prieto, D. An adaptive semiparametric and context-based approach to unsupervised change detection in multitemporal remote-sensing images. IEEE Trans. Image Process. 2002, 11, 452–466. [Google Scholar] [CrossRef]
  13. Liu, Z.; Li, G.; Mercier, G.; He, Y.; Pan, Q. Change Detection in Heterogenous Remote Sensing Images via Homogeneous Pixel Transformation. IEEE Trans. Image Process. 2018, 27, 1822–1834. [Google Scholar] [CrossRef]
  14. Luppino, L.T.; Bianchi, F.M.; Moser, G.; Anfinsen, S.N. Unsupervised Image Regression for Heterogeneous Change Detection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9960–9975. [Google Scholar] [CrossRef] [Green Version]
  15. Lv, Z.; Wang, F.; Liu, T.; Kong, X.; Benediktsson, J.A. Novel Automatic Approach for Land Cover Change Detection by Using VHR Remote Sensing Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  16. Luppino, L.T.; Hansen, M.A.; Kampffmeyer, M.; Bianchi, F.M.; Moser, G.; Jenssen, R.; Anfinsen, S.N. Code-Aligned Autoencoders for Unsupervised Change Detection in Multimodal Remote Sensing Images. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–13. [Google Scholar] [CrossRef] [PubMed]
  17. Luppino, L.T.; Kampffmeyer, M.; Bianchi, F.M.; Moser, G.; Serpico, S.B.; Jenssen, R.; Anfinsen, S.N. Deep Image Translation with an Affinity-Based Change Prior for Unsupervised Multimodal Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4700422. [Google Scholar] [CrossRef]
  18. Malila, W.A. Change vector analysis: An approach for detecting forest changes with Landsat. In Proceedings of the LARS Symposia, West Lafayette, IN, USA, 3–6 June 1980; p. 385. [Google Scholar]
  19. Nielsen, A.A.; Conradsen, K.; Simpson, J.J. Multivariate alteration detection (MAD) and MAF postprocessing in multispectral, bitemporal image data: New approaches to change detection studies. Remote Sens. Environ. 1998, 64, 1–19. [Google Scholar] [CrossRef] [Green Version]
  20. Nielsen, A.A. The Regularized Iteratively Reweighted MAD Method for Change Detection in Multi- and Hyperspectral Data. IEEE Trans. Image Process. 2007, 16, 463–478. [Google Scholar] [CrossRef] [Green Version]
  21. Saha, S.; Bovolo, F.; Bruzzone, L. Unsupervised Deep Change Vector Analysis for Multiple-Change Detection in VHR Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 3677–3693. [Google Scholar] [CrossRef]
  22. Wu, C.; Du, B.; Zhang, L. Slow Feature Analysis for Change Detection in Multispectral Imagery. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2858–2874. [Google Scholar] [CrossRef]
  23. Du, B.; Ru, L.; Wu, C.; Zhang, L. Unsupervised Deep Slow Feature Analysis for Change Detection in Multi-Temporal Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9976–9992. [Google Scholar] [CrossRef] [Green Version]
  24. Moser, G.; Serpico, S. Generalized minimum-error thresholding for unsupervised change detection from SAR amplitude imagery. IEEE Trans. Geosci. Remote Sens. 2006, 44, 2972–2982. [Google Scholar] [CrossRef]
  25. Bovolo, F.; Bruzzone, L. A detail-preserving scale-driven approach to change detection in multitemporal SAR images. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2963–2972. [Google Scholar] [CrossRef]
  26. Inglada, J.; Mercier, G. A New Statistical Similarity Measure for Change Detection in Multitemporal SAR Images and Its Extension to Multiscale Change Analysis. IEEE Trans. Geosci. Remote Sens. 2007, 45, 1432–1445. [Google Scholar] [CrossRef] [Green Version]
  27. Nar, F.; Özgür, A.; Saran, A.N. Sparsity-Driven Change Detection in Multitemporal SAR Images. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1032–1036. [Google Scholar] [CrossRef]
  28. Sun, Y.; Lei, L.; Guan, D.; Li, X.; Kuang, G. SAR Image Change Detection Based on Nonlocal Low-Rank Model and Two-Level Clustering. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 293–306. [Google Scholar] [CrossRef]
  29. Zhang, W.; Jiao, L.; Liu, F.; Yang, S.; Liu, J. Adaptive Contourlet Fusion Clustering for SAR Image Change Detection. IEEE Trans. Image Process. 2022, 31, 2295–2308. [Google Scholar] [CrossRef] [PubMed]
  30. Wang, J.; Zhao, T.; Jiang, X.; Lan, K. A Hierarchical Heterogeneous Graph for Unsupervised SAR Image Change Detection. IEEE Geosci. Remote. Sens. Lett. 2022, 19, 4516605. [Google Scholar] [CrossRef]
  31. Planinšič, P.; Gleich, D. Temporal Change Detection in SAR Images Using Log Cumulants and Stacked Autoencoder. IEEE Geosci. Remote Sens. Lett. 2018, 15, 297–301. [Google Scholar] [CrossRef]
  32. Zhang, X.; Su, X.; Yuan, Q.; Wang, Q. Spatial–Temporal Gray-Level Co-Occurrence Aware CNN for SAR Image Change Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4018605. [Google Scholar] [CrossRef]
  33. Wang, R.; Wang, L.; Wei, X.; Chen, J.W.; Jiao, L. Dynamic Graph-Level Neural Network for SAR Image Change Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4501005. [Google Scholar] [CrossRef]
  34. Mignotte, M. A Fractal Projection and Markovian Segmentation-Based Approach for Multimodal Change Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8046–8058. [Google Scholar] [CrossRef]
  35. Sun, Y.; Lei, L.; Guan, D.; Wu, J.; Kuang, G.; Liu, L. Image Regression with Structure Cycle Consistency for Heterogeneous Change Detection. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–15. [Google Scholar] [CrossRef]
  36. Sun, Y.; Lei, L.; Tan, X.; Guan, D.; Wu, J.; Kuang, G. Structured graph based image regression for unsupervised multimodal change detection. ISPRS J. Photogramm. Remote Sens. 2022, 185, 16–31. [Google Scholar] [CrossRef]
  37. Sun, Y.; Lei, L.; Guan, D.; Li, M.; Kuang, G. Sparse-Constrained Adaptive Structure Consistency-Based Unsupervised Image Regression for Heterogeneous Remote-Sensing Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4405814. [Google Scholar] [CrossRef]
  38. Sun, Y.; Lei, L.; Guan, D.; Kuang, G. Iterative Robust Graph for Unsupervised Change Detection of Heterogeneous Remote Sensing Images. IEEE Trans. Image Process. 2021, 30, 6277–6291. [Google Scholar] [CrossRef]
  39. Sun, Y.; Lei, L.; Li, X.; Sun, H.; Kuang, G. Nonlocal patch similarity based heterogeneous remote sensing change detection. Pattern Recognit. 2021, 109, 107598. [Google Scholar] [CrossRef]
  40. Sun, Y.; Lei, L.; Guan, D.; Kuang, G.; Liu, L. Graph Signal Processing for Heterogeneous Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4415823. [Google Scholar] [CrossRef]
  41. Zheng, Y.; Zhang, X.; Hou, B.; Liu, G. Using Combined Difference Image and k -Means Clustering for SAR Image Change Detection. IEEE Geosci. Remote Sens. Lett. 2014, 11, 691–695. [Google Scholar] [CrossRef]
  42. Hou, B.; Wei, Q.; Zheng, Y.; Wang, S. Unsupervised Change Detection in SAR Image Based on Gauss–Log Ratio Image Fusion and Compressed Projection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 3297–3317. [Google Scholar] [CrossRef]
  43. Krähenbühl, P.; Koltun, V. Efficient inference in fully connected crfs with gaussian edge potentials. In Proceedings of the 21st Annual Conference on Neural Information Processing Systems, Granada, Spain, 12–17 December 2011. [Google Scholar]
  44. Zhang, X.; Su, H.; Zhang, C.; Gu, X.; Tan, X.; Atkinson, P.M. Robust unsupervised small area change detection from SAR imagery using deep learning. ISPRS J. Photogramm. Remote Sens. 2021, 173, 79–94. [Google Scholar] [CrossRef]
  45. Jimenez-Sierra, D.A.; Quintero-Olaya, D.A.; Alvear-Muñoz, J.C.; Benítez-Restrepo, H.D.; Florez-Ospina, J.F.; Chanussot, J. Graph Learning Based on Signal Smoothness Representation for Homogeneous and Heterogeneous Change Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4410416. [Google Scholar] [CrossRef]
  46. Ban, Z.; Liu, J.; Cao, L. Superpixel Segmentation Using Gaussian Mixture Model. IEEE Trans. Image Process. 2018, 27, 4105–4117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Zheng, X.; Guan, D.; Li, B.; Chen, Z.; Li, X. Change Smoothness-Based Signal Decomposition Method for Multimodal Change Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 2507605. [Google Scholar] [CrossRef]
  48. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  49. Kittler, J.; Illingworth, J. Minimum error thresholding. Pattern Recognit. 1986, 19, 41–47. [Google Scholar] [CrossRef]
  50. Hartigan, J.A.; Wong, M.A. Algorithm AS 136: A k-means clustering algorithm. J. R. Stat. Soc. Ser. C (Appl. Stat.) 1979, 28, 100–108. [Google Scholar] [CrossRef]
  51. Bezdek, J.C.; Ehrlich, R.; Full, W. FCM: The fuzzy c-means clustering algorithm. Comput. Geosci. 1984, 10, 191–203. [Google Scholar] [CrossRef]
  52. Sun, Y.; Lei, L.; Guan, D.; Wu, J.; Kuang, G. Iterative structure transformation and conditional random field based method for unsupervised multimodal change detection. Pattern Recognit. 2022, 131, 108845. [Google Scholar] [CrossRef]
  53. Gong, M.; Cao, Y.; Wu, Q. A Neighborhood-Based Ratio Approach for Change Detection in SAR Images. IEEE Geosci. Remote Sens. Lett. 2012, 9, 307–311. [Google Scholar] [CrossRef]
  54. Lei, L.; Sun, Y.; Kuang, G. Adaptive Local Structure Consistency-Based Heterogeneous Remote Sensing Change Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 8003905. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed GLGDE.
Figure 1. Framework of the proposed GLGDE.
Remotesensing 15 01194 g001
Figure 2. Datasets #1 (top row) and #2 (bottom row). From left to right are: (a) pre-event image; (b) post-event image; and (c) ground truth.
Figure 2. Datasets #1 (top row) and #2 (bottom row). From left to right are: (a) pre-event image; (b) post-event image; and (c) ground truth.
Remotesensing 15 01194 g002
Figure 3. Initial and enhanced DIs of Datasets #1 and #2. From top to bottom, they correspond to initial DIs of Dataset #1, enhancement DIs of Dataset #1, initial DIs of Dataset #2, and enhancement DIs of Dataset #2. From left to right are DIs generated using: (a1a4) Diff/E-Diff; (b1b4) LR/E-LR; (c1c4) MR/E-MR; (d1d4) NR/E-NR; (e1e4) SDCD/E-SDCD; and (f1f4) INLPG/E-INLPG.
Figure 3. Initial and enhanced DIs of Datasets #1 and #2. From top to bottom, they correspond to initial DIs of Dataset #1, enhancement DIs of Dataset #1, initial DIs of Dataset #2, and enhancement DIs of Dataset #2. From left to right are DIs generated using: (a1a4) Diff/E-Diff; (b1b4) LR/E-LR; (c1c4) MR/E-MR; (d1d4) NR/E-NR; (e1e4) SDCD/E-SDCD; and (f1f4) INLPG/E-INLPG.
Remotesensing 15 01194 g003
Figure 4. ROC and PR curves of Datasets #1 and #2. From left to right are: (a) ROC curves on Dataset #1; (b) PR curves on Dataset #1; (c) ROC curves on Dataset #2; and (d) PR curves on Dataset #2.
Figure 4. ROC and PR curves of Datasets #1 and #2. From left to right are: (a) ROC curves on Dataset #1; (b) PR curves on Dataset #1; (c) ROC curves on Dataset #2; and (d) PR curves on Dataset #2.
Remotesensing 15 01194 g004
Figure 5. CMs computed from initial and enhanced DIs of Datasets #1 and #2. From top to bottom, they correspond to initial CMs of Dataset #1, enhancement CMs of Dataset #1, initial CMs of Dataset #2, and enhancement CMs of Dataset #2. From left to right are CMs generated by: (a1a4) Diff/E-Diff; (b1b4) LR/E-LR; (c1c4) MR/E-MR; (d1d4) NR/E-NR; (e1e4) SDCD/E-SDCD; and (f1f4) INLPG/E-INLPG. In the CM, White: true positives (TP); Red: false positives (FP); Black: true negatives (TN); and Cyan: false negatives (FN).
Figure 5. CMs computed from initial and enhanced DIs of Datasets #1 and #2. From top to bottom, they correspond to initial CMs of Dataset #1, enhancement CMs of Dataset #1, initial CMs of Dataset #2, and enhancement CMs of Dataset #2. From left to right are CMs generated by: (a1a4) Diff/E-Diff; (b1b4) LR/E-LR; (c1c4) MR/E-MR; (d1d4) NR/E-NR; (e1e4) SDCD/E-SDCD; and (f1f4) INLPG/E-INLPG. In the CM, White: true positives (TP); Red: false positives (FP); Black: true negatives (TN); and Cyan: false negatives (FN).
Remotesensing 15 01194 g005
Figure 6. Datasets #3 (top row) and #4 (bottom row). From left to right are: (a) pre-event image; (b) post-event image; and (c) ground truth.
Figure 6. Datasets #3 (top row) and #4 (bottom row). From left to right are: (a) pre-event image; (b) post-event image; and (c) ground truth.
Remotesensing 15 01194 g006
Figure 7. Initial and enhanced DIs of Datasets #3 and #4. From top to bottom, they correspond to initial DIs of Dataset #3, enhancement DIs of Dataset #3, initial DIs of Dataset #4, and enhancement DIs of Dataset #4. From left to right are DIs generated using: (a1a4) CVA/E-CVA; (b1b4) MAD/E-MAD; (c1c4) IRMAD/E-IRMAD; (d1d4) DSFA/E-DSFA; (e1e4) DCVA/E-DCVA; and (f1f4) INLPG/E-INLPG.
Figure 7. Initial and enhanced DIs of Datasets #3 and #4. From top to bottom, they correspond to initial DIs of Dataset #3, enhancement DIs of Dataset #3, initial DIs of Dataset #4, and enhancement DIs of Dataset #4. From left to right are DIs generated using: (a1a4) CVA/E-CVA; (b1b4) MAD/E-MAD; (c1c4) IRMAD/E-IRMAD; (d1d4) DSFA/E-DSFA; (e1e4) DCVA/E-DCVA; and (f1f4) INLPG/E-INLPG.
Remotesensing 15 01194 g007
Figure 8. ROC and PR curves of Datasets #3 and #4. From left to right are: (a) ROC curves on Dataset #3; (b) PR curves on Dataset #3; (c) ROC curves on Dataset #4; (d) PR curves on Dataset #4.
Figure 8. ROC and PR curves of Datasets #3 and #4. From left to right are: (a) ROC curves on Dataset #3; (b) PR curves on Dataset #3; (c) ROC curves on Dataset #4; (d) PR curves on Dataset #4.
Remotesensing 15 01194 g008
Figure 9. CMs computed from initial and enhanced DIs of Datasets #3 and #4. From top to bottom, they correspond to initial CMs of Dataset #3, enhancement CMs of Dataset #3, initial CMs of Dataset #4, enhancement CMs of Dataset #4. From left to right are CMs generated by: (a1a4) CVA/E-CVA; (b1b4) MAD/E-MAD; (c1c4) IRMAD/E-IRMAD; (d1d4) DSFA/E-DSFA; (e1e4) DCVA/E-DCVA; and (f1f4) INLPG/E-INLPG.
Figure 9. CMs computed from initial and enhanced DIs of Datasets #3 and #4. From top to bottom, they correspond to initial CMs of Dataset #3, enhancement CMs of Dataset #3, initial CMs of Dataset #4, enhancement CMs of Dataset #4. From left to right are CMs generated by: (a1a4) CVA/E-CVA; (b1b4) MAD/E-MAD; (c1c4) IRMAD/E-IRMAD; (d1d4) DSFA/E-DSFA; (e1e4) DCVA/E-DCVA; and (f1f4) INLPG/E-INLPG.
Remotesensing 15 01194 g009
Figure 10. Datasets #5 (top row) and #6 (bottom row). From left to right are: (a) pre-event image; (b) post-event image; and (c) ground truth.
Figure 10. Datasets #5 (top row) and #6 (bottom row). From left to right are: (a) pre-event image; (b) post-event image; and (c) ground truth.
Remotesensing 15 01194 g010
Figure 11. Initial and enhanced DIs of Datasets #5 and #6. From top to bottom, they correspond to initial DIs of Dataset #5, enhancement DIs of Dataset #5, initial DIs of Dataset #6, and enhancement DIs of Dataset #6. From left to right are DIs generated using: (a1a4) CVA/E-CVA; (b1b4) MAD/E-MAD; (c1c4) IRMAD/E-IRMAD; (d1d4) DSFA/E-DSFA; (e1e4) DCVA/E-DCVA; and (f1f4) INLPG/E-INLPG.
Figure 11. Initial and enhanced DIs of Datasets #5 and #6. From top to bottom, they correspond to initial DIs of Dataset #5, enhancement DIs of Dataset #5, initial DIs of Dataset #6, and enhancement DIs of Dataset #6. From left to right are DIs generated using: (a1a4) CVA/E-CVA; (b1b4) MAD/E-MAD; (c1c4) IRMAD/E-IRMAD; (d1d4) DSFA/E-DSFA; (e1e4) DCVA/E-DCVA; and (f1f4) INLPG/E-INLPG.
Remotesensing 15 01194 g011
Figure 12. ROC and PR curves of Datasets #5 and #6. From left to right are: (a) ROC curves on Dataset #5; (b) PR curves on Dataset #5; (c) ROC curves on Dataset #6; and (d) PR curves on Dataset #6.
Figure 12. ROC and PR curves of Datasets #5 and #6. From left to right are: (a) ROC curves on Dataset #5; (b) PR curves on Dataset #5; (c) ROC curves on Dataset #6; and (d) PR curves on Dataset #6.
Remotesensing 15 01194 g012
Figure 13. CMs computed from initial and enhanced DIs of Datasets #5 and #6. From top to bottom, they correspond to initial CMs of Dataset #5, enhancement CMs of Dataset #5, initial CMs of Dataset #6, and enhancement CMs of Dataset #6. From left to right are CMs generated by: (a1a4) CVA/E-CVA; (b1b4) MAD/E-MAD; (c1c4) IRMAD/E-IRMAD; (d1d4) DSFA/E-DSFA; (e1e4) DCVA/E-DCVA; and (f1f4) INLPG/E-INLPG.
Figure 13. CMs computed from initial and enhanced DIs of Datasets #5 and #6. From top to bottom, they correspond to initial CMs of Dataset #5, enhancement CMs of Dataset #5, initial CMs of Dataset #6, and enhancement CMs of Dataset #6. From left to right are CMs generated by: (a1a4) CVA/E-CVA; (b1b4) MAD/E-MAD; (c1c4) IRMAD/E-IRMAD; (d1d4) DSFA/E-DSFA; (e1e4) DCVA/E-DCVA; and (f1f4) INLPG/E-INLPG.
Remotesensing 15 01194 g013
Figure 14. Sensitivity analysis of parameter  α  in GLGDE.
Figure 14. Sensitivity analysis of parameter  α  in GLGDE.
Remotesensing 15 01194 g014
Table 1. Description of the six datasets.
Table 1. Description of the six datasets.
DatasetSensor (or Source)Size (Pixels)DateLocationEvent (and Spatial Resolution)
#1Radarsat-2/Radarsat-2   289 × 257 × 1   ( 1 ) June 2008–June 2009Yellow River, ChinaFarmland irrigation (8 m)
#2Radarsat-2/Radarsat-2   291 × 306 × 1   ( 1 ) June 2008–June 2009Yellow River, ChinaFarmland irrigation (8 m)
#3Google Earth/Google Earth   500 × 500 × 3   ( 3 ) September 2012–March 2013Beijing, ChinaConstruction (1 m)
#4Google Earth/Google Earth   500 × 500 × 3   ( 3 ) September 2012–March 2013Beijing, ChinaConstruction (1 m)
#5Landsat-5/Google Earth   300 × 412 × 1   ( 3 ) September 1995–July 1996Sardinia, ItalyLake expansion (30 m)
#6Radarsat-2/Google Earth   593 × 921 × 1   ( 3 ) June 2008–September 2012Shuguang Village, ChinaBuilding construction (8 m)
Table 2. Quantitative measures of DIs and CMs on the Datasets #1 and #2. Avg.ipv represents the average improvement.
Table 2. Quantitative measures of DIs and CMs on the Datasets #1 and #2. Avg.ipv represents the average improvement.
MethodsDataset #1Dataset #2
AUR ↑AUP ↑Fa ↓Mr ↓Oa ↑Kc ↑AUR ↑AUP ↑Fa ↓Mr ↓Oa ↑Kc ↑
Diff0.6570.2480.3170.4530.6590.1670.8180.1940.3260.1580.6840.154
LR0.7640.4780.1850.4040.7750.3510.9160.5250.1430.1560.8560.352
MR0.9020.8050.2260.1430.7890.4700.9660.8440.2630.0380.7500.238
NR0.9050.7940.1660.1780.8320.5360.9780.8600.1140.0470.8900.460
SDCD0.9000.6010.2360.0980.7890.4840.9710.7880.1880.0330.8210.326
INLPG0.9780.9380.0080.2640.9460.7980.9900.9090.0130.1630.9780.809
E-Diff0.9590.8810.0210.2520.9370.7740.9860.9220.0050.1540.9860.869
E-LR0.9710.9110.0170.2280.9450.8020.9930.9430.0050.1720.9850.863
E-MR0.9730.9290.0150.1800.9550.8410.9900.9450.0040.1210.9890.898
E-NR0.9810.9380.0120.1830.9570.8470.9940.9570.0030.1350.9890.897
E-SDCD0.9760.9300.0230.1350.9570.8530.9920.9470.0060.1050.9880.893
E-INLPG0.9830.9450.0130.1820.9560.8450.9960.9620.0040.1470.9880.885
Avg.ipv0.1230.278−0.173−0.0630.1530.3590.0520.259−0.1700.0400.1570.494
Table 3. Quantitative measures of DIs and CMs on the Datasets #3 and #4. Avg.ipv represents the average improvement.
Table 3. Quantitative measures of DIs and CMs on the Datasets #3 and #4. Avg.ipv represents the average improvement.
MethodsDataset #3Dataset #4
AUR ↑AUP ↑Fa ↓Mr ↓Oa ↑Kc ↑AUR ↑AUP ↑Fa ↓Mr ↓Oa ↑Kc ↑
CVA0.7120.1600.1910.4740.7870.1850.7980.1270.1680.3320.8300.073
MAD0.8850.4200.1930.1980.8060.3120.8590.1920.3170.1660.6850.042
IRMAD0.9110.6870.0060.5690.9500.5490.8560.2240.3060.1790.6960.043
DSFA0.7690.1940.0810.6850.8710.2070.8240.1800.0820.3950.9130.139
DCVA0.7150.2120.3350.3290.6660.1270.9550.6140.1400.0950.8600.127
INLPG0.9550.6520.1240.0860.8790.4840.9920.7960.0070.2280.9900.674
E-CVA0.9750.6850.0760.0470.9260.6310.9780.7060.0930.0660.9070.195
E-MAD0.9930.9020.0330.0160.9690.8150.9820.8090.1370.0280.8650.142
E-IRMAD0.9950.9420.0220.0350.9770.8580.9800.7990.1400.0330.8610.137
E-DSFA0.9760.6800.0730.0180.9320.6580.9790.7020.0470.1600.9520.306
E-DCVA0.9750.7440.0930.0490.9110.5800.9940.8670.0500.0360.9500.328
E-INLPG0.9920.8980.0350.0120.9660.8040.9990.9490.0010.1940.9970.876
Avg.ipv0.1600.421−0.100−0.3610.1200.4140.1050.450−0.092−0.1460.0930.148
Table 4. Quantitative measures of DIs and CMs on the Datasets #5 and #6. Avg.ipv represents the average improvement.
Table 4. Quantitative measures of DIs and CMs on the Datasets #5 and #6. Avg.ipv represents the average improvement.
MethodsDataset #5Dataset #6
AUR ↑AUP ↑Fa ↓Mr ↓Oa ↑Kc ↑AUR ↑AUP ↑Fa ↓Mr ↓Oa ↑Kc ↑
HPT0.8890.3730.1960.1380.8080.2860.9220.5640.1170.1810.8800.339
AMDIR0.7950.1550.2240.2770.7730.2030.9110.4700.1510.1800.8480.279
ALSC0.9720.7930.0260.2180.9620.6960.9800.6230.0380.1150.9580.641
INLPG0.9300.6040.0270.2970.9560.6420.9850.8080.0290.1510.9650.672
FPMS0.9250.4060.0830.1890.9110.4860.9940.9040.0040.3100.9820.768
SCASC0.8850.3830.0430.4200.9340.4850.9680.6950.0020.7130.9650.418
E-HPT0.9480.7570.0350.1860.9560.6720.9920.9220.0190.0440.9790.800
E-AMDIR0.9260.5760.0920.1810.9030.4650.9910.8890.0600.0390.9410.571
E-ALSC0.9800.8510.0110.2290.9750.7810.9940.9110.0210.0400.9790.794
E-INLPG0.9590.7610.0170.2610.9680.7230.9940.9200.0200.0380.9800.802
E-FPMS0.9530.6690.0400.2210.9490.6260.9960.9480.0110.0390.9870.869
E-SCASC0.9640.7550.0180.2800.9660.7050.9920.8920.0210.0380.9780.792
Avg.ipv0.0560.276−0.064−0.0300.0620.1960.0330.236−0.032−0.2350.0410.252
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, X.; Guan, D.; Li, B.; Chen, Z.; Pan, L. Global and Local Graph-Based Difference Image Enhancement for Change Detection. Remote Sens. 2023, 15, 1194. https://doi.org/10.3390/rs15051194

AMA Style

Zheng X, Guan D, Li B, Chen Z, Pan L. Global and Local Graph-Based Difference Image Enhancement for Change Detection. Remote Sensing. 2023; 15(5):1194. https://doi.org/10.3390/rs15051194

Chicago/Turabian Style

Zheng, Xiaolong, Dongdong Guan, Bangjie Li, Zhengsheng Chen, and Lefei Pan. 2023. "Global and Local Graph-Based Difference Image Enhancement for Change Detection" Remote Sensing 15, no. 5: 1194. https://doi.org/10.3390/rs15051194

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop