Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (29)

Search Parameters:
Keywords = bitemporal image transformer

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 10556 KiB  
Article
HSAA-CD: A Hierarchical Semantic Aggregation Mechanism and Attention Module for Non-Agricultural Change Detection in Cultivated Land
by Fangting Li, Fangdong Zhou, Guo Zhang, Jianfeng Xiao and Peng Zeng
Remote Sens. 2024, 16(8), 1372; https://doi.org/10.3390/rs16081372 - 13 Apr 2024
Viewed by 721
Abstract
Cultivated land plays a fundamental role in the sustainable development of the world. Monitoring the non-agricultural changes is important for the development of land-use policies. A bitemporal image transformer (BIT) can achieve high accuracy for change detection (CD) tasks and also become a [...] Read more.
Cultivated land plays a fundamental role in the sustainable development of the world. Monitoring the non-agricultural changes is important for the development of land-use policies. A bitemporal image transformer (BIT) can achieve high accuracy for change detection (CD) tasks and also become a key scientific tool to support decision-making. Because of the diversity of high-resolution RSIs in series, the complexity of agricultural types, and the irregularity of hierarchical semantics in different types of changes, the accuracy of non-agricultural CD is far below the need for the management of the land and for resource planning. In this paper, we proposed a novel non-agricultural CD method to improve the accuracy of machine processing. First, multi-resource surveying data are collected to produce a well-tagged dataset with cultivated land and non-agricultural changes. Secondly, a hierarchical semantic aggregation mechanism and attention module (HSAA) bitemporal image transformer method named HSAA-CD is performed for non-agricultural CD in cultivated land. The proposed HSAA-CD added a hierarchical semantic aggregation mechanism for clustering the input data for U-Net as the backbone network and an attention module to improve the feature edge. Experiments were performed on the open-source LEVIR-CD and WHU Building-CD datasets as well as on the self-built RSI dataset. The F1-score, intersection over union (IoU), and overall accuracy (OA) of these three datasets were 88.56%, 84.29%, and 68.50%; 79.84%, 73.41%, and 59.29%; and 98.83%, 98.39%, and 93.56%, respectively. The results indicated that the proposed HSAA-CD method outperformed the BIT and some other state-of-the-art methods and proved to be suitable accuracy for non-agricultural CD in cultivated land. Full article
Show Figures

Graphical abstract

19 pages, 1404 KiB  
Article
MFINet: Multi-Scale Feature Interaction Network for Change Detection of High-Resolution Remote Sensing Images
by Wuxu Ren, Zhongchen Wang, Min Xia and Haifeng Lin
Remote Sens. 2024, 16(7), 1269; https://doi.org/10.3390/rs16071269 - 4 Apr 2024
Cited by 8 | Viewed by 1297
Abstract
Change detection is widely used in the field of building monitoring. In recent years, the progress of remote sensing image technology has provided high-resolution data. However, unlike other tasks, change detection focuses on the difference between dual-input images, so the interaction between bi-temporal [...] Read more.
Change detection is widely used in the field of building monitoring. In recent years, the progress of remote sensing image technology has provided high-resolution data. However, unlike other tasks, change detection focuses on the difference between dual-input images, so the interaction between bi-temporal features is crucial. However, the existing methods have not fully tapped the potential of multi-scale bi-temporal features to interact layer by layer. Therefore, this paper proposes a multi-scale feature interaction network (MFINet). The network realizes the information interaction of multi-temporal images by inserting a bi-temporal feature interaction layer (BFIL) between backbone networks at the same level, guides the attention to focus on the difference region, and suppresses the interference. At the same time, a double temporal feature fusion layer (BFFL) is used at the end of the coding layer to extract subtle difference features. By introducing the transformer decoding layer and improving the recovery effect of the feature size, the ability of the network to accurately capture the details and contour information of the building is further improved. The F1 of our model on the public dataset LEVIR-CD reaches 90.12%, which shows better accuracy and generalization performance than many state-of-the-art change detection models. Full article
Show Figures

Figure 1

30 pages, 73805 KiB  
Article
DTT-CGINet: A Dual Temporal Transformer Network with Multi-Scale Contour-Guided Graph Interaction for Change Detection
by Ming Chen, Wanshou Jiang and Yuan Zhou
Remote Sens. 2024, 16(5), 844; https://doi.org/10.3390/rs16050844 - 28 Feb 2024
Cited by 1 | Viewed by 1053
Abstract
Deep learning has dramatically enhanced remote sensing change detection. However, existing neural network models often face challenges like false positives and missed detections due to factors like lighting changes, scale differences, and noise interruptions. Additionally, change detection results often fail to capture target [...] Read more.
Deep learning has dramatically enhanced remote sensing change detection. However, existing neural network models often face challenges like false positives and missed detections due to factors like lighting changes, scale differences, and noise interruptions. Additionally, change detection results often fail to capture target contours accurately. To address these issues, we propose a novel transformer-based hybrid network. In this study, we analyze the structural relationship in bi-temporal images and introduce a cross-attention-based transformer to model this relationship. First, we use a tokenizer to express the high-level features of the bi-temporal image into several semantic tokens. Then, we use a dual temporal transformer (DTT) encoder to capture dense spatiotemporal contextual relationships among the tokens. The features extracted at the coarse scale are refined into finer details through the DTT decoder. Concurrently, we input the backbone’s low-level features into a contour-guided graph interaction module (CGIM) that utilizes joint attention to capture semantic relationships between object regions and the contour. Then, we use the feature pyramid decoder to integrate the multi-scale outputs of the CGIM. The convolutional block attention modules (CBAMs) employ channel and spatial attention to reweight feature maps. Finally, the classifier discriminates change pixels and generates the final change map of the difference feature map. Several experiments have demonstrated that our model shows significant advantages over other methods in terms of efficiency, accuracy, and visual effects. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Graphical abstract

18 pages, 4775 KiB  
Article
A Hyperspectral Change Detection (HCD-Net) Framework Based on Double Stream Convolutional Neural Networks and an Attention Module
by Seyd Teymoor Seydi, Mahboubeh Boueshagh, Foad Namjoo, Seyed Mohammad Minouei, Zahir Nikraftar and Meisam Amani
Remote Sens. 2024, 16(5), 827; https://doi.org/10.3390/rs16050827 - 28 Feb 2024
Cited by 1 | Viewed by 1316
Abstract
Human activities and natural phenomena continually transform the Earth’s surface, presenting ongoing challenges to the environment. Therefore, the accurate and timely monitoring and prediction of these alterations are essential for devising effective solutions and mitigating environmental impacts in advance. This study introduces a [...] Read more.
Human activities and natural phenomena continually transform the Earth’s surface, presenting ongoing challenges to the environment. Therefore, the accurate and timely monitoring and prediction of these alterations are essential for devising effective solutions and mitigating environmental impacts in advance. This study introduces a novel framework, called HCD-Net, for detecting changes using bi-temporal hyperspectral images. HCD-Net is built upon a dual-stream deep feature extraction process, complemented by an attention mechanism. The first stream employs 3D convolution layers and 3D Squeeze-and-Excitation (SE) blocks to extract deep features, while the second stream utilizes 2D convolution and 2D SE blocks for the same purpose. The deep features from both streams are then concatenated and processed through dense layers for decision-making. The performance of HCD-Net is evaluated against existing state-of-the-art change detection methods. For this purpose, the bi-temporal Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) hyperspectral dataset was utilized to assess the change detection performance. The findings indicate that HCD-Net achieves superior accuracy and the lowest false alarm rate among the compared methods, with an overall classification accuracy exceeding 96%, and a kappa coefficient greater than 0.9. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Graphical abstract

20 pages, 49541 KiB  
Article
Assessing Sentinel-2, Sentinel-1, and ALOS-2 PALSAR-2 Data for Large-Scale Wildfire-Burned Area Mapping: Insights from the 2017–2019 Canada Wildfires
by Puzhao Zhang, Xikun Hu, Yifang Ban, Andrea Nascetti and Maoguo Gong
Remote Sens. 2024, 16(3), 556; https://doi.org/10.3390/rs16030556 - 31 Jan 2024
Cited by 3 | Viewed by 2041
Abstract
Wildfires play a crucial role in the transformation of forest ecosystems and exert a significant influence on the global climate over geological timescales. Recent shifts in climate patterns and intensified human–forest interactions have led to an increase in the incidence of wildfires. These [...] Read more.
Wildfires play a crucial role in the transformation of forest ecosystems and exert a significant influence on the global climate over geological timescales. Recent shifts in climate patterns and intensified human–forest interactions have led to an increase in the incidence of wildfires. These fires are characterized by their extensive coverage, higher frequency, and prolonged duration, rendering them increasingly destructive. To mitigate the impact of wildfires on climate change, ecosystems, and biodiversity, it is imperative to conduct systematic monitoring of wildfire progression and evaluate their environmental repercussions on a global scale. Satellite remote sensing is a powerful tool, offering precise and timely data on terrestrial changes, and has been extensively utilized for wildfire identification, tracking, and impact assessment at both local and regional levels. The Canada Centre for Mapping and Earth Observation, in collaboration with the Canadian Forest Service, has developed a comprehensive National Burned Area Composite (NBAC). This composite serves as a benchmark for curating a bi-temporal multi-source satellite image dataset for change detection, compiled from the archives of Sentinel-2, Sentinel-1, and ALOS-2 PALSAR-2. To our knowledge, this dataset is the inaugural large-scale, multi-source, and multi-frequency satellite image dataset with 20 m spatial resolution for wildfire mapping, monitoring, and evaluation. It harbors significant potential for enhancing wildfire management strategies, building upon the profound advancements in deep learning that have contributed to the field of remote sensing. Based on our curated dataset, which encompasses major wildfire events in Canada, we conducted a systematic evaluation of the capability of multi-source satellite earth observation data in identifying wildfire-burned areas using statistical analysis and deep learning. Our analysis compares the difference between burned and unburned areas using post-event observation solely or bi-temporal (pre- and post-event) observations across diverse land cover types. We demonstrate that optical satellite data yield higher separability than C-Band and L-Band Synthetic Aperture Radar (SAR), which exhibit considerable overlap in burned and unburned sample distribution, as evidenced by SAR-based boxplots. With U-Net, we further explore how different input channels influence the detection accuracy. Our findings reveal that deep neural networks enhance SAR’s performance in mapping burned areas. Notably, C-Band SAR shows a higher dependency on pre-event data than L-Band SAR for effective detection. A comparative analysis of U-Net and its variants indicates that U-Net works best with single-sensor data, while the late fusion architecture marginally surpasses others in the fusion of optical and SAR data. Accuracy across sensors is highest in closed forests, with sequentially lower performance in open forests, shrubs, and grasslands. Future work will extend the data from both spatial and temporal dimensions to encompass varied vegetation types and climate zones, furthering our understanding of multi-source and multi-frequency satellite remote sensing capabilities in wildfire detection and monitoring. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

21 pages, 2913 KiB  
Article
Interactive Change-Aware Transformer Network for Remote Sensing Image Change Captioning
by Chen Cai, Yi Wang and Kim-Hui Yap
Remote Sens. 2023, 15(23), 5611; https://doi.org/10.3390/rs15235611 - 3 Dec 2023
Cited by 2 | Viewed by 1507
Abstract
Remote sensing image change captioning (RSICC) aims to automatically generate sentences describing the difference in content in remote sensing bitemporal images. Recent works extract the changes between bitemporal features and employ a hierarchical approach to fuse multiple changes of interest, yielding change captions. [...] Read more.
Remote sensing image change captioning (RSICC) aims to automatically generate sentences describing the difference in content in remote sensing bitemporal images. Recent works extract the changes between bitemporal features and employ a hierarchical approach to fuse multiple changes of interest, yielding change captions. However, these methods directly aggregate all features, potentially incorporating non-change-focused information from each encoder layer into the change caption decoder, adversely affecting the performance of change captioning. To address this problem, we proposed an Interactive Change-Aware Transformer Network (ICT-Net). ICT-Net is able to extract and incorporate the most critical changes of interest in each encoder layer to improve change description generation. It initially extracts bitemporal visual features from the CNN backbone and employs an Interactive Change-Aware Encoder (ICE) to capture the crucial difference between these features. Specifically, the ICE captures the most change-aware discriminative information between the paired bitemporal features interactively through difference and content attention encoding. A Multi-Layer Adaptive Fusion (MAF) module is proposed to adaptively aggregate the relevant change-aware features in the ICE layers while minimizing the impact of irrelevant visual features. Moreover, we extend the ICE to extract multi-scale changes and introduce a novel Cross Gated-Attention (CGA) module into the change caption decoder to select essential discriminative multi-scale features to improve the change captioning performance. We evaluate our method on two RSICC datasets (e.g., LEVIR-CC and LEVIRCCD), and the experimental results demonstrate that our method achieves a state-of-the-art performance. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

27 pages, 26208 KiB  
Article
A Multi-Task Consistency Enhancement Network for Semantic Change Detection in HR Remote Sensing Images and Application of Non-Agriculturalization
by Haihan Lin, Xiaoqin Wang, Mengmeng Li, Dehua Huang and Ruijiao Wu
Remote Sens. 2023, 15(21), 5106; https://doi.org/10.3390/rs15215106 - 25 Oct 2023
Cited by 1 | Viewed by 1349
Abstract
It is challenging to investigate semantic change detection (SCD) in bi-temporal high-resolution (HR) remote sensing images. For the non-changing surfaces in the same location of bi-temporal images, existing SCD methods often obtain the results with frequent errors or incomplete change detection due to [...] Read more.
It is challenging to investigate semantic change detection (SCD) in bi-temporal high-resolution (HR) remote sensing images. For the non-changing surfaces in the same location of bi-temporal images, existing SCD methods often obtain the results with frequent errors or incomplete change detection due to insufficient performance on overcoming the phenomenon of intraclass differences. To address the above-mentioned issues, we propose a novel multi-task consistency enhancement network (MCENet) for SCD. Specifically, a multi-task learning-based network is constructed by combining CNN and Transformer as the backbone. Moreover, a multi-task consistency enhancement module (MCEM) is introduced, and cross-task mapping connections are selected as auxiliary designs in the network to enhance the learning of semantic consistency in non-changing regions and the integrity of change features. Furthermore, we establish a novel joint loss function to alleviate the negative effect of class imbalances in quantity during network training optimization. We performed experiments on publicly available SCD datasets, including the SECOND and HRSCD datasets. MCENet achieved promising results, with a 22.06% Sek and a 37.41% Score on the SECOND dataset and a 14.87% Sek and a 30.61% Score on the HRSCD dataset. Moreover, we evaluated the applicability of MCENet on the NAFZ dataset that was employed for cropland change detection and non-agricultural identification, with a 21.67% Sek and a 37.28% Score. The relevant comparative and ablation experiments suggested that MCENet possesses superior performance and effectiveness in network design. Full article
Show Figures

Graphical abstract

20 pages, 8917 KiB  
Article
TTNet: A Temporal-Transform Network for Semantic Change Detection Based on Bi-Temporal Remote Sensing Images
by Liangcun Jiang, Feng Li, Li Huang, Feifei Peng and Lei Hu
Remote Sens. 2023, 15(18), 4555; https://doi.org/10.3390/rs15184555 - 15 Sep 2023
Cited by 2 | Viewed by 1465
Abstract
Semantic change detection (SCD) holds a critical place in remote sensing image interpretation, as it aims to locate changing regions and identify their associated land cover classes. Presently, post-classification techniques stand as the predominant strategy for SCD due to their simplicity and efficacy. [...] Read more.
Semantic change detection (SCD) holds a critical place in remote sensing image interpretation, as it aims to locate changing regions and identify their associated land cover classes. Presently, post-classification techniques stand as the predominant strategy for SCD due to their simplicity and efficacy. However, these methods often overlook the intricate relationships between alterations in land cover. In this paper, we argue that comprehending the interplay of changes within land cover maps holds the key to enhancing SCD’s performance. With this insight, a Temporal-Transform Module (TTM) is designed to capture change relationships across temporal dimensions. TTM selectively aggregates features across all temporal images, enhancing the unique features of each temporal image at distinct pixels. Moreover, we build a Temporal-Transform Network (TTNet) for SCD, comprising two semantic segmentation branches and a binary change detection branch. TTM is embedded into the decoder of each semantic segmentation branch, thus enabling TTNet to obtain better land cover classification results. Experimental results on the SECOND dataset show that TTNet achieves enhanced performance when compared to other benchmark methods in the SCD task. In particular, TTNet elevates mIoU accuracy by a minimum of 1.5% in the SCD task and 3.1% in the semantic segmentation task. Full article
Show Figures

Figure 1

20 pages, 4986 KiB  
Article
A VHR Bi-Temporal Remote-Sensing Image Change Detection Network Based on Swin Transformer
by Yunhe Teng, Shuo Liu, Weichao Sun, Huan Yang, Bin Wang and Jintong Jia
Remote Sens. 2023, 15(10), 2645; https://doi.org/10.3390/rs15102645 - 19 May 2023
Cited by 4 | Viewed by 1889
Abstract
Change detection (CD), as a special remote-sensing (RS) segmentation task, faces challenges, including alignment errors and illumination variation, dense small targets, and large background intraclass variance in very high-resolution (VHR) remote-sensing images. Recent methods have avoided the misjudgment caused by illumination variation and [...] Read more.
Change detection (CD), as a special remote-sensing (RS) segmentation task, faces challenges, including alignment errors and illumination variation, dense small targets, and large background intraclass variance in very high-resolution (VHR) remote-sensing images. Recent methods have avoided the misjudgment caused by illumination variation and alignment errors by increasing the ability of global modeling, but the latter two problems have still not been fully addressed. In this paper, we propose a new CD model called SFCD, which increases the feature extraction capability for small targets by introducing a shifted-window (Swin) transformer. We designed a foreground-aware fusion module to use attention gates to trim low-level feature responses, enabling increased attention to the changed region compared to the background when recovering the changed region, thus reducing background interference. We evaluated our model on two CD datasets, LEVIR-CD and CDD, and obtained F1 scores of 91.78 and 97.87, respectively. The experimental results and visual interpretation show that our model outperforms several previous CD models. In addition, we adjusted the parameters and structure of the standard model to develop a lightweight version that achieves an accuracy beyond most models with only 1.55 M parameters, further validating the effectiveness of our design. Full article
Show Figures

Figure 1

26 pages, 2303 KiB  
Article
A CNN-Transformer Network Combining CBAM for Change Detection in High-Resolution Remote Sensing Images
by Mengmeng Yin, Zhibo Chen and Chengjian Zhang
Remote Sens. 2023, 15(9), 2406; https://doi.org/10.3390/rs15092406 - 4 May 2023
Cited by 10 | Viewed by 3769
Abstract
Current deep learning-based change detection approaches mostly produce convincing results by introducing attention mechanisms to traditional convolutional networks. However, given the limitation of the receptive field, convolution-based methods fall short of fully modelling global context and capturing long-range dependencies, thus insufficient in discriminating [...] Read more.
Current deep learning-based change detection approaches mostly produce convincing results by introducing attention mechanisms to traditional convolutional networks. However, given the limitation of the receptive field, convolution-based methods fall short of fully modelling global context and capturing long-range dependencies, thus insufficient in discriminating pseudo changes. Transformers have an efficient global spatio-temporal modelling capability, which is beneficial for the feature representation of changes of interest. However, the lack of detailed information may cause the transformer to locate the boundaries of changed regions inaccurately. Therefore, in this article, a hybrid CNN-transformer architecture named CTCANet, combining the strengths of convolutional networks, transformer, and attention mechanisms, is proposed for high-resolution bi-temporal remote sensing image change detection. To obtain high-level feature representations that reveal changes of interest, CTCANet utilizes tokenizer to embed the features of each image extracted by convolutional network into a sequence of tokens, and the transformer module to model global spatio-temporal context in token space. The optimal bi-temporal information fusion approach is explored here. Subsequently, the reconstructed features carrying deep abstract information are fed to the cascaded decoder to aggregate with features containing shallow fine-grained information, through skip connections. Such an aggregation empowers our model to maintain the completeness of changes and accurately locate small targets. Moreover, the integration of the convolutional block attention module enables the smoothing of semantic gaps between heterogeneous features and the accentuation of relevant changes in both the channel and spatial domains, resulting in more impressive outcomes. The performance of the proposed CTCANet surpasses that of recent certain state-of-the-art methods, as evidenced by experimental results on two publicly accessible datasets, LEVIR-CD and SYSU-CD. Full article
Show Figures

Figure 1

30 pages, 47919 KiB  
Article
DCAT: Dual Cross-Attention-Based Transformer for Change Detection
by Yuan Zhou, Chunlei Huo, Jiahang Zhu, Leigang Huo and Chunhong Pan
Remote Sens. 2023, 15(9), 2395; https://doi.org/10.3390/rs15092395 - 3 May 2023
Cited by 8 | Viewed by 3748
Abstract
Several transformer-based methods for change detection (CD) in remote sensing images have been proposed, with Siamese-based methods showing promising results due to their two-stream feature extraction structure. However, these methods ignore the potential of the cross-attention mechanism to improve change feature discrimination and [...] Read more.
Several transformer-based methods for change detection (CD) in remote sensing images have been proposed, with Siamese-based methods showing promising results due to their two-stream feature extraction structure. However, these methods ignore the potential of the cross-attention mechanism to improve change feature discrimination and thus, may limit the final performance. Additionally, using either high-frequency-like fast change or low-frequency-like slow change alone may not effectively represent complex bi-temporal features. Given these limitations, we have developed a new approach that utilizes the dual cross-attention-transformer (DCAT) method. This method mimics the visual change observation procedure of human beings and interacts with and merges bi-temporal features. Unlike traditional Siamese-based CD frameworks, the proposed method extracts multi-scale features and models patch-wise change relationships by connecting a series of hierarchically structured dual cross-attention blocks (DCAB). DCAB is based on a hybrid dual branch mixer that combines convolution and transformer to extract and fuse local and global features. It calculates two types of cross-attention features to effectively learn comprehensive cues with both low- and high-frequency information input from paired CD images. This helps enhance discrimination between the changed and unchanged regions during feature extraction. The feature pyramid fusion network is more lightweight than the encoder and produces powerful multi-scale change representations by aggregating features from different layers. Experiments on four CD datasets demonstrate the advantages of DCAT architecture over other state-of-the-art methods. Full article
(This article belongs to the Special Issue Pattern Recognition in Remote Sensing)
Show Figures

Figure 1

21 pages, 7190 KiB  
Article
PBFormer: Point and Bi-Spatiotemporal Transformer for Pointwise Change Detection of 3D Urban Point Clouds
by Ming Han, Jianjun Sha, Yanheng Wang and Xiangwei Wang
Remote Sens. 2023, 15(9), 2314; https://doi.org/10.3390/rs15092314 - 27 Apr 2023
Cited by 3 | Viewed by 1506
Abstract
Change detection (CD) is a technique widely used in remote sensing for identifying the differences between data acquired at different times. Most existing 3D CD approaches voxelize point clouds into 3D grids, project them into 2D images, or rasterize them into digital surface [...] Read more.
Change detection (CD) is a technique widely used in remote sensing for identifying the differences between data acquired at different times. Most existing 3D CD approaches voxelize point clouds into 3D grids, project them into 2D images, or rasterize them into digital surface models due to the irregular format of point clouds and the variety of changes in three-dimensional (3D) objects. However, the details of the geometric structure and spatiotemporal sequence information may not be fully utilized. In this article, we propose PBFormer, a transformer network with Siamese architecture, for directly inferring pointwise changes in bi-temporal 3D point clouds. First, we extract point sequences from irregular 3D point clouds using the k-nearest neighbor method. Second, we uniquely use a point transformer network as an encoder to extract point feature information from bitemporal 3D point clouds. Then, we design a module for fusing the spatiotemporal features of bi-temporal point clouds to effectively detect change features. Finally, multilayer perceptrons are used to obtain the CD results. Extensive experiments conducted on the Urb3DCD benchmark show that PBFormer outperforms other excellent approaches for 3D point cloud CD tasks. Full article
Show Figures

Figure 1

13 pages, 3126 KiB  
Article
CA-BIT: A Change Detection Method of Land Use in Natural Reserves
by Bin Jia, Zhiyou Cheng, Chuanjian Wang, Jinling Zhao and Ning An
Agronomy 2023, 13(3), 635; https://doi.org/10.3390/agronomy13030635 - 23 Feb 2023
Cited by 2 | Viewed by 1512
Abstract
Natural reserves play a leading role in safeguarding national ecological security. Remote sensing change detection (CD) technology can identify the dynamic changes of land use and warn of ecological risks in natural reserves in a timely manner, which can provide technical support for [...] Read more.
Natural reserves play a leading role in safeguarding national ecological security. Remote sensing change detection (CD) technology can identify the dynamic changes of land use and warn of ecological risks in natural reserves in a timely manner, which can provide technical support for the management of natural reserves. We propose a CD method (CA-BIT) based on the improved bitemporal image transformer (BIT) model to realize the change detection of remote sensing data of Anhui Natural Reserves in 2018 and 2021. Resnet34-CA is constructed through the combination of Resnet34 and a coordinate attention mechanism to effectively extract high-level semantic features. The BIT module is also used to efficiently enhance the original semantic features. Compared with the overall accuracy of the existing deep learning-based CD methods, that of CA-BIT is 98.34% on the natural protected area CD datasets and 99.05% on LEVIR_CD. Our method can effectively satisfy the need of CD of different land categories such as construction land, farmland, and forest land. Full article
Show Figures

Figure 1

19 pages, 6868 KiB  
Article
SMNet: Symmetric Multi-Task Network for Semantic Change Detection in Remote Sensing Images Based on CNN and Transformer
by Yiting Niu, Haitao Guo, Jun Lu, Lei Ding and Donghang Yu
Remote Sens. 2023, 15(4), 949; https://doi.org/10.3390/rs15040949 - 9 Feb 2023
Cited by 18 | Viewed by 3065 | Correction
Abstract
Deep learning has achieved great success in remote sensing image change detection (CD). However, most methods focus only on the changed regions of images and cannot accurately identify their detailed semantic categories. In addition, most CD methods using convolutional neural networks (CNN) have [...] Read more.
Deep learning has achieved great success in remote sensing image change detection (CD). However, most methods focus only on the changed regions of images and cannot accurately identify their detailed semantic categories. In addition, most CD methods using convolutional neural networks (CNN) have difficulty capturing sufficient global information from images. To address the above issues, we propose a novel symmetric multi-task network (SMNet) that integrates global and local information for semantic change detection (SCD) in this paper. Specifically, we employ a hybrid unit consisting of pre-activated residual blocks (PR) and transformation blocks (TB) to construct the (PRTB) backbone, which obtains more abundant semantic features with local and global information from bi-temporal images. To accurately capture fine-grained changes, the multi-content fusion module (MCFM) is introduced, which effectively enhances change features by distinguishing foreground and background information in complex scenes. In the meantime, the multi-task prediction branches are adopted, and the multi-task loss function is used to jointly supervise model training to improve the performance of the network. Extensive experimental results on the challenging SECOND and Landsat-SCD datasets, demonstrate that our SMNet obtains 71.95% and 85.65% at mean Intersection over Union (mIoU), respectively. In addition, the proposed SMNet achieves 20.29% and 51.14% at Separated Kappa coefficient (Sek) on the SECOND and Landsat-SCD datasets, respectively. All of the above proves the effectiveness and superiority of the proposed method. Full article
(This article belongs to the Special Issue Recent Progress of Change Detection Based on Remote Sensing)
Show Figures

Graphical abstract

24 pages, 10438 KiB  
Article
A Dual Neighborhood Hypergraph Neural Network for Change Detection in VHR Remote Sensing Images
by Junzheng Wu, Ruigang Fu, Qiang Liu, Weiping Ni, Kenan Cheng, Biao Li and Yuli Sun
Remote Sens. 2023, 15(3), 694; https://doi.org/10.3390/rs15030694 - 24 Jan 2023
Cited by 5 | Viewed by 2130
Abstract
The very high spatial resolution (VHR) remote sensing images have been an extremely valuable source for monitoring changes occurring on the Earth’s surface. However, precisely detecting relevant changes in VHR images still remains a challenge, due to the complexity of the relationships among [...] Read more.
The very high spatial resolution (VHR) remote sensing images have been an extremely valuable source for monitoring changes occurring on the Earth’s surface. However, precisely detecting relevant changes in VHR images still remains a challenge, due to the complexity of the relationships among ground objects. To address this limitation, a dual neighborhood hypergraph neural network is proposed in this article, which combines multiscale superpixel segmentation and hypergraph convolution to model and exploit the complex relationships. First, the bi-temporal image pairs are segmented under two scales and fed to a pre-trained U-net to obtain node features by treating each object under the fine scale as a node. The dual neighborhood is then defined using the father-child and adjacent relationships of the segmented objects to construct the hypergraph, which permits models to represent higher-order structured information far more complex than the conventional pairwise relationships. The hypergraph convolutions are conducted on the constructed hypergraph to propagate the label information from a small amount of labeled nodes to the other unlabeled ones by the node-edge-node transformation. Moreover, to alleviate the problem of imbalanced sampling, the focal loss function is adopted to train the hypergraph neural network. The experimental results on optical, SAR and heterogeneous optical/SAR data sets demonstrate that the proposed method offersbetter effectiveness and robustness compared to many state-of-the-art methods. Full article
Show Figures

Figure 1

Back to TopTop