Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (10,042)

Search Parameters:
Keywords = remote-sensing imaging

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 3215 KiB  
Article
Flexible Natural Language-Based Image Data Downlink Prioritization for Nanosatellites
by Ezra Fielding and Akitoshi Hanazawa
Aerospace 2024, 11(11), 888; https://doi.org/10.3390/aerospace11110888 (registering DOI) - 28 Oct 2024
Abstract
Nanosatellites increasingly produce more data than can be downlinked within a reasonable time due to their limited bandwidth and power. Therefore, an on-board system is required to prioritize scientifically significant data for downlinking, as described by scientists. This paper determines whether natural language [...] Read more.
Nanosatellites increasingly produce more data than can be downlinked within a reasonable time due to their limited bandwidth and power. Therefore, an on-board system is required to prioritize scientifically significant data for downlinking, as described by scientists. This paper determines whether natural language processing can be used to prioritize remote sensing images on CubeSats with more flexibility compared to existing methods. Two approaches implementing the same conceptual prioritization pipeline are compared. The first uses YOLOv8 and Llama2 to extract image features and compare them with text descriptions via cosine similarity. The second approach employs CLIP, fine-tuned on remote sensing data, to achieve the same. Both approaches are evaluated on real nanosatellite hardware, the VERTECS Camera Control Board. The CLIP approach, particularly the ResNet50-based model, shows the best performance in prioritizing and sequencing remote sensing images. This paper demonstrates that on-orbit prioritization using natural language descriptions is viable and allows for more flexibility than existing methods. Full article
(This article belongs to the Special Issue Small Satellite Missions)
Show Figures

Figure 1

18 pages, 5084 KiB  
Article
Activation of Ms 6.9 Milin Earthquake on Sedongpu Disaster Chain, China with Multi-Temporal Optical Images
by Yubin Xin, Chaoying Zhao, Bin Li, Xiaojie Liu, Yang Gao and Jianqi Lou
Remote Sens. 2024, 16(21), 4003; https://doi.org/10.3390/rs16214003 (registering DOI) - 28 Oct 2024
Abstract
In recent years, disaster chains caused by glacier movements have occurred frequently in the lower Yarlung Tsangpo River in southwest China. However, it is still unclear whether earthquakes significantly contribute to glacier movements and disaster chains. In addition, it is difficult to measure [...] Read more.
In recent years, disaster chains caused by glacier movements have occurred frequently in the lower Yarlung Tsangpo River in southwest China. However, it is still unclear whether earthquakes significantly contribute to glacier movements and disaster chains. In addition, it is difficult to measure the high-frequency and large gradient displacement time series with optical remote sensing images due to cloud coverage. To this end, we take the Sedongpu disaster chain as an example, where the Milin earthquake, with an epicenter 11 km away, occurred on 18 November 2017. Firstly, to deal with the cloud coverage problem for single optical remote sensing analysis, we employed multiple platform optical images and conducted a cross-platform correlation technique to invert the two-dimensional displacement rate and the cumulative displacement time series of the Sedongpu glacier. To reveal the correlation between earthquakes and disaster chains, we divided the optical images into three classes according to the Milin earthquake event. Lastly, to increase the accuracy and reliability, we propose two strategies for displacement monitoring, that is, a four-quadrant block registration strategy and a multi-window fusion strategy. Results show that the RMSE reduction percentage of the proposed registration method reaches 80%, and the fusion method can retrieve the large magnitude displacements and complete displacement field. Secondly, the Milin earthquake accelerated the Sedongpu glacier movement, where the pre-seismic velocities were less than 0.5 m/day, the co-seismic velocities increased to 1 to 6 m/day, and the post-seismic velocities decreased to 0.5 to 3 m/day. Lastly, the earthquake had a triggering effect around 33 days on the Sedongpu disaster chain event on 21 December 2017. The failure pattern can be summarized as ice and rock collapse in the source area, large magnitude glacier displacement in the moraine area, and a large volume of sediment in the deposition area, causing a river blockage. Full article
Show Figures

Figure 1

17 pages, 11054 KiB  
Article
Advanced Plant Phenotyping: Unmanned Aerial Vehicle Remote Sensing and CimageA Software Technology for Precision Crop Growth Monitoring
by Hongyu Fu, Jianning Lu, Guoxian Cui, Jihao Nie, Wei Wang, Wei She and Jinwei Li
Agronomy 2024, 14(11), 2534; https://doi.org/10.3390/agronomy14112534 (registering DOI) - 28 Oct 2024
Abstract
In production activities and breeding programs, large-scale investigations of crop high-throughput phenotype information are needed to help improve management and decision-making. The development of UAV (unmanned aerial vehicle) remote sensing technology provides a new means for the large-scale, efficient, and accurate acquisition of [...] Read more.
In production activities and breeding programs, large-scale investigations of crop high-throughput phenotype information are needed to help improve management and decision-making. The development of UAV (unmanned aerial vehicle) remote sensing technology provides a new means for the large-scale, efficient, and accurate acquisition of crop phenotypes, but its practical application and popularization are hindered due to the complicated data processing required. To date, there is no automated system that can utilize the canopy images acquired through UAV to conduct a phenotypic character analysis. To address this bottleneck, we developed a new scalable software called CimageA. CimageA uses crop canopy images obtained by UAV as materials. It can combine machine vision technology and machine learning technology to conduct the high-throughput processing and phenotyping of crop remote sensing data. First, zoning tools are applied to draw an area-of-interest (AOI). Then, CimageA can rapidly extract vital remote sensing information such as the color, texture, and spectrum of the crop canopy in the plots. In addition, we developed data analysis modules that estimate and quantify related phenotypes (such as leaf area index, canopy coverage, and plant height) by analyzing the association between measured crop phenotypes and CimageA-derived remote sensing eigenvalues. Through a series of experiments, we confirmed that CimageA performs well in extracting high-throughput remote sensing information regarding crops, and verified the reliability of retrieving LAI (R2 = 0.796) and estimating plant height (R2 = 0.989) and planting area using CimageA. In short, CimageA is an efficient and non-destructive tool for crop phenotype analysis, which is of great value for monitoring crop growth and guiding breeding decisions. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

15 pages, 6433 KiB  
Technical Note
RSPS-SAM: A Remote Sensing Image Panoptic Segmentation Method Based on SAM
by Zhuoran Liu, Zizhen Li, Ying Liang, Claudio Persello, Bo Sun, Guangjun He and Lei Ma
Remote Sens. 2024, 16(21), 4002; https://doi.org/10.3390/rs16214002 (registering DOI) - 28 Oct 2024
Abstract
Satellite remote sensing images contain complex and diverse ground object information and the images exhibit spatial multi-scale characteristics, making the panoptic segmentation of satellite remote sensing images a highly challenging task. Due to the lack of large-scale annotated datasets for panoramic segmentation, existing [...] Read more.
Satellite remote sensing images contain complex and diverse ground object information and the images exhibit spatial multi-scale characteristics, making the panoptic segmentation of satellite remote sensing images a highly challenging task. Due to the lack of large-scale annotated datasets for panoramic segmentation, existing methods still suffer from weak model generalization capabilities. To mitigate this issue, this paper leverages the advantages of the Segment Anything Model (SAM), which can segment any object in remote sensing images without requiring any annotations and proposes a high-resolution remote sensing image panoptic segmentation method called Remote Sensing Panoptic Segmentation SAM (RSPS-SAM). Firstly, to address the problem of global information loss caused by cropping large remote sensing images for training, a Batch Attention Pyramid was designed to extract multi-scale features from remote sensing images and capture long-range contextual information between cropped patches, thereby enhancing the semantic understanding of remote sensing images. Secondly, we constructed a Mask Decoder to address the limitation of SAM requiring manual input prompts and its inability to output category information. This decoder utilized mask-based attention for mask segmentation, enabling automatic prompt generation and category prediction of segmented objects. Finally, the effectiveness of the proposed method was validated on the high-resolution remote sensing image airport scene dataset RSAPS-ASD. The results demonstrate that the proposed method achieves segmentation and recognition of foreground instances and background regions in high-resolution remote sensing images without the need for prompt input, while providing smooth segmentation boundaries with a panoptic segmentation quality (PQ) of 57.2, outperforming current mainstream methods. Full article
Show Figures

Figure 1

22 pages, 4798 KiB  
Article
Advancing Algorithmic Adaptability in Hyperspectral Anomaly Detection with Stacking-Based Ensemble Learning
by Bradley J. Wheeler and Hassan A. Karimi
Remote Sens. 2024, 16(21), 3994; https://doi.org/10.3390/rs16213994 (registering DOI) - 28 Oct 2024
Abstract
Anomaly detection in hyperspectral imaging is crucial for remote sensing, driving the development of numerous algorithms. However, systematic studies reveal a dichotomy where algorithms generally excel at either detecting anomalies in specific datasets or generalizing across heterogeneous datasets (i.e., lack adaptability). A key [...] Read more.
Anomaly detection in hyperspectral imaging is crucial for remote sensing, driving the development of numerous algorithms. However, systematic studies reveal a dichotomy where algorithms generally excel at either detecting anomalies in specific datasets or generalizing across heterogeneous datasets (i.e., lack adaptability). A key source of this dichotomy may center on the singular and like biases frequently employed by existing algorithms. Current research lacks experimentation into how integrating insights from diverse biases might counteract problems in singularly biased approaches. Addressing this gap, we propose stacking-based ensemble learning for hyperspectral anomaly detection (SELHAD). SELHAD introduces the integration of hyperspectral anomaly detection algorithms with diverse biases (e.g., Gaussian, density, partition) into a singular ensemble learning model and learns the factor to which each bias should contribute so anomaly detection performance is optimized. Additionally, it introduces bootstrapping strategies into hyperspectral anomaly detection algorithms to further increase robustness. We focused on five representative algorithms embodying common biases in hyperspectral anomaly detection and demonstrated how they result in the previously highlighted dichotomy. Subsequently, we demonstrated how SELHAD learns the interplay between these biases, enabling their collaborative utilization. In doing so, SELHAD transcends the limitations inherent in individual biases, thereby alleviating the dichotomy and advancing toward more adaptable solutions. Full article
(This article belongs to the Special Issue Remote Sensing: 15th Anniversary)
Show Figures

Figure 1

22 pages, 7881 KiB  
Article
Bidirectional Mamba with Dual-Branch Feature Extraction for Hyperspectral Image Classification
by Ming Sun, Jie Zhang, Xiaoou He and Yihe Zhong
Sensors 2024, 24(21), 6899; https://doi.org/10.3390/s24216899 (registering DOI) - 28 Oct 2024
Abstract
The hyperspectral image (HSI) classification task is widely used in remote sensing image analysis. The HSI classification methods based on convolutional neural networks (CNNs) have greatly improved the classification performance. However, they cannot well utilize the sequential properties of spectral features and face [...] Read more.
The hyperspectral image (HSI) classification task is widely used in remote sensing image analysis. The HSI classification methods based on convolutional neural networks (CNNs) have greatly improved the classification performance. However, they cannot well utilize the sequential properties of spectral features and face the challenge of increasing computational cost with the increase in network depth. To address these shortcomings, this paper proposes a novel network with a CNN-Mamba architecture, called DBMamba, which uses a bidirectional Mamba to process spectral feature sequences at a linear computational cost. In the DBMamba, principal component analysis (PCA) is first used to extract the main features of the data. Then, a dual-branch CNN structure, with the fused features from spectral–spatial features by 3D-CNN and spatial features by 2D-CNN, is used to extract shallow spectral–spatial features. Finally, a bidirectional Mamba is used to effectively capture global contextual information in features and significantly enhance the extraction of spectral features. Experimental results on the Indian Pines, Salinas, and Pavia University datasets demonstrate that the classification performance surpasses that of many cutting-edge methods, improving by 1.04%, 0.15%, and 0.09%, respectively, over the competing SSFTT method. The research in this paper enhances the existing knowledge on HSI classification and provides valuable insights for future research in this field. Full article
(This article belongs to the Special Issue Recent Advances in Synthetic Aperture Radar (SAR) Remote Sensing)
Show Figures

Figure 1

20 pages, 1971 KiB  
Article
A Patch-Level Region-Aware Module with a Multi-Label Framework for Remote Sensing Image Captioning
by Yunpeng Li, Xiangrong Zhang, Tianyang Zhang, Guanchun Wang, Xinlin Wang and Shuo Li
Remote Sens. 2024, 16(21), 3987; https://doi.org/10.3390/rs16213987 (registering DOI) - 27 Oct 2024
Abstract
Recent Transformer-based works can generate high-quality captions for remote sensing images (RSIs). However, these methods generally feed global or grid visual features to a Transformer-based captioning model for associating cross-modal information, which limits performance. In this work, we investigate unexplored ideas for a [...] Read more.
Recent Transformer-based works can generate high-quality captions for remote sensing images (RSIs). However, these methods generally feed global or grid visual features to a Transformer-based captioning model for associating cross-modal information, which limits performance. In this work, we investigate unexplored ideas for a remote sensing image captioning task, using a novel patch-level region-aware module with a multi-label framework. Due to an overhead perspective and a significantly larger scale in RSIs, a patch-level region-aware module is designed to filter the redundant information in the RSI scene, which benefits the Transformer-based decoder by attaining improved image perception. Technically, the trainable multi-label classifier capitalizes on semantic features as supplementary to the region-aware features. Moreover, modeling the inner relations of inputs is essential for understanding the RSI. Thus, we introduce region-oriented attention, which associates region features and semantic labels, omits the irrelevant regions to highlight relevant regions, and learns related semantic information. Extensive qualitative and quantitative experimental results show the superiority of our approach on the RSICD, UCM-Captions, and Sydney-Captions. The code for our method will be publicly available. Full article
Show Figures

Figure 1

24 pages, 11586 KiB  
Article
Integrating Thermal Infrared Imaging and Weather Data for Short-Term Prediction of Building Envelope Thermal Appearance
by Nikolay Golosov and Guido Cervone
Remote Sens. 2024, 16(21), 3981; https://doi.org/10.3390/rs16213981 (registering DOI) - 26 Oct 2024
Abstract
This study presents a novel deep-learning framework for predicting the thermal appearance of building envelopes under varying weather conditions based on a new dataset collected using a thermal infrared camera at 10 min intervals over a one-and-a-half-year period. Unlike existing studies that rely [...] Read more.
This study presents a novel deep-learning framework for predicting the thermal appearance of building envelopes under varying weather conditions based on a new dataset collected using a thermal infrared camera at 10 min intervals over a one-and-a-half-year period. Unlike existing studies that rely on simulated data or physical models that do not always accurately reflect the complex heat transfer processes in real buildings, we have collected a large dataset showing how a building behaves under different climatic conditions. We propose a novel deep-learning approach that integrates weather data and thermal imagery to predict the temperature distribution on the building façade for the next 24 and 48 h. The model uses a state-of-the-art recurrent neural network architecture, PredRNN V2, with an action conditioning mechanism to incorporate weather forecasting data into the prediction process. We evaluate this approach in terms of average accuracy, prediction accuracy in specific regions, and visual-perceptual performance of the images. The proposed framework achieves a prediction accuracy of 1.5 °C (root mean square error—RMSE) for the 24 h prediction and 2.04 °C (RMSE) for the 48 h prediction, outperforming baseline models in terms of temperature prediction accuracy and structural similarity of the predicted images. Full article
Show Figures

Figure 1

24 pages, 18639 KiB  
Article
National-Scale Detection of New Forest Roads in Sentinel-2 Time Series
by Øivind Due Trier and Arnt-Børre Salberg
Remote Sens. 2024, 16(21), 3972; https://doi.org/10.3390/rs16213972 - 25 Oct 2024
Abstract
The Norwegian Environment Agency is responsible for updating a map of undisturbed nature, which is performed every five years based on aerial photos. Some of the aerial photos are already up to five years old when a new version of the map of [...] Read more.
The Norwegian Environment Agency is responsible for updating a map of undisturbed nature, which is performed every five years based on aerial photos. Some of the aerial photos are already up to five years old when a new version of the map of undisturbed nature is published. Thus, several new nature interventions may have been missed. To address this issue, the timeliness and mapping accuracy were improved by integrating Sentinel-2 satellite imagery for the detection of new roads across Norway. The focus on new roads was due to the fact that most new nature interventions include the construction of new roads. The proposed methodology is based on applying U-Net on all the available summer images with less than 10% cloud cover over a five-year period, with an aggregation step to summarize the predictions. The observed detection rate was 98%. Post-processing steps reduced the false positive rate to 46%. However, as the false positive rate was still substantial, the manual verification of the predicted new roads was needed. The false negative rate was low, except in areas without vegetation. Full article
(This article belongs to the Special Issue Remote Sensing: 15th Anniversary)
Show Figures

Figure 1

22 pages, 4161 KiB  
Article
A Multi-Tiered Collaborative Network for Optical Remote Sensing Fine-Grained Ship Detection in Foggy Conditions
by Wenbo Zhou, Ligang Li, Bo Liu, Yuan Cao and Wei Ni
Remote Sens. 2024, 16(21), 3968; https://doi.org/10.3390/rs16213968 - 25 Oct 2024
Abstract
Ship target detection faces the challenges of complex and changing environments combined with the varied characteristics of ship targets. In practical applications, the complexity of meteorological conditions, uncertainty of lighting, and the diversity of ship target characteristics can affect the accuracy and efficiency [...] Read more.
Ship target detection faces the challenges of complex and changing environments combined with the varied characteristics of ship targets. In practical applications, the complexity of meteorological conditions, uncertainty of lighting, and the diversity of ship target characteristics can affect the accuracy and efficiency of ship target detection algorithms. Most existing target detection methods perform well in conditions of a general scenario but underperform in complex conditions. In this study, a collaborative network for target detection under foggy weather conditions is proposed, aiming to achieve improved accuracy while satisfying the need for real-time detection. First, a collaborative block was designed and SCConv and PCA modules were introduced to enhance the detection of low-quality images. Second, the PAN + FPN structure was adopted to take full advantage of its lightweight and efficient features. Finally, four detection heads were used to enhance the performance. In addition to this, a dataset for foggy ship detection was constructed based on ShipRSImageNet, and the mAP on the dataset reached 48.7%. The detection speed reached 33.3 frames per second (FPS), which is ultimately comparable to YOLOF. It shows that the model proposed has good detection effectiveness for remote sensing ship images during low-contrast foggy days. Full article
Show Figures

Figure 1

14 pages, 6043 KiB  
Article
Developing Site-Specific Prescription Maps for Sugarcane Weed Control Using High-Spatial-Resolution Images and Light Detection and Ranging (LiDAR)
by Kerin F. Romero and Muditha K. Heenkenda
Land 2024, 13(11), 1751; https://doi.org/10.3390/land13111751 - 25 Oct 2024
Abstract
Sugarcane is a perennial grass species mainly for sugar production and one of the significant crops in Costa Rica, where ideal growing conditions support its cultivation. Weed control is a critical aspect of sugarcane farming, traditionally managed through preventive or corrective mechanical and [...] Read more.
Sugarcane is a perennial grass species mainly for sugar production and one of the significant crops in Costa Rica, where ideal growing conditions support its cultivation. Weed control is a critical aspect of sugarcane farming, traditionally managed through preventive or corrective mechanical and chemical methods. However, these methods can be time-consuming and costly. This study aimed to develop site-specific, variable rate prescription maps for weed control using remote sensing. High-spatial-resolution images (5 cm) and Light Detection And Ranging (LiDAR) were acquired using a Micasense Rededge-P camera and a DJI L1 sensor mounted on a drone. Precise locations of weeds were collected for calibration and validation. Normalized Difference Vegetation Index derived from multispectral images separated vegetation coverage and soil. A deep learning (DL) algorithm further classified vegetation coverage into sugarcane and weeds. The DL model performed well without overfitting. The classification accuracy was 87% compared to validation samples. The density and average heights of weed patches were extracted from the canopy height model (LiDAR). They were used to derive site-specific prescription maps for weed control. This efficient and precise alternative to traditional methods could optimize weed control, reduce herbicide usage and provide more profitable yield. Full article
Show Figures

Figure 1

14 pages, 8742 KiB  
Article
Estimating Winter Canola Aboveground Biomass from Hyperspectral Images Using Narrowband Spectra-Texture Features and Machine Learning
by Xia Liu, Ruiqi Du, Youzhen Xiang, Junying Chen, Fucang Zhang, Hongzhao Shi, Zijun Tang and Xin Wang
Plants 2024, 13(21), 2978; https://doi.org/10.3390/plants13212978 - 25 Oct 2024
Abstract
Aboveground biomass (AGB) is a critical indicator for monitoring the crop growth status and predicting yields. UAV remote sensing technology offers an efficient and non-destructive method for collecting crop information in small-scale agricultural fields. High-resolution hyperspectral images provide abundant spectral-textural information, but whether [...] Read more.
Aboveground biomass (AGB) is a critical indicator for monitoring the crop growth status and predicting yields. UAV remote sensing technology offers an efficient and non-destructive method for collecting crop information in small-scale agricultural fields. High-resolution hyperspectral images provide abundant spectral-textural information, but whether they can enhance the accuracy of crop biomass estimations remains subject to further investigation. This study evaluates the predictability of winter canola AGB by integrating the narrowband spectra and texture features from UAV hyperspectral images. Specifically, narrowband spectra and vegetation indices were extracted from the hyperspectral images. The Gray Level Co-occurrence Matrix (GLCM) method was employed to compute texture indices. Correlation analysis and autocorrelation analysis were utilized to determine the final spectral feature scheme, texture feature scheme, and spectral-texture feature scheme. Subsequently, machine learning algorithms were applied to develop estimation models for winter canola biomass. The results indicate: (1) For spectra features, narrow-bands at 450~510 nm, 680~738 nm, 910~940 nm wavelength, as well as vegetation indices containing red-edge narrow-bands, showed outstanding performance with correlation coefficients ranging from 0.49 to 0.65; For texture features, narrow-band texture parameters CON, DIS, ENT, ASM, and vegetation index texture parameter COR demonstrated significant performance, with correlation coefficients between 0.65 and 0.72; (2) The Adaboost model using the spectra-texture feature scheme exhibited the best performance in estimating winter canola biomass (R2 = 0.91; RMSE = 1710.79 kg/ha; NRMSE = 19.88%); (3) The combined use of narrowband spectra and texture feature significantly improved the estimation accuracy of winter canola biomass. Compared to the spectra feature scheme, the model’s R2 increased by 11.2%, RMSE decreased by 29%, and NRMSE reduced by 17%. These findings provide a reference for studies on UAV hyperspectral remote sensing monitoring of crop growth status. Full article
Show Figures

Figure 1

32 pages, 7183 KiB  
Article
Learn from Simulations, Adapt to Observations: Super-Resolution of Isoprene Emissions via Unpaired Domain Adaptation
by Antonio Giganti, Sara Mandelli, Paolo Bestagini and Stefano Tubaro
Remote Sens. 2024, 16(21), 3963; https://doi.org/10.3390/rs16213963 - 24 Oct 2024
Abstract
Plants emit biogenic volatile organic compounds (BVOCs), such as isoprene, significantly influencing atmospheric chemistry and climate. BVOC emissions estimated from bottom-up (BU) approaches (derived from numerical simulations) usually exhibit denser and more detailed spatial information compared to those estimated through top-down (TD) approaches [...] Read more.
Plants emit biogenic volatile organic compounds (BVOCs), such as isoprene, significantly influencing atmospheric chemistry and climate. BVOC emissions estimated from bottom-up (BU) approaches (derived from numerical simulations) usually exhibit denser and more detailed spatial information compared to those estimated through top-down (TD) approaches (derived from satellite observations). Moreover, numerically simulated emissions are typically easier to obtain, even if they are less reliable than satellite acquisitions, which, being derived from actual measurements, are considered a more trustworthy instrument for performing chemistry and climate investigations. Given the coarseness and relative lack of satellite-derived measurements, fine-grained numerically simulated emissions could be exploited to enhance them. However, simulated (BU) and observed (TD) emissions usually differ regarding value range and spatiotemporal resolution. In this work, we present a novel deep learning (DL)-based approach to increase the spatial resolution of satellite-derived isoprene emissions, investigating the adoption of efficient domain adaptation (DA) techniques to bridge the gap between numerically simulated emissions and satellite-derived emissions, avoiding the need for retraining a specific SR algorithm on them. For this, we propose a methodology based on the CycleGAN architecture, which has been extensively used for adapting natural images (like digital photographs) of different domains. In our work, we depart from the standard CycleGAN framework, proposing additional loss terms that allow for better DA and emissions’ SR. We extensively demonstrate the proposed method’s effectiveness and robustness in restoring fine-grained patterns of observed isoprene emissions. Moreover, we compare different setups and validate our approach using different emission inventories from both domains. Eventually, we show that the proposed DA strategy paves the way towards robust SR solutions even in the case of spatial resolution mismatch between the training and testing domains and in the case of unknown testing data. Full article
16 pages, 2602 KiB  
Article
Multi-Scale and Multi-Network Deep Feature Fusion for Discriminative Scene Classification of High-Resolution Remote Sensing Images
by Baohua Yuan, Sukhjit Singh Sehra and Bernard Chiu
Remote Sens. 2024, 16(21), 3961; https://doi.org/10.3390/rs16213961 - 24 Oct 2024
Abstract
The advancement in satellite image sensors has enabled the acquisition of high-resolution remote sensing (HRRS) images. However, interpreting these images accurately and obtaining the computational power needed to do so is challenging due to the complexity involved. This manuscript proposed a multi-stream convolutional [...] Read more.
The advancement in satellite image sensors has enabled the acquisition of high-resolution remote sensing (HRRS) images. However, interpreting these images accurately and obtaining the computational power needed to do so is challenging due to the complexity involved. This manuscript proposed a multi-stream convolutional neural network (CNN) fusion framework that involves multi-scale and multi-CNN integration for HRRS image recognition. The pre-trained CNNs were used to learn and extract semantic features from multi-scale HRRS images. Feature extraction using pre-trained CNNs is more efficient than training a CNN from scratch or fine-tuning a CNN. Discriminative canonical correlation analysis (DCCA) was used to fuse deep features extracted across CNNs and image scales. DCCA reduced the dimension of the features extracted from CNNs while providing a discriminative representation by maximizing the within-class correlation and minimizing the between-class correlation. The proposed model has been evaluated on NWPU-RESISC45 and UC Merced datasets. The accuracy associated with DCCA was 10% and 6% higher than discriminant correlation analysis (DCA) in the NWPU-RESISC45 and UC Merced datasets. The advantage of DCCA was better demonstrated in the NWPU-RESISC45 dataset due to the incorporation of richer within-class variability in this dataset. While both DCA and DCCA minimize between-class correlation, only DCCA maximizes the within-class correlation and, therefore, attains better accuracy. The proposed framework achieved higher accuracy than all state-of-the-art frameworks involving unsupervised learning and pre-trained CNNs and 2–3% higher than the majority of fine-tuned CNNs. The proposed framework offers computational time advantages, requiring only 13 s for training in NWPU-RESISC45, compared to a day for fine-tuning the existing CNNs. Thus, the proposed framework achieves a favourable balance between efficiency and accuracy in HRRS image recognition. Full article
Show Figures

Figure 1

23 pages, 5508 KiB  
Article
YOLO-DroneMS: Multi-Scale Object Detection Network for Unmanned Aerial Vehicle (UAV) Images
by Xueqiang Zhao and Yangbo Chen
Drones 2024, 8(11), 609; https://doi.org/10.3390/drones8110609 - 24 Oct 2024
Abstract
In recent years, research on Unmanned Aerial Vehicles (UAVs) has developed rapidly. Compared to traditional remote-sensing images, UAV images exhibit complex backgrounds, high resolution, and large differences in object scales. Therefore, UAV object detection is an essential yet challenging task. This paper proposes [...] Read more.
In recent years, research on Unmanned Aerial Vehicles (UAVs) has developed rapidly. Compared to traditional remote-sensing images, UAV images exhibit complex backgrounds, high resolution, and large differences in object scales. Therefore, UAV object detection is an essential yet challenging task. This paper proposes a multi-scale object detection network, namely YOLO-DroneMS (You Only Look Once for Drone Multi-Scale Object), for UAV images. Targeting the pivotal connection between the backbone and neck, the Large Separable Kernel Attention (LSKA) mechanism is adopted with the Spatial Pyramid Pooling Factor (SPPF), where weighted processing of multi-scale feature maps is performed to focus more on features. And Attentional Scale Sequence Fusion DySample (ASF-DySample) is introduced to perform attention scale sequence fusion and dynamic upsampling to conserve resources. Then, the faster cross-stage partial network bottleneck with two convolutions (named C2f) in the backbone is optimized using the Inverted Residual Mobile Block and Dilated Reparam Block (iRMB-DRB), which balances the advantages of dynamic global modeling and static local information fusion. This optimization effectively increases the model’s receptive field, enhancing its capability for downstream tasks. By replacing the original CIoU with WIoUv3, the model prioritizes anchoring boxes of superior quality, dynamically adjusting weights to enhance detection performance for small objects. Experimental findings on the VisDrone2019 dataset demonstrate that at an Intersection over Union (IoU) of 0.5, YOLO-DroneMS achieves a 3.6% increase in mAP@50 compared to the YOLOv8n model. Moreover, YOLO-DroneMS exhibits improved detection speed, increasing the number of frames per second (FPS) from 78.7 to 83.3. The enhanced model supports diverse target scales and achieves high recognition rates, making it well-suited for drone-based object detection tasks, particularly in scenarios involving multiple object clusters. Full article
(This article belongs to the Special Issue Intelligent Image Processing and Sensing for Drones 2nd Edition)
Show Figures

Figure 1

Back to TopTop