Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (130)

Search Parameters:
Keywords = hierarchical semantic feature

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 7668 KiB  
Article
Semantics-Preserving RDB2RDF Data Transformation Using Hierarchical Direct Mapping
by Hee-Gook Jun and Dong-Hyuk Im
Appl. Sci. 2020, 10(20), 7070; https://doi.org/10.3390/app10207070 - 12 Oct 2020
Cited by 3 | Viewed by 2440
Abstract
Direct mapping is an automatic transformation method used to generate resource description framework (RDF) data from relational data. In the field of direct mapping, semantics preservation is critical to ensure that the mapping method outputs RDF data without information loss or incorrect semantic [...] Read more.
Direct mapping is an automatic transformation method used to generate resource description framework (RDF) data from relational data. In the field of direct mapping, semantics preservation is critical to ensure that the mapping method outputs RDF data without information loss or incorrect semantic data generation. However, existing direct-mapping methods have problems that prevent semantics preservation in specific cases. For this reason, a mapping method is developed to perform a semantics-preserving transformation of relational databases (RDB) into RDF data without semantic information loss and to reduce the volume of incorrect RDF data. This research reviews cases that do not generate semantics-preserving results, and the corresponding problems into categories are arranged. This paper defines lemmas that represent the features of RDF data transformation to resolve those problems. Based on the lemmas, this work develops a hierarchical direct-mapping method to strictly abide by the definition of semantics preservation and to prevent semantic information loss, reducing the volume of incorrect RDF data generated. Experiments demonstrate the capability of the proposed method to perform semantics-preserving RDB2RDF data transformation, generating semantically accurate results. This work impacts future studies, which should involve the development of synchronization methods to achieve RDF data consistency when original RDB data are modified. Full article
(This article belongs to the Special Issue Advances in Deep Learning Ⅱ)
Show Figures

Figure 1

15 pages, 1463 KiB  
Article
Dual Attention-Guided Multiscale Dynamic Aggregate Graph Convolutional Networks for Skeleton-Based Human Action Recognition
by Zeyuan Hu and Eung-Joo Lee
Symmetry 2020, 12(10), 1589; https://doi.org/10.3390/sym12101589 - 24 Sep 2020
Cited by 7 | Viewed by 2385
Abstract
Traditional convolution neural networks have achieved great success in human action recognition. However, it is challenging to establish effective associations between different human bone nodes to capture detailed information. In this paper, we propose a dual attention-guided multiscale dynamic aggregate graph convolution neural [...] Read more.
Traditional convolution neural networks have achieved great success in human action recognition. However, it is challenging to establish effective associations between different human bone nodes to capture detailed information. In this paper, we propose a dual attention-guided multiscale dynamic aggregate graph convolution neural network (DAG-GCN) for skeleton-based human action recognition. Our goal is to explore the best correlation and determine high-level semantic features. First, a multiscale dynamic aggregate GCN module is used to capture important semantic information and to establish dependence relationships for different bone nodes. Second, the higher level semantic feature is further refined, and the semantic relevance is emphasized through a dual attention guidance module. In addition, we exploit the relationship of joints hierarchically and the spatial temporal correlations through two modules. Experiments with the DAG-GCN method result in good performance on the NTU-60-RGB+D and NTU-120-RGB+D datasets. The accuracy is 95.76% and 90.01%, respectively, for the cross (X)-View and X-Subon the NTU60dataset. Full article
(This article belongs to the Section Computer)
Show Figures

Graphical abstract

18 pages, 9851 KiB  
Article
Semantic-Based Building Extraction from LiDAR Point Clouds Using Contexts and Optimization in Complex Environment
by Yongjun Wang, Tengping Jiang, Min Yu, Shuaibing Tao, Jian Sun and Shan Liu
Sensors 2020, 20(12), 3386; https://doi.org/10.3390/s20123386 - 15 Jun 2020
Cited by 25 | Viewed by 3085
Abstract
The extraction of buildings has been an essential part of the field of LiDAR point clouds processing in recent years. However, it is still challenging to extract buildings from huge amount of point clouds due to the complicated and incomplete structures, occlusions and [...] Read more.
The extraction of buildings has been an essential part of the field of LiDAR point clouds processing in recent years. However, it is still challenging to extract buildings from huge amount of point clouds due to the complicated and incomplete structures, occlusions and local similarities between different categories in a complex environment. Taking the urban and campus scene as examples, this paper presents a versatile and hierarchical semantic-based method for building extraction using LiDAR point clouds. The proposed method first performs a series of preprocessing operations, such as removing ground points, establishing super-points and using them as primitives for subsequent processing, and then semantically labels the raw LiDAR data. In the feature engineering process, considering the purpose of this article is to extract buildings, we tend to choose the features extracted from super-points that can describe building for the next classification. There are a portion of inaccurate labeling results due to incomplete or overly complex scenes, a Markov Random Field (MRF) optimization model is constructed for postprocessing and segmentation results refinement. Finally, the buildings are extracted from the labeled points. Experimental verification was performed on three datasets in different scenes, our results were compared with the state-of-the-art methods. These evaluation results demonstrate the feasibility and effectiveness of the proposed method for extracting buildings from LiDAR point clouds in multiple environments. Full article
(This article belongs to the Special Issue LiDAR-Based Creation of Virtual Cities)
Show Figures

Figure 1

20 pages, 4741 KiB  
Article
Image Collection Summarization Method Based on Semantic Hierarchies
by Zahra Riahi Samani and Mohsen Ebrahimi Moghaddam
AI 2020, 1(2), 209-228; https://doi.org/10.3390/ai1020014 - 18 May 2020
Cited by 2 | Viewed by 4771
Abstract
The size of internet image collections is increasing drastically. As a result, new techniques are required to facilitate users in browsing, navigation, and summarization of these large volume collections. Image collection summarization methods present users with a set of exemplar images as the [...] Read more.
The size of internet image collections is increasing drastically. As a result, new techniques are required to facilitate users in browsing, navigation, and summarization of these large volume collections. Image collection summarization methods present users with a set of exemplar images as the most representative ones from the initial image collection. In this study, an image collection summarization technique was introduced according to semantic hierarchies among them. In the proposed approach, images were mapped to the nodes of a pre-defined domain ontology. In this way, a semantic hierarchical classifier was used, which finally mapped images to different nodes of the ontology. We made a compromise between the degree of freedom of the classifier and the goodness of the summarization method. The summarization was done using a group of high-level features that provided a semantic measurement of information in images. Experimental outcomes indicated that the introduced image collection summarization method outperformed the recent techniques for the summarization of image collections. Full article
Show Figures

Figure 1

16 pages, 2892 KiB  
Article
Redesigned Skip-Network for Crowd Counting with Dilated Convolution and Backward Connection
by Sorn Sooksatra, Toshiaki Kondo, Pished Bunnun and Atsuo Yoshitaka
J. Imaging 2020, 6(5), 28; https://doi.org/10.3390/jimaging6050028 - 2 May 2020
Cited by 3 | Viewed by 3761
Abstract
Crowd counting is a challenging task dealing with the variation of an object scale and a crowd density. Existing works have emphasized on skip connections by integrating shallower layers with deeper layers, where each layer extracts features in a different object scale and [...] Read more.
Crowd counting is a challenging task dealing with the variation of an object scale and a crowd density. Existing works have emphasized on skip connections by integrating shallower layers with deeper layers, where each layer extracts features in a different object scale and crowd density. However, only high-level features are emphasized while ignoring low-level features. This paper proposes an estimation network by passing high-level features to shallow layers and emphasizing its low-level feature. Since an estimation network is a hierarchical network, a high-level feature is also emphasized by an improved low-level feature. Our estimation network consists of two identical networks for extracting a high-level feature and estimating the final result. To preserve semantic information, dilated convolution is employed without resizing the feature map. Our method was tested in three datasets for counting humans and vehicles in a crowd image. The counting performance is evaluated by mean absolute error and root mean squared error indicating the accuracy and robustness of an estimation network, respectively. The experimental result shows that our network outperforms other related works in a high crowd density and is effective for reducing over-counting error in the overall case. Full article
Show Figures

Figure 1

16 pages, 2557 KiB  
Article
EHANet: An Effective Hierarchical Aggregation Network for Face Parsing
by Ling Luo, Dingyu Xue and Xinglong Feng
Appl. Sci. 2020, 10(9), 3135; https://doi.org/10.3390/app10093135 - 30 Apr 2020
Cited by 18 | Viewed by 4576
Abstract
In recent years, benefiting from deep convolutional neural networks (DCNNs), face parsing has developed rapidly. However, it still has the following problems: (1) Existing state-of-the-art frameworks usually do not satisfy real-time while pursuing performance; (2) similar appearances cause incorrect pixel label assignments, especially [...] Read more.
In recent years, benefiting from deep convolutional neural networks (DCNNs), face parsing has developed rapidly. However, it still has the following problems: (1) Existing state-of-the-art frameworks usually do not satisfy real-time while pursuing performance; (2) similar appearances cause incorrect pixel label assignments, especially in the boundary; (3) to promote multi-scale prediction, deep features and shallow features are used for fusion without considering the semantic gap between them. To overcome these drawbacks, we propose an effective and efficient hierarchical aggregation network called EHANet for fast and accurate face parsing. More specifically, we first propose a stage contextual attention mechanism (SCAM), which uses higher-level contextual information to re-encode the channel according to its importance. Secondly, a semantic gap compensation block (SGCB) is presented to ensure the effective aggregation of hierarchical information. Thirdly, the advantages of weighted boundary-aware loss effectively make up for the ambiguity of boundary semantics. Without any bells and whistles, combined with a lightweight backbone, we achieve outstanding results on both CelebAMask-HQ (78.19% mIoU) and Helen datasets (90.7% F1-score). Furthermore, our model can achieve 55 FPS on a single GTX 1080Ti card with 640 × 640 input and further reach over 300 FPS with a resolution of 256 × 256, which is suitable for real-world applications. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

17 pages, 23988 KiB  
Article
Semantic Segmentation of Natural Materials on a Point Cloud Using Spatial and Multispectral Features
by J. M. Jurado, J. L. Cárdenas, C. J. Ogayar, L. Ortega and F. R. Feito
Sensors 2020, 20(8), 2244; https://doi.org/10.3390/s20082244 - 15 Apr 2020
Cited by 12 | Viewed by 3425
Abstract
The characterization of natural spaces by the precise observation of their material properties is highly demanded in remote sensing and computer vision. The production of novel sensors enables the collection of heterogeneous data to get a comprehensive knowledge of the living and non-living [...] Read more.
The characterization of natural spaces by the precise observation of their material properties is highly demanded in remote sensing and computer vision. The production of novel sensors enables the collection of heterogeneous data to get a comprehensive knowledge of the living and non-living entities in the ecosystem. The high resolution of consumer-grade RGB cameras is frequently used for the geometric reconstruction of many types of environments. Nevertheless, the understanding of natural spaces is still challenging. The automatic segmentation of homogeneous materials in nature is a complex task because there are many overlapping structures and an indirect illumination, so the object recognition is difficult. In this paper, we propose a method based on fusing spatial and multispectral characteristics for the unsupervised classification of natural materials in a point cloud. A high-resolution camera and a multispectral sensor are mounted on a custom camera rig in order to simultaneously capture RGB and multispectral images. Our method is tested in a controlled scenario, where different natural objects coexist. Initially, the input RGB images are processed to generate a point cloud by applying the structure-from-motion (SfM) algorithm. Then, the multispectral images are mapped on the three-dimensional model to characterize the geometry with the reflectance captured from four narrow bands (green, red, red-edge and near-infrared). The reflectance, the visible colour and the spatial component are combined to extract key differences among all existing materials. For this purpose, a hierarchical cluster analysis is applied to pool the point cloud and identify the feature pattern for every material. As a result, the tree trunk, the leaves, different species of low plants, the ground and rocks can be clearly recognized in the scene. These results demonstrate the feasibility to perform a semantic segmentation by considering multispectral and spatial features with an unknown number of clusters to be detected on the point cloud. Moreover, our solution is compared to other method based on supervised learning in order to test the improvement of the proposed approach. Full article
(This article belongs to the Collection Multi-Sensor Information Fusion)
Show Figures

Figure 1

19 pages, 6163 KiB  
Article
Improved Remote Sensing Image Classification Based on Multi-Scale Feature Fusion
by Chengming Zhang, Yan Chen, Xiaoxia Yang, Shuai Gao, Feng Li, Ailing Kong, Dawei Zu and Li Sun
Remote Sens. 2020, 12(2), 213; https://doi.org/10.3390/rs12020213 - 8 Jan 2020
Cited by 34 | Viewed by 4799
Abstract
When extracting land-use information from remote sensing imagery using image segmentation, obtaining fine edges for extracted objects is a key problem that is yet to be solved. In this study, we developed a new weight feature value convolutional neural network (WFCNN) to perform [...] Read more.
When extracting land-use information from remote sensing imagery using image segmentation, obtaining fine edges for extracted objects is a key problem that is yet to be solved. In this study, we developed a new weight feature value convolutional neural network (WFCNN) to perform fine remote sensing image segmentation and extract improved land-use information from remote sensing imagery. The WFCNN includes one encoder and one classifier. The encoder obtains a set of spectral features and five levels of semantic features. It uses the linear fusion method to hierarchically fuse the semantic features, employs an adjustment layer to optimize every level of fused features to ensure the stability of the pixel features, and combines the fused semantic and spectral features to form a feature graph. The classifier then uses a Softmax model to perform pixel-by-pixel classification. The WFCNN was trained using a stochastic gradient descent algorithm; the former and two variants were subject to experimental testing based on Gaofen 6 images and aerial images that compared them with the commonly used SegNet, U-NET, and RefineNet models. The accuracy, precision, recall, and F1-Score of the WFCNN were higher than those of the other models, indicating certain advantages in pixel-by-pixel segmentation. The results clearly show that the WFCNN can improve the accuracy and automation level of large-scale land-use mapping and the extraction of other information using remote sensing imagery. Full article
Show Figures

Graphical abstract

16 pages, 1204 KiB  
Article
Named-Entity Recognition in Sports Field Based on a Character-Level Graph Convolutional Network
by Xieraili Seti, Aishan Wumaier, Turgen Yibulayin, Diliyaer Paerhati, Lulu Wang and Alimu Saimaiti
Information 2020, 11(1), 30; https://doi.org/10.3390/info11010030 - 5 Jan 2020
Cited by 9 | Viewed by 3643
Abstract
Traditional methods for identifying naming ignore the correlation between named entities and lose hierarchical structural information between the named entities in a given text. Although traditional named-entity methods are effective for conventional datasets that have simple structures, they are not as effective for [...] Read more.
Traditional methods for identifying naming ignore the correlation between named entities and lose hierarchical structural information between the named entities in a given text. Although traditional named-entity methods are effective for conventional datasets that have simple structures, they are not as effective for sports texts. This paper proposes a Chinese sports text named-entity recognition method based on a character graph convolutional neural network (Char GCN) with a self-attention mechanism model. In this method, each Chinese character in the sports text is regarded as a node. The edge between the nodes is constructed using a similar character position and the character feature of the named-entity in the sports text. The internal structural information of the entity is extracted using a character map convolutional neural network. The hierarchical semantic information of the sports text is captured by the self-attention model to enhance the relationship between the named entities and capture the relevance and dependency between the characters. The conditional random fields classification function can accurately identify the named entities in the Chinese sports text. The results conducted on four datasets demonstrate that the proposed method improves the F-Score values significantly to 92.51%, 91.91%, 93.98%, and 95.01%, respectively, in comparison to the traditional naming methods. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

19 pages, 4375 KiB  
Article
Using Geographic Ontologies and Geo-Characterization to Represent Geographic Scenarios
by Yi Huang, May Yuan, Yehua Sheng, Xiangqiang Min and Yuwei Cao
ISPRS Int. J. Geo-Inf. 2019, 8(12), 566; https://doi.org/10.3390/ijgi8120566 - 10 Dec 2019
Cited by 11 | Viewed by 3751
Abstract
Traditional Geographic Information Systems (GIS) represent the environment under reductionist thinking, which disaggregates a geographic environment into independent geographic themes. The reductionist approach makes the spatiotemporal characteristics of geo-features explicit, but neglects the holistic nature of the environment, such as the hierarchical structure [...] Read more.
Traditional Geographic Information Systems (GIS) represent the environment under reductionist thinking, which disaggregates a geographic environment into independent geographic themes. The reductionist approach makes the spatiotemporal characteristics of geo-features explicit, but neglects the holistic nature of the environment, such as the hierarchical structure and interactions among environmental elements. To fill this gap, we integrate the concept geographic scenario with the fundamental principles of General System Theory to realize the environmental complexity in GIS. With the integration, a geographic scenario constitutes a hierarchy of spatiotemporal frameworks for organizing environmental elements and subserving the exploration of their relationships. Furthermore, we propose geo-characterization with ontological commitments to both static and dynamic properties of a geographic scenario and prescribe spatial, temporal, semantic, interactive, and causal relationships among environmental elements. We have tested the utility of the proposed representation in OWL and the associated reasoning process in Semantic Web Rule Language (SWRL) rules in a case study in Nanjing, China. The case study represents Nanjing and the Nanjing presidential palace to demonstrate the connections among environmental elements in different scenarios and the support for information queries, evolution process simulation, and semantic inferences. The proposed representation encodes geographic knowledge of the environment, makes the interactions among environmental elements explicit, supports geographic process simulation, opens opportunities for deep knowledge mining, and grounds a foundation for GeoAI to discover geographic complexity and dynamics beyond the support of conventional theme-centric inquiries in GIS. Full article
Show Figures

Figure 1

17 pages, 2334 KiB  
Article
Hierarchical Semantic Correspondence Analysis on Feature Classes between Two Geospatial Datasets Using a Graph Embedding Method
by Yong Huh
ISPRS Int. J. Geo-Inf. 2019, 8(11), 479; https://doi.org/10.3390/ijgi8110479 - 24 Oct 2019
Cited by 1 | Viewed by 1848
Abstract
A method to find corresponding feature class pairs, including hierarchical M:N pairs between two geospatial datasets is proposed. Applying an overlapping analysis to the object sets within the feature classes, the similarities of the feature classes are estimated and projected onto a lower-dimensional [...] Read more.
A method to find corresponding feature class pairs, including hierarchical M:N pairs between two geospatial datasets is proposed. Applying an overlapping analysis to the object sets within the feature classes, the similarities of the feature classes are estimated and projected onto a lower-dimensional vector space after applying the graph embedding method. In this space, conventional mathematical tools—agglomerative hierarchical clustering in this study—could be used to analyze semantic correspondences between the datasets and identify their hierarchical M:N corresponding pairs. The proposed method was applied to two cadastral parcel datasets; one for latest land-use records in an urban information system, and the other, for original land-use categories in the Korea land information system. To quantitatively assess identified feature pairs, F-measures for each pair are presented. The results showed that it was possible to find various semantic correspondences of the feature classes and infer regional land development characteristics. Full article
Show Figures

Graphical abstract

22 pages, 1262 KiB  
Review
Deep Learning on Point Clouds and Its Application: A Survey
by Weiping Liu, Jia Sun, Wanyi Li, Ting Hu and Peng Wang
Sensors 2019, 19(19), 4188; https://doi.org/10.3390/s19194188 - 26 Sep 2019
Cited by 149 | Viewed by 13744
Abstract
Point cloud is a widely used 3D data form, which can be produced by depth sensors, such as Light Detection and Ranging (LIDAR) and RGB-D cameras. Being unordered and irregular, many researchers focused on the feature engineering of the point cloud. Being able [...] Read more.
Point cloud is a widely used 3D data form, which can be produced by depth sensors, such as Light Detection and Ranging (LIDAR) and RGB-D cameras. Being unordered and irregular, many researchers focused on the feature engineering of the point cloud. Being able to learn complex hierarchical structures, deep learning has achieved great success with images from cameras. Recently, many researchers have adapted it into the applications of the point cloud. In this paper, the recent existing point cloud feature learning methods are classified as point-based and tree-based. The former directly takes the raw point cloud as the input for deep learning. The latter first employs a k-dimensional tree (Kd-tree) structure to represent the point cloud with a regular representation and then feeds these representations into deep learning models. Their advantages and disadvantages are analyzed. The applications related to point cloud feature learning, including 3D object classification, semantic segmentation, and 3D object detection, are introduced, and the datasets and evaluation metrics are also collected. Finally, the future research trend is predicted. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

23 pages, 5129 KiB  
Article
Web-Net: A Novel Nest Networks with Ultra-Hierarchical Sampling for Building Extraction from Aerial Imageries
by Yan Zhang, Weiguo Gong, Jingxi Sun and Weihong Li
Remote Sens. 2019, 11(16), 1897; https://doi.org/10.3390/rs11161897 - 14 Aug 2019
Cited by 38 | Viewed by 3950
Abstract
How to efficiently utilize vast amounts of easily accessed aerial imageries is a critical challenge for researchers with the proliferation of high-resolution remote sensing sensors and platforms. Recently, the rapid development of deep neural networks (DNN) has been a focus in remote sensing, [...] Read more.
How to efficiently utilize vast amounts of easily accessed aerial imageries is a critical challenge for researchers with the proliferation of high-resolution remote sensing sensors and platforms. Recently, the rapid development of deep neural networks (DNN) has been a focus in remote sensing, and the networks have achieved remarkable progress in image classification and segmentation tasks. However, the current DNN models inevitably lose the local cues during the downsampling operation. Additionally, even with skip connections, the upsampling methods cannot properly recover the structural information, such as the edge intersections, parallelism, and symmetry. In this paper, we propose the Web-Net, which is a nested network architecture with hierarchical dense connections, to handle these issues. We design the Ultra-Hierarchical Sampling (UHS) block to absorb and fuse the inter-level feature maps to propagate the feature maps among different levels. The position-wise downsampling/upsampling methods in the UHS iteratively change the shape of the inputs while preserving the number of their parameters, so that the low-level local cues and high-level semantic cues are properly preserved. We verify the effectiveness of the proposed Web-Net in the Inria Aerial Dataset and WHU Dataset. The results of the proposed Web-Net achieve an overall accuracy of 96.97% and an IoU (Intersection over Union) of 80.10% on the Inria Aerial Dataset, which surpasses the state-of-the-art SegNet 1.8% and 9.96%, respectively; the results on the WHU Dataset also support the effectiveness of the proposed Web-Net. Additionally, benefitting from the nested network architecture and the UHS block, the extracted buildings on the prediction maps are obviously sharper and more accurately identified, and even the building areas that are covered by shadows can also be correctly extracted. The verified results indicate that the proposed Web-Net is both effective and efficient for building extraction from high-resolution remote sensing images. Full article
(This article belongs to the Special Issue Remote Sensing based Building Extraction)
Show Figures

Graphical abstract

20 pages, 3958 KiB  
Article
Finer Resolution Mapping of Marine Aquaculture Areas Using WorldView-2 Imagery and a Hierarchical Cascade Convolutional Neural Network
by Yongyong Fu, Ziran Ye, Jinsong Deng, Xinyu Zheng, Yibo Huang, Wu Yang, Yaohua Wang and Ke Wang
Remote Sens. 2019, 11(14), 1678; https://doi.org/10.3390/rs11141678 - 15 Jul 2019
Cited by 34 | Viewed by 4333
Abstract
Marine aquaculture plays an important role in seafood supplement, economic development, and coastal ecosystem service provision. The precise delineation of marine aquaculture areas from high spatial resolution (HSR) imagery is vital for the sustainable development and management of coastal marine resources. However, various [...] Read more.
Marine aquaculture plays an important role in seafood supplement, economic development, and coastal ecosystem service provision. The precise delineation of marine aquaculture areas from high spatial resolution (HSR) imagery is vital for the sustainable development and management of coastal marine resources. However, various sizes and detailed structures of marine objects make it difficult for accurate mapping from HSR images by using conventional methods. Therefore, this study attempts to extract marine aquaculture areas by using an automatic labeling method based on the convolutional neural network (CNN), i.e., an end-to-end hierarchical cascade network (HCNet). Specifically, for marine objects of various sizes, we propose to improve the classification performance by utilizing multi-scale contextual information. Technically, based on the output of a CNN encoder, we employ atrous convolutions to capture multi-scale contextual information and aggregate them in a hierarchical cascade way. Meanwhile, for marine objects with detailed structures, we propose to refine the detailed information gradually by using a series of long-span connections with fine resolution features from the shallow layers. In addition, to decrease the semantic gaps between features in different levels, we propose to refine the feature space (i.e., channel and spatial dimensions) using an attention-based module. Experimental results show that our proposed HCNet can effectively identify and distinguish different kinds of marine aquaculture, with 98% of overall accuracy. It also achieves better classification performance compared with object-based support vector machine and state-of-the-art CNN-based methods, such as FCN-32s, U-Net, and DeeplabV2. Our developed method lays a solid foundation for the intelligent monitoring and management of coastal marine resources. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Graphical abstract

20 pages, 9161 KiB  
Article
A Multi-Primitive-Based Hierarchical Optimal Approach for Semantic Labeling of ALS Point Clouds
by Xuming Ge, Bo Wu, Yuan Li and Han Hu
Remote Sens. 2019, 11(10), 1243; https://doi.org/10.3390/rs11101243 - 24 May 2019
Cited by 8 | Viewed by 3786
Abstract
There are normally three main steps to carrying out the labeling of airborne laser scanning (ALS) point clouds. The first step is to use appropriate primitives to represent the scanning scenes, the second is to calculate the discriminative features of each primitive, and [...] Read more.
There are normally three main steps to carrying out the labeling of airborne laser scanning (ALS) point clouds. The first step is to use appropriate primitives to represent the scanning scenes, the second is to calculate the discriminative features of each primitive, and the third is to introduce a classifier to label the point clouds. This paper investigates multiple primitives to effectively represent scenes and exploit their geometric relationships. Relationships are graded according to the properties of related primitives. Then, based on initial labeling results, a novel, hierarchical, and optimal strategy is developed to optimize semantic labeling results. The proposed approach was tested using two sets of representative ALS point clouds, namely the Vaihingen datasets and Hong Kong’s Central District dataset. The results were compared with those generated by other typical methods in previous work. Quantitative assessments for the two experimental datasets showed that the performance of the proposed approach was superior to reference methods in both datasets. The scores for correctness attained over 98% in all cases of the Vaihingen datasets and up to 96% in the Hong Kong dataset. The results reveal that our approach of labeling different classes in terms of ALS point clouds is robust and bears significance for future applications, such as 3D modeling and change detection from point clouds. Full article
(This article belongs to the Special Issue Point Cloud Processing in Remote Sensing)
Show Figures

Figure 1

Back to TopTop