Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
DRFM-Based Repeater Jamming Reconstruction and Cancellation Method with Accurate Edge Detection
Next Article in Special Issue
Hyperspectral Anomaly Detection Using Spatial–Spectral-Based Union Dictionary and Improved Saliency Weight
Previous Article in Journal
HY1C/D-CZI Noctiluca scintillans Bloom Recognition Network Based on Hybrid Convolution and Self-Attention
Previous Article in Special Issue
From Video to Hyperspectral: Hyperspectral Image-Level Feature Extraction with Transfer Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Spectral-Spatial Attention Network for Hyperspectral Image Classification Combining 2D Octave and 3D Convolutional Neural Networks

1
College of Electrical and Information Engineering, Hunan University, Changsha 418002, China
2
Jiangxi Province Key Laboratory of Water Information Cooperative Sensing and Intelligent Processing, School of Information Engineering, Nanchang Institute of Technology, Nanchang 330099, China
3
Hubei Key Laboratory of Intelligent Geo-Information Processing, School of Computer Science, China University of Geosciences, Wuhan 430078, China
4
Hyperspectral Computing Laboratory, Department of Technology of Computers and Communications, University of Extremadura, E-10071 Caceres, Spain
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(7), 1758; https://doi.org/10.3390/rs15071758
Submission received: 23 February 2023 / Revised: 20 March 2023 / Accepted: 22 March 2023 / Published: 24 March 2023
(This article belongs to the Special Issue Computational Intelligence in Hyperspectral Remote Sensing)

Abstract

:
Traditional convolutional neural networks (CNNs) can be applied to obtain the spectral-spatial feature information from hyperspectral images (HSIs). However, they often introduce significant redundant spatial feature information. The octave convolution network is frequently utilized instead of traditional CNN to decrease spatial redundant information of the network and extend its receptive field. However, the 3D octave convolution-based approaches may introduce extensive parameters and complicate the network. To solve these issues, we propose a new HSI classification approach with a multi-scale spectral-spatial network-based framework that combines 2D octave and 3D CNNs. Our method, called MOCNN, first utilizes 2D octave convolution and 3D DenseNet branch networks with various convolutional kernel sizes to obtain complex spatial contextual feature information and spectral characteristics, separately. Moreover, the channel and the spectral attention mechanisms are, respectively, applied to these two branch networks to emphasize significant feature regions and certain important spectral bands that comprise discriminative information for the categorization. Furthermore, a sample balancing strategy is applied to address the sample imbalance problem. Expansive experiments are undertaken on four HSI datasets, demonstrating that our MOCNN approach outperforms several other methods for HSI classification, especially in scenarios dominated by limited and imbalanced sample data.

1. Introduction

Hyperspectral images (HSIs) are captured by hyperspectral imagers, and consist of tens or even a few hundred consecutive narrow bands of the spectrum. These instruments are capable of collecting abundant spectral and spatial information [1,2,3]. The resulting data cubes can be used to accurately identify ground materials and discriminate ground covers. Thus, HSIs have been extensively applied in various domains, including environmental monitoring [4], agriculture construction [5], military defense [6], anomaly detection [7], and so on. The task of HSI categorization aims to allocate a unique and concrete category to individual HSI pixels based on their feature information. HSI classification is an essential branch in HSI processing tasks which has already become a research hotspot among remote sensing fields [1,2,3].
During the past decades, many approaches have been proposed to classify HSIs. Early HSIs classification approaches mainly used spectral features for classification, such as SVM [8], decision trees [9], and multi-nomial logistic regression (MLR) [10]. As is well known, adjacent pixels of HSI are strongly correlated, which leads to a high probability that they belong to the same ground category. The classification methods merely based on spectral characteristics overlook the relevance of spatial information and the consistency of local feature information, resulting in an insufficient utilization of spatial contextual feature information. Therefore, aiming to further enhance their classification performance, many machine learning-based approaches for HSI classification tasks integrate both spectral and spatial information in their models. For example, Li et al. [11] proposed a spectral-spatial feature categorization method that combines subspace polynomial logistic regression and Markov random field (MRF), based on MLR to learn posterior probability distributions from spectral signatures and the subspace projection-based approach to represent noisy and height-mixing pixels. The spatial contextual information is then included by employing multi-level MRF prior. Notably, MRF usually exploits the continuity of neighboring image pixel labels in the probabilistic sense, and it establishes spatial context dependencies by specifying local conditional probabilities [12,13]. Shen et al. [14] proposed an approach for spectral-spatial feature information extraction by conducting complex 3D Gabor wavelet groups with various frequencies and directions. In [15], a pixel-level and object-level feature extraction method was first proposed by using a composite kernel scheme, which merges the features of these two levels into a composite kernel and feeds them into an SVM for classification, and then the voting fusion scheme is employed to determine the final categorization results. Besides, conditional random fields [16], extended attribute profiles [17], multiple kernels [18], and extinction profiles [19] have also been proposed. However, although the above traditional HSI classification approaches have achieved good categorization results, they are largely based on shallow learning and handcrafted features, relying on professional domain knowledge in the parameter setting stage. It is, therefore, difficult to obtain the deep-level abstract characteristic information contained in HSIs using these approaches.
Recently, with the rapid evolution of deep learning (DL) techniques in machine vision, audio processing, natural language, face recognition, and so on, DL-based techniques have been extensively introduced to remote sensing image processing tasks [1,2,20]. Some typical DL approaches have been widely employed in HSI classification, involving stacked auto-encoders, recurrent neural networks, deep belief networks, and CNNs, which significantly boost the performance of the HSI categorization task [21,22,23,24,25]. DL-based techniques are capable of automatically extracting HSI features from a low to high level and learning efficient and deeper abstract features. Chen et al. [26] exploited 3D CNNs to obtain spectral-spatial joint feature information from the raw data directly. Hang et al. [27] proposed an attention-aided CNN model that merges the attention modules into convolutional layers, making the CNN concentrate on more discriminative channels and spatial locations. Although deeper DL networks are capable of extracting more refined discriminative features, it is more difficult to train a deeper model due to the vanishing gradient phenomenon. Generally speaking, the deeper the DL network is, the easier it is to lead to gradient disappearance or explosion problems. ResNets [28] and DenseNets [29] effectively alleviate the vanishing gradient phenomenon. Based on ResNets and 2D CNNs, Lee et al. [30] utilized multi-scale convolutional filter groups to construct a nine-layer, fully connected CNN to realize the joint extraction of spectral-spatial features. In [31], two 3D CNN dense blocks with different kernel sizes, spatial densely block and spectral densely block, were proposed to acquire deeper spatial and spectral characteristics. Li et al. [32] proposed a deep multi-layer fusion DenseNet model, which utilizes 2D CNN and 3D CNN dense blocks to capture different levels of spatial and spectral signatures from the original HSI, respectively. In [33], a spatial and spectral signature extractor combining multi-scale 2D DenseNet with Bi-RNN is proposed to enhance the acquisition of complex spatial contextual feature information and the propagation of features among various convolutional layers via utilizing multi-scale 2D DenseNet.
Based on 3D CNNs and DenseNet, Ma et al. [34] developed a double branch model (DBMA) combined with 3D DenseNet blocks and multiple attention mechanism blocks, applying both attention mechanisms to emphasize important spectral and spatial components. Inspired by [34], a double branch and dual attention mechanism framework (DBDA) is developed, employing input patches of various sizes, to refine and spotlight the extracted feature maps by introducing different channel and spatial attention modules [35]. Since substantial spatial redundant information exists in traditional CNN-based networks, Chen et al. [36] first proposed to adopt octave convolution instead, which not only reduced redundant spatial information, but also improved performance through effective communication between high and low frequencies. In order to deeply mine discriminative spectral-spatial features and simultaneously decrease the spatial feature redundancy caused by CNNs, Tang et al. [37] developed a 3D octave convolution (3Doc-conv) model with both spectral and spatial attention networks by replacing the traditional CNN with octave convolution. Xu et al. [38] presented a multi-scale 3Doc-conv with spatial and channel attention to obtain the spatial contextual signatures at different scales via multi-scale octave convolution kernels. Wang et al. [39] designed a fast multi-scale capsule model with octave convolution, reducing the spatial redundancy parameters and obtaining competitive classification performance. Combining the characteristics of 2D CNNs and 3D CNNs, a hybrid spectral CNN model is proposed to use 2D CNN after the 3D CNN to capture more deep spatial representation features, reducing the complexity by using the 3D CNN alone [40]. Moreover, a cross-level spectral-spatial joint coding model was also introduced for the categorization of HSIs, which enhances the weight of small samples with classification difficulties by developing a category-proportional sampling strategy and a weighted loss method [41]. In [42], a model combined with global convolutional long short-term memory and global joint attention mechanism is introduced, utilizing a hierarchical balanced sampling method and a weighted softmax loss strategy to further overcome the problem of insufficient and unbalanced samples in HSIs.
Inspired by the above-mentioned works, here, we develop a new multi-scale spectral-spatial attention network framework combining 2D octave and 3D CNNs (MOCNN) for classification tasks. The spectral-spatial feature extraction module of our MOCNN is composed of two primary branch networks: the multi-scale DenseNet based on 3D CNNs (multi-scale 3D DenseNet) and the multi-scale 2D octave convolution network (multi-scale 2D octave) are, respectively, applied to spectral signatures and spatial contextual feature information extraction. For the spatial feature extraction sub-network, a multi-scale 2D octave is employed to adequately mine the spatial contextual feature information, while decomposing the obtained feature map into low-frequency (LF) and high-frequency (HF) components to decrease the redundancy of spatial feature information. For the spectral feature extraction sub-network, the multi-scale 3D DenseNet is utilized to sufficiently mine the spectral signatures at various scales and simultaneously merge the spectral features in shallow and deep convolutional layers to improve the reuse of spectral characteristics. Additionally, the spectral attention mechanism module is used to allocate proper weight values for each spectral band while suppressing insignificant spectral bands to mitigate the effect of redundant HSI bands in the classification. The channel attention mechanism module is used in the two feature extraction sub-networks to enhance the interactions of the feature map information among feature channels and improve the feature extraction capability. Finally, a sample balancing strategy based on the weighted cross-entropy loss function (WCL) is used to address the problem of sample imbalance. The MOCNN proposed in the present paper is inspired from [43]. The major contributions of this study are outlined below:
  • Based on 2D octave, a multi-scale 2D octave convolution sub-network is proposed to capture spatial feature information. It can not only reduce the spatial feature information redundancy, but also extract complex spatial structure information adequately.
  • A multi-scale DenseNet based on 3D CNNs is exploited to adequately explore the discriminative spectral signatures at various scales, while fusing the spectral signatures in both shallow and deep convolutional layers to enhance the reuse of spectral features.
  • Two types of attention models are utilized to highlight the features containing important information to boost the spectral-spatial feature capture capability of the model. Furthermore, in order to address the problem of sample imbalance, a sample balancing strategy based on the WCL is employed to achieve the balance of the weight probabilities for each category, resulting in the fact that the model focuses more on categories with scarce training samples.
The organization of the remaining parts of this article is described below. The related materials and methods (i.e., the proposed MOCNN model, 2D octave, 3D DenseNet, attention mechanism models, and balanced sampling strategy) are briefly presented in Section 2. The parameter configuration, experimental setup, extensive experimentation and discussion, analysis of experimental results at various sample ratios, and ablation experiments are detailed in Section 3. Ultimately, Section 4 provides conclusions and directions regarding future research.

2. Materials and Methods

The proposed MOCNN model primarily consists of three modules: a branch network of the multi-scale 2D octave for spatial contextual feature information extraction, a branch network of the multi-scale 3D DenseNet for exploring the spectral feature information, and a spectral-spatial feature fusion sub-network. The flowchart of the network of the MOCNN approach for HSI classification is illustrated in Figure 1. For spatial feature extraction, after dimensionality reduction of the raw HSI data by principal component analysis (PCA), two 2D octave convolutional networks (where each network contains four 2Doc-conv) are utilized to decompose the feature map into LF and HF components to decrease the redundancy of spatial feature information. Furthermore, the multi-scale 2D octave with two distinct convolution kernel sizes is used to further capture complex spatial contextual feature information of various scale sizes. In multi-scale 3D DenseNet, firstly, a band attention mechanism (BAM) is used to band select and weight each spectral band, to improve the acquiring of characteristic information contained in vital spectral bands and decrease band redundant feature information. Subsequently, the multi-scale 3D DenseNet is exploited to adequately explore the spectral characteristic information at various scales, while fusing the spectral features among various layers in the shallow and deeper convolutional layers. In the final part of the above two sub-networks, the channels of the feature maps connected by concat operation are weighted using the efficient channel attention mechanism (ECA) for enhancing the information capture capability of the spectral and spatial feature channels. Lastly, in the feature fusion sub-network, the final feature maps gained via connecting these two sub-networks are fed through the fully connected layer to yield the new spectral and spatial fusion features. The fused new features are then imported to the softmax layer to gain prediction results.

2.1. 2D Octave Convolution

The image information is capable of being divided into the LF information component and the HF information component. Among them, the LF information denotes the relatively smooth or homogeneous regions of the image, which have relatively slow image gray value transformations. The HF information refers to the image boundary information and the detail information of small areas of the image, where the image gray value transforms relatively dramatically. Likewise, the output characteristic maps in the CNNs can be considered as the characteristic signal made up of LF and HF information, and each position in the spatial dimension of the characteristic map stores its own feature signal individually. Octave convolution stores the LF information that transforms slowly between adjacent positions together in a low-resolution tensor by sharing positions so as to decrease the resolution of LF features and decrease the superfluity of spatial characteristics and operational complexity. Hence, 2D octave convolution (2Doc-conv) is applied to obtain spatial contextual feature in this work, which can effectively minimize the spatial information superfluity and lower the computational effort of the model. The schematic of the overall structure of 2Doc-conv network is depicted in Figure 2.
Specifically, let us suppose that the input feature maps of the 2Doc-conv are denoted as X=( X h , X l ), and the output at the m-th layer of the octave convolution is denoted by Y m =( Y m h , Y m l ), where the upper label h and the upper label l stand for the corresponding feature maps belonging to HF and LF information, respectively. Notice that L L and H H denote the exchange of information between intra-frequencies. H L and L H express the exchange of information between HF and LF. As the raw HSI contains abundant spatial and spectral signature detail information, it can be regarded as HF feature data. As displayed in Figure 2, the 2Doc-conv network is made up of four 2Doc-convs, an averaging pooling (AP) operation, and an upsample layer. The output characteristic information of the first layer in this 2Doc-conv is expressed as:
Y 1 h = C o n ( X h ) ,
Y 1 l = C o n ( A p o o l ( X h ) ) ,
where C o n ( . ) and A p o o l ( . ) represent the 2D convolution and AP operation, respectively. In the second layer of the 2Doc-conv network, it is mainly primarily utilized to complete the update of same-frequency (HF to HF or LF to LF) feature information and the transformation between different-frequency (HF to LF or LF to HF) ones. Thus, the output features of the second layer in this 2Doc-conv can be formulated as:
Y 2 h = C o n ( Y 1 h + U p s a m p l e ( Y 1 l ) ) ,
Y 2 l = C o n ( Y 1 l + C o n ( A p o o l ( Y 1 h ) ) ,
where U p s a m p l e ( . ) represents the upsample operation based on bi-linear interpolation. For reducing the redundancy of feature information, the HF information acquired in the second layer is downsampled, and then the LF feature information gained in this layer is merged to yield a novel feature Y p o o l , whose calculation formula is:
Y p o o l = C o n ( A p o o l ( Y 2 h ) + C o n ( Y 2 l ) ) .
Likewise, the final output feature information of the third and fourth layer, Y 3 h , Y 3 l , Y 4 h , and Y 4 l , in this network can be updated by the communication mechanism between intra-frequency and inter-frequency, respectively. Finally, Y o can be represented by:
Y o = C o n ( C o n ( Y 4 h ) + U p s a m p l e ( Y 4 l ) ) .
To sum up, based on 2Doc-conv, the slowly varying LF feature information is stored in the tensor with LF resolution, and then the LF feature resolution is lowered by sharing information among adjacent positions. Compared with traditional CNNs, octave convolution has two obvious advantages, as it can not only reduce spatial redundancy but also extend the receptive field of the network. Therefore, the 2D octave is exploited instead of CNNs to improve the categorization performance of HSI in this work.

2.2. 3D DenseNet

Generally, traditional CNN models increase the depth of the network model by stacking convolutional layers, thereby improving the classification accuracy. However, the problem of gradient descent or gradient explosion will occur when too many convolutional layers are stacked. Each convolutional layer of DenseNet gains external input information from all previous convolution layers and delivers their own feature information to all succeeding convolution layers. The gradient value gained in each convolution layer is the gradient addition from the previous convolution layers, which can effectively alleviate the gradient descent problem [29]. Compared with ResNet [28], the DenseNet structure with feature multiplexing can utilize hierarchical features more efficiently and reinforce feature transfer among convolution layers.
All convolutional layers in 3D DenseNet are connected by directly skip-connecting to achieve maximum information transfer among layers of the network. In short, the input of each convolutional layer is the concatenation of the outputs of all previous convolution layers. The first 3D convolutional layer processes raw pixel data of size a × a × B (where a × a denotes the spatial size and B denotes the number of bands) to generate m 0 feature maps of size a × a × b. These maps serve as input to dense blocks. As illustrated in Figure 3, the input of the dense block is of the size a × a × b with m 0 feature maps, where the subscript indicates the amount of convolutional layers within the dense block. The convolution layer embedded within the dense block is represented as D(.), after each convolution layer applies m kernels of size k × k × c to extract rich spectral information, the output feature map of each layer is size of a × a × b. The number of output spectral feature maps of the t-th layer can be formulated as m 0 + ( t 1 ) × m , where m 0 means the number of channels contained in the initial feature map. The input to the t-th layer of the densely connected model can be expressed as:
H t = D ( H 0 , H 1 , H 2 , , H t 1 ) t N + ,
where D denotes a function module, which includes convolutional layer, batch normalization (BN), and activation function layer. N + denotes a positive integer.

2.3. Attention Mechanism Module

The attention mechanism can be used to focus on a certain part of the HSI, assign independent and appropriate weights, weaken redundant and useless information, and highlight effective features that are beneficial for classification. This section will describe the BAM and ECA attention modules in detail.

2.3.1. Spectral Attention Mechanism

The bands of hyperspectral images are redundant and noisy, whereas a band attention module has the ability to conduct band selection and weight each band, which can mitigate the negative effects of band superfluity in HSI classification [44]. The schematic of the structure of BAM is shown in Figure 4. The BAM module primarily contains three 3 × 3 2D convolutional layers, two 1D convolutional layers, two pooling layers, and an activation function layer, where the number of channels of 2D convolutional layers is 16, 32, and 32, respectively. In this attention model, after using a 2D convolutional layer for expanding the receptive field, the resolution is reduced by pooling layers to obtain global information in the spatial domain, and then a 1D convolutional layer is employed to further learn the non-linear correlation among the bands. Ultimately, a sigmoid function is adopted to obtain the spectral band weight vector. Notably, the sizes of the two 1D convolution kernels are 1 × 1 × 32 × B/r and 1 × 1 × B/r × B, in which B denotes the number of spectral bands and r is a hyper-parameter that controls the degree of information aggregation layers in the 1D convolution layer. The forward propagation expression of BAM is represented as:
U o = ψ 2 ( W 12 ψ 1 ( W 11 f G ( ψ 1 ( ( W 23 ψ 1 ( W 22 f P ( ψ 1 ( W 21 U i ) ) ) ) ) ) ) ) ,
where U i and U o denote the HSI data input to the BAM and the output of the BAM, W i j denotes the j-th i-dimensional convolutional kernel matrix in the BAM, f P denotes an AP layer for resolution reduction, and f G denotes a global AP layer for the sufficient fusion of the spatial feature information contained in the generated feature maps to form a band mask. Moreover, ψ 1 , ψ 2 denote the Relu and sigmoid activation functions. After obtaining the spectral band weighted vector by the ψ 2 function, the inner product operation is performed with the input of BAM to obtain more representative spectral band features.

2.3.2. Channel Attention Mechanism

The proper cross-channel feature interaction and avoidance of dimensionality reduction for channel attention learning can enhance the feature learning capability of the model while lowering its complexity. ECA is a lightweight and local cross-channel feature interaction attention model without dimensionality reduction [45]. The schematic of the structure of ECA is given in Figure 5. It generates channel attention weight parameters via 1D convolution, with the size R of the convolution kernel being self-adaptively determined by a non-linear mapping of channel dimensions C, which leads ECA to significantly lower the model complexity while sustaining cross-channel interaction. Its computational expression can be presented as:
Q C = P ( I C ) = 1 S × S i S j S I C ( i , j ) ,
where I C ( i , j ) refers to the characteristic information of the c-th channel from the input feature maps at the ( i , j ) coordinate position. Q C denotes the feature information of the c-th channel from the feature map obtained after the AP layer of I C . The size R of the 1D convolution kernel (the coverage of local cross-channel interactions) is expressed as:
R = α ( C ) = log 2 ( C ) β + b β o d ,
where r o d represents taking the nearest odd value of r and the values of parameters b and β are given as 1 and 2, respectively. Hence, after the values of parameters b, β are defined, the value of R is exclusively defined adaptively by the value of channels C. Finally, the corresponding weights of the channels can be calculated as:
w C = ψ 1 ( j = 1 R W j Q c j ) , Q c j θ c R ,
where ψ 1 denotes the sigmoid activation functions and θ c R represents the set of R adjacency channels of Q C .

2.4. Balanced Sampling Strategy

The WCL can balance the weight probabilities of different categories, which is a common method to address the imbalance problem [46]. To tackle the problem of sample imbalance in this paper, the WCL is applied to lower the weights of easily classified sample categories, while making the model more focused on categories with small samples and misclassification. Based on this balanced sample strategy, the composite loss function of the MOCNN model can be represented as:
L i m ( y , P ( Z i m ) ) = i = 1 n W i y l o g ( P ( Z i m ) + ( 1 y ) l o g ( 1 P ( Z i m ) ) ) ,
L s p a ( y , P ( Z s p a ) ) = i = 1 n y l o g ( P ( Z s p a ) + ( 1 y ) l o g ( 1 P ( Z s p a ) ) ) ,
L s p e ( y , P ( Z s p e ) ) = i = 1 n y l o g ( P ( Z s p e ) + ( 1 y ) l o g ( 1 P ( Z s p e ) ) ) ,
L = L s p a ( y , P ( Z s p a ) + L s p e ( y , P ( Z s p e ) ) + η L i m ( y , P ( Z i m ) ) ) ,
where Z i m , Z s p a , and Z s p e denote the logits of the spatial-spectral feature fusion, spatial feature extraction, and spectral feature extraction sub-network in the proposed model, respectively, L denotes the composite loss function of the proposed MOCNN model, L i m , L s p a , and L s p e denote the loss functions of the corresponding sub-networks, separately, y denotes the true label corresponding to the i-th training sample, η denotes the parameter that controls the weight of L i m , which is set empirically to 0.9, and W i denotes the weight parameter corresponding to the category loss, which can be calculated from:
W i = N T i ,
where N stands for the number of occurrences of the median across all categories and T i denotes the frequency of category i, and it represents the proportion of the number of categories i to the sum of all pixels in the training sample. Thus, T i is expressed as:
T i = M i j M j ,
where M i refers to the number of samples of category i. During each iteration of training, the network loss corresponding to the training set is updated by the counter-propagation operation per layer, and the optimal network is selected in a finite number of iterations.

3. Experimental Results and Discussion

3.1. Experimental Datasets Description

In this section, to verify the performance of the MOCNN model and contrast it with other approaches, four HSI datasets are applied in the experiments, i.e., the Pavia University (UP), Salinas Valley (SV), Indian Pines (IN), and Zaoyuan (ZY) datasets. The reference maps (i.e., false-color images and ground truth) for the four HSIs are presented in Figure 6. Three common quantitative indicators, namely, overall accuracy (OA), average accuracy (AA), and Kappa coefficient (Ka), are utilized to evaluate the classification performance of all methodologies [47]. In particular, OA means the percentage of correct categorization of all pixels, AA refers to the mean classification precision of the whole category, and Ka takes into account the influence of uncertainty on the classification results and reflects the coherence between classification outcomes and the ground truth. The values of OA, AA, and Ka vary from 0 to 1, where closer to 1 implies better classification outcomes of the model.
Pavia University dataset (UP): The UP contains 610 × 340 pixels and encompasses 103 spectral channels after abandoning noisy spectral channels, covering wavelengths varying from 430 to 860 nm, whose spatial resolution (SR) is up to 1.3 m. It encompasses nine classes of labeled samples of land cover objects for classification.
Salinas Valley dataset (SV): The SV contains 512 × 217 pixels and encompasses 204 spectral channels after abandoning 20 noisy spectral channels, covering wavelengths varying from 360 to 2500 nm with 3.7 m SR. It encompasses 16 classes of various land cover labeled samples for classification.
Indian Pines dataset (IN): The IN contains 145 × 145 pixels and encompasses 220 spectral channels. It covers wavelengths varying from 400 to 2500 nm, while its SR is 20 m. It encompasses a total of 16 classes of land cover labeled samples for classification.
Zaoyuan dataset (ZY): The ZY was obtained by OMIS sensor in 2001 in the Zaoyuan zone of China. It encompasses 80 spectral channels after discarding 48 noisy spectral channels and covers wavelengths varying from 400 to 1700 nm. The spatial size of ZY is 137 × 202 with 23,821 labeled pixels. It encompasses a total of eight classes of land cover labeled samples for classification.
During each experimental execution, the selected experimental samples (training samples, validation sets, and test samples) are taken through a random sampling method. Table 1, Table 2, Table 3 and Table 4 provide the categories information and corresponding sample numbers for each HSI dataset.

3.2. Parameter Setting

In this section, the configuration of the model parameters is introduced at length according to the flowchart of the network of the proposed MOCNN model given in Figure 1. The network parameters settings for 2D octave and 3D DenseNet in the MOCNN model are given in Table 5. For spatial feature extraction, after the input raw 3D cube is down-sampled by PCA, the size of the resulting data cubes is 23 × 23 × 9, 29 × 29 × 9, 21 × 21 × 9, and 27 × 27 × 9 for UP, SV, IN, and ZY, respectively. After that, the reduced data cubes are input to the multi-scale 2D Octave sub-network for spatial feature extraction, where the detailed parameters of the two 2D Octaves in this sub-network configuration are shown in Table 5. For the UP, SV, IN, and ZY, the two convolution kernel sizes k s p a in multi-scale 2D octave are set to 5 and 7, 3 and 9, 5 and 7, 3 and 7, respectively. The quantities of convolutional kernel channels m s p a are given as 32, 64, 32, and 16 for each 2D octave layer, respectively. For spectral feature extraction, corresponding to the above HSI dataset, the size of the input raw HSI data is 7 × 7 × 103, 7 × 7 × 204, 7 × 7 × 220, and 3 × 3 × 80, respectively. After the channels are weighted by a BAM module, the data are fed into a multi-scale 3D DenseNet to learn the spectrum signatures at various scales. Likewise, the detailed parameters of the two 3D DenseNet in this sub-network configuration are presented in Table 5, where the two convolution kernel size k s p e are set to 5 and 7, 5 and 7, 5 and 7, 3 and 5, on UP, SV, IN and ZY, respectively. The number of convolutional kernel channels m s p e for each convolutional layer and other detailed configuration settings are also given in Table 5. Besides, the value of the convolution kernel k s p e 0 depends on the value of bands B. When the value of B is even, k s p e 0 takes ceil(B/2 − k s p e 0 ). Conversely, k s p e 0 takes ceil (B/2 − k s p e 0 ) + 1. The ceil(.) represents the upward rounding calculation function.
In addition, the parameter settings on four different HSI datasets concerning the learning rate, dropout, batch size, and epochs are shown in Table 6. Due to space considerations, the parameter settings are not described in detail here, and the optimal parameters derived from the summary of the experiments are given directly.

3.3. Experimental Setup

For assessing the validity and superiority of the MOCNN approach, the DL-based classifiers CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], and HRAM [49] are compared with the proposed method.
(1) CDCNN: The CDCNN approach realizes the joint extraction of spectral-spatial features through utilizing multi-scale convolutional filter groups and exploiting the modules of ResNet to significantly improve the depth of the network.
(2) FDSSC: The FDSSC approach utilizes two 3D CNN dense blocks with different kernel sizes, spatial densely block and spectral densely block, to capture deeper spatial and spectral characteristics, respectively.
(3) DBMA: The DBMA approach is a double branch model combined with 3D DenseNet blocks and multiple attention mechanism blocks, applying both attention mechanisms to emphasize important spectral and spatial components.
(4) DBDA: Although DBDA is analogous to the DBMA framework, DBDA employs different size input patches and introduces different channel and spatial attention modules. Moreover, the activation function Mish is introduced to accelerate counter-propagation and prevent overfitting in severely restricted samples.
(5) SSAN: The SSAN approach is a spectral-spatial two-branch two-attention framework model, which adds spectral and spatial attention mechanisms on the basis of Bi-RNN and CNN, respectively.
(6) TriCNN: The TriCNN approach is a three-branch network model based on 3D CNN using various scale convolutional kernels of 1 × 1 × 3, 3 × 3 × 1, and 3 × 3 × 3 sizes to extract spectral, spatial, and spectral-spatial features, respectively, and then fuse the different features obtained from the three sub-networks by feature flattening and concatenation.
(7) 3DOC-CNN: The 3DOC-CNN approach adopts 3Doc-conv to acquire the spectral-spatial characteristics, and introduces attention networks to spotlight the more meaningful characteristics in HSI.
(8) HRAM: The HRAM approach employs a hierarchical residual network to extract spatial and spectral characteristics at the granularity level, utilizing a two-branch structure in parallel with the corresponding convolution kernel. Besides, to boost the discriminative power of the model, it exploits attention mechanisms to assign adaptive weights to spatial and spectral features at various scales.
All experiments were run on a 2080Ti GPU, i9-9900K CPU, 128GB RAM, python 3.6, using tensorflow-gpu 1.14 for implementation. The outcomes reported from all experiments are the average and variance of the results of twenty Monte Carlo runs. To guarantee impartial comparison, the experimental arguments of each classifier are set with default settings of the original paper. In the following, the outcomes of the quantitative comparison experiments between MOCNN and all compared approaches on UP, SV, IN, and ZY are presented in Table 7, Table 8, Table 9 and Table 10. Among all classification accuracy comparisons, the highest accuracy is marked in bold. Besides, we visualize the classification maps in Figure 7, Figure 8, Figure 9 and Figure 10.

3.4. Classification Maps and Results

The accuracy metrics, including OA, AA, and Ka, for the UP dataset from our proposed MOCNN approach and several other DL-based comparison models are provided in Table 7, while corresponding classification maps are given in Figure 7.
From Table 7, it can be seen that the MOCNN approach achieves the optimal classification precision, with values of 98.92% for OA, 97.94% for AA, and 98.57% for Ka. The CDCNN approach based on multi-scale convolutional filters and ResNet structure results in the worst accuracy, with 92.86% OA, 91.41% AA, and 90.48% Ka. The reason may be that the limitation of this network structure design leads to a weak feature extraction ability of the model under the limited training samples. Owing to the fact that the Bi-RNN model based on the attention mechanism effectively enhances the extraction of spectral features, the OA of the SSAN approach is 2.08% higher than the CDCNN. However, since SSAN employs the traditional CNN network for spatial feature extraction, it lacks the adequate exploitation of spatial features from convolutional layers of various depths. The TriCNN increases the accuracy of OA by 2.55% and 0.47% over CDCNN and SSAN, respectively, which may be owing to three convolutional kernels of various scales utilized by TriCNN and its strengthened ability of the network to explore complex features. Nevertheless, TriCNN still also lacks the utilization of characteristics across various convolutional layers. The HRAM effectively improves the utilization of hierarchical features by using hierarchical ResNets, and its OA accuracy enhances by 2.76% compared with TriCNN. Similarly, the FDSSCN, DBMA, and DBDA approaches employ 3D DenseNet to improve the reuse and propagation of hierarchical characteristic information and further enhance the capability of the network for feature extraction. Compared with SSAN, the OA accuracy of FDSSCN, DBMA, and DBDA approaches improved by 3.39%, 3.06%, and 3.43%, respectively. Comparing our method with 3DOC-CNN using single-scale octave convolution alone, OA improves by 1.53% and the accuracy of the two metrics (AA and Ka) is also superior to that of 3DOC-CNN. In spite of the fact that the proposed MOCNN method achieves worse results than 3DOC-CNN in certain categories, such as Meadows and Bricks, while the classification accuracy is the highest in other categories.
It is evident from the corresponding Figure 7 that the classification maps of the MOCNN and DBDA approaches are significantly superior to other methods, with better homogeneity and smoother classification maps for the “Baresoil” category than those generated by the other methods. Additionally, our proposed approach contains fewer misclassified pixels, and noticeably better classification maps in the “Gravel” category. That is because the MOCNN combining multi-scale 2D octave convolution and multi-scale 3D DenseNet not only takes into account the utilization of features across hierarchical convolution layers, but also sufficiently exploits the complex characteristics at various scales.
As shown in Table 8, the MOCNN approach realizes the best categorization results with 98.69% OA, 99.05% AA, and 98.54% Ka. It is notable that CDCNN has the worst classification accuracy, which is 14.92% lower than our method. This also indicates the fact that limitations of CDCNN may exist in the design of the network structure, failing to exploit the superiority of multi-scale convolutional filters and residual network structures to mine distinguishing characteristics at various scales from complicated scenes. Although TriCNN and SSAN methods are both traditional CNN-based in spatial feature extraction, the OA accuracy of TriCNN is 96.59%, which is 1.92% higher than the value of 94.67% gained by SSAN. In the corresponding categorization maps in Figure 8, TriCNN has significantly fewer misclassification noise points in terms of spatial details than SSAN. The reason for this is that TriCNN adopts a three-branch network with convolutional kernels of various scales, which greatly boosts the network’s ability to capture complex characteristics. The HRAM method based on hierarchical ResNet achieves an OA accuracy of 97.21%, which exceeds the conventional CNN-based TriCNN and SSAN, respectively, by 0.62% and 2.54%. In addition, in comparison to FDSSC, DBMA, and DBDA based on single-scale DenseNet, our method’s classification accuracy OA is 0.86%, 2.78%, and 2.20% higher, respectively. This may be explained by the fact that MOCNN uses not only DenseNet structures but also multi-scale convolutional kernels, which further facilitate the extraction of complex characteristic information at various scales. Notably, despite the fact that MOCNN achieves optimal categorization accuracy in many categories, certain categories in other comparison methods, such as FDSSC in categories “Brocoli-green-weeds-1”, “Brocoli-green-weeds-2”, and “Vinyard-vertical-trellis”, also obtain 100% categorization accuracy. It demonstrates that other DL-based models also have strong feature learning capabilities and may assign more weights to certain categories to realize accurate classification of that category.
Meanwhile, as seen in the corresponding classification map in Figure 8, the MOCNN approach provides the most accurate categorization result map and is smoother in homogeneous regions. Its causes can be attributed to the following: first, the densely connected based structure merges features from convolutional layers of various depths, enhancing the reuse and propagation of feature information. Then, the use of multi-scale convolutional kernels facilitates the extraction of complex characteristic information at various scales. In addition, two attention mechanism modules are utilized to highlight important spatial feature information and reinforce the extraction of feature channel information in the network, respectively.
As shown in Table 9, MOCNN achieves the optimal categorization performance in OA, AA, and Ka, achieving 98.58%, 98.29%, and 98.39%. Compared with the suboptimal classification results, our method improves over the FDSSC by 0.42%, 0.64%, and 0.48%. Compared with HRAM and TriCNN, the accuracy improvement of AA by MOCNN is more remarkable, i.e., 6.1% and 2.96% higher, respectively. This is primarily attributed to employing a WCL-based sample balancing strategy that allows MOCNN to allocate appropriate weights to each category and to pay more attention to the categories with small sample sizes. Furthermore, it is clear from Table 3 that the sample distribution of the IN is highly unbalanced. Certain categories, such as “Soybean-mintill” and “Corn-notill”, have 295 and 172 samples. Yet, there are also some categories with only a few or dozens, such as “Oats”, “Grass-pasture-mowed”, “Alfalfa”, and “Corn”, which have only 3,4, 6, and 29 samples, respectively. Nonetheless, the MOCNN also achieved an OA accuracy of 100.00%, 95.00%, 97.06%, and 100.00% for these categories, respectively. It illustrates that the compound loss function based on WCL proposed in this paper, allocating appropriate loss weights to various categories, can make the model more focused on the categories with small samples, which effectively alleviates the problem of poor classification results due to sample imbalance. As we can see from Figure 9, although the MOCNN method has few misclassified pixels on the categories Corn-notill, Corn-mintill, and Soybean-mintill, the overall classification results are relatively less noisy than the other approaches. Moreover, the smoothing in homogeneous areas (such as Corn-notill and Grass-pasture-mowed) is significantly better than the other methods.
As seen in Table 10, the MOCNN approach still realizes the best categorization results with 98.16% OA, 98.34% AA, and 97.59% Ka. Although the categorization accuracy of MOCNN is not the highest in the “Grape”, “Peach”, “Corn”, and “Terrace/Grass regions”, the accuracy of our method exceeds 97.04% in each of these categories. This illustrates well the effectiveness of our method in extracting discriminable features between different categories. Moreover, we can see from Figure 10 that the classification map of the CDCNN approach with the lowest categorization accuracy has a large number of misclassified pixels and pepper noise. The OA accuracy of SSAN and 3DOC are 96.06% and 96.31%, respectively, which are both better than 84.14% of CDCNN. In contrast, SSAN and 3DOC-CNN also have fewer mislabels, but the classification results in the “Pear” and “Dry vegetable” regions are still not satisfactory. Conversely, in the densely connected based approaches, i.e., FDSSC, DBMA, DBDA and MOCNN, categorization results are significantly better than the previous two methods. This illustrates that the complex spatial structure information of HSI can be effectively exploited by using the complementary yet related information between the features of different convolutional layers. Besides, the ResNet-based HRAM method achieves 97.59% categorization accuracy with less classification noise on the categorization map. Nonetheless, since it adopts the single-scale convolutional kernel, it suffers from limitations in capturing boundary detail information. However, the MOCNN method not only produces a smoother appearance in homogeneous areas but also has clear boundaries and preserves edge detail information well, which also demonstrates that the proposed MOCNN based on multi-scale convolutional structure and attention network mechanism can adequately capture complex characteristic information at various scales.

3.5. Discussion

In this section, we first discuss the classification performance of the proposed method and other research methods with varying training sample sizes. Following that, we discuss the effects of different functional modules on the classification performance of the model in the ablation experiments.

3.5.1. Performance under Different Numbers of Training Samples

DL is a data-driven approach that relies on the availability of massive amounts of labeled data. Therefore, to further investigate the categorization performance under various numbers of training samples, we conducted extensive experiments on UP, SV, IN, and ZY using different percentages of training samples, respectively. Figure 11 exhibits the corresponding experimental results. Notably, since the highest values of CDCNN on the IN and ZY are lower than the minimum coordinate values on the corresponding sub-figures, the plot of CDCNN is not shown in the corresponding sub-figures. As anticipated, the categorization precision of the DL-based approach improves along with the increase in the number of training samples. As seen in Figure 11, it is evident that our method can still achieve superior classification precision with limited training samples. Hence, our method can save a great deal of labor and cost in labeling samples.

3.5.2. Ablation Experiment

In this section, ablation experiments were designed to validate the effectiveness of the spectral attention module BAM, the channel attention mechanism ECA, multi-scale 3D DenseNet, and multi-scale 2D octave. Specifically, no-ECA denotes that the proposed model does not use the ECA module, and only the BAM module is employed. Similarly, no-BAM denotes that the proposed method utilizes only the ECA module. MONA indicates that ECA and BAM attention networks are not used in the proposed MOCNN approach. The MCNN denotes that the proposed MOCNN method only adopts multi-scale 3D DenseNet for spectral feature extraction, and there is no branch network employed for spatial information extraction. The MONN denotes that the proposed MOCNN method only adopts multi-scale 2D octave for feature extraction, and lacks the network for the extraction of spectral information.
From Figure 12, the classification accuracy of the MONA approach without using any attention mechanism is significantly lower than that of the no-ECA and no-BAM approaches on the four datasets, thus fully proving the effectiveness of the ECA and BAM attention modules. In addition, it is apparent that the classification outcomes of MONN and MCNN are significantly lower than MOCNN, which is due to the fact that MONN based on multi-scale 2D octave can only extract spatial feature information at various scales, while lacking access to spectral information. Similarly, MCNN based on multi-scale 3D DenseNet can only withdraw spectral signatures at various scales and lacks the mining of complex spatial information. It also further demonstrates that the single extraction network, either for spectral features alone or for spatial feature information alone, fails to simultaneously sufficiently exploit the spectral-spatial feature information that is favorable for HSI categorization.

4. Conclusions

In this study, a new multi-scale spectral-spatial attention network combining 2D octave and 3D CNN is proposed for HSI classification. Concerning spatial features, based on 2D octave, multi-scale 2D octave is proposed to obtain complicated spatial characteristic information. It can not only adequately mine the spatial feature information under the complex structure by using convolutional kernels of various scales, but also decompose the obtained feature map into LF and HF components to lessen the redundancy of spatial feature information. Concerning spectral features, we use the multi-scale 3D DenseNet to sufficiently extract discriminative spectral characteristics at various scales, while fusing spectral features in both shallow and deeper convolutional layers to enhance the transmission and reuse of feature information among various convolutional layers. Besides, Two attention models (BAM and ECA) are used to improve network performance. Among them, BAM is exploited to assign proper weight values for each spectral band while suppressing insignificant spectral bands to alleviate the effect of redundant HSI bands in the classification. ECA is used in the two feature extraction sub-networks to enhance the interactions of feature information among feature channels and boost the feature extraction capability. Moreover, a sample balancing strategy based on WCL is applied to address the problem of sample imbalance. Experimental outcomes indicate that the proposed MOCNN approach outperforms several other compared approaches for classification.
In the future, we will develop semi-supervised or unsupervised HSI classification methods able to operate in scenarios with limited and imbalanced samples.

Author Contributions

L.L. and S.Z. proposed the core framework design idea of MOCNN network. L.L., S.Z., and Z.C. participated in conducting the code implementation, experimental analysis, and manuscript writing. J.L. and A.P. provided valuable advice on the methodology and carefully modified the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the National Natural Science Fund of China (T2225019), the Jiangxi Province Key Subject Academic and Technical Leader Funding Project (20225BCJ23019), the Jiangxi Provincial Natural Science Foundation (20224BAB202007 and 20224ACB202002), the Key R & D Program of Hunan Province (2019SK2102), and the China Scholarship Council.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNconvolutional neural network
HSIhyperspectral image
MLRmulti-nomial logistic regression
MRFMarkov random field
DLdeep learning
3Doc-conv3D octave convolution
MOCNNmulti-scale spectral-spatial attention network
framework combining 2D octave and 3D CNNs
multi-scale 3D DenseNetmulti-scale DenseNet based on 3D CNNs
multi-scale 2D octavemulti-scale 2D octave convolution network
LFlow frequency
HFhigh frequency
PCAprincipal component analysis
BAMband attention mechanism
ECAefficient channel attention mechanism
WCLweighted cross-entropy loss function
APaveraging pooling
BNbatch normalization
2Doc-conv2D octave convolution
OAoverall accuracy
AAaverage accuracy
KaKappa cofficient
UPPavia University dataset
SVSalinas Valley dataset
INIndian Pines dataset
ZYZaoyuan dataset
SRspatial resolution

References

  1. He, L.; Li, J.; Liu, C.; Li, S. Recent Advances on Spectral–Spatial Hyperspectral Image Classification: An Overview and New Guidelines. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1579–1597. [Google Scholar] [CrossRef]
  2. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep Learning for Hyperspectral Image Classification: An Overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef] [Green Version]
  3. Thoreau, R.; Achard, V.; Risser, L.; Berthelot, B.; Briottet, X. Active Learning for Hyperspectral Image Classification: A comparative review. IEEE Trans. Geosci. Remote Sens. Mag. 2022, 10, 256–278. [Google Scholar] [CrossRef]
  4. Yang, X.; Yu, Y. Estimating Soil Salinity Under Various Moisture Conditions: An Experimental Study. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2525–2533. [Google Scholar] [CrossRef]
  5. Liang, L.; Di, L.; Zhang, L.; Deng, M.; Qin, Z.; Zhao, S.; Lin, H. Estimation of crop LAI using hyperspectral vegetation indices and a hybrid inversion method. Remote Sens. Environ. 2015, 165, 123–134. [Google Scholar] [CrossRef]
  6. Makki, I.; Younes, R.; Francis, C.; Bianchi, T.; Zucchetti, M. A survey of landmine detection using hyperspectral imaging. ISPRS J. Photogramm. Remote Sens. 2017, 124, 40–53. [Google Scholar] [CrossRef]
  7. Tao, R.; Zhao, X.; Li, W.; Li, H.C.; Du, Q. Hyperspectral Anomaly Detection by Fractional Fourier Entropy. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2019, 12, 4920–4929. [Google Scholar] [CrossRef]
  8. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  9. Brodley, C.; Friedl, M. Decision tree classification of land cover from remotely sensed data. Remote Sens. Environ. 1997, 61, 399–409. [Google Scholar]
  10. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Semisupervised Hyperspectral Image Segmentation Using Multinomial Logistic Regression with Active Learning. IEEE Trans. Geosci. Remote Sens. 2004, 48, 4085–4098. [Google Scholar] [CrossRef] [Green Version]
  11. Li, J.; Bioucas-Dias, J.M.; Plaza, A. Spectral–Spatial Hyperspectral Image Segmentation Using Subspace Multinomial Logistic Regression and Markov Random Fields. IEEE Trans. Geosci. Remote Sens. 2011, 50, 809–823. [Google Scholar] [CrossRef]
  12. Tariq, A.; Shu, H. CA-Markov Chain Analysis of Seasonal Land Surface Temperature and Land Use Land Cover Change Using Optical Multi-Temporal Satellite Data of Faisalabad, Pakistan. Remote Sens. 2020, 12, 3402. [Google Scholar] [CrossRef]
  13. Dimov, I.; Maire, S.; Todorov, V. An unbiased Monte Carlo method to solve linear Volterra equations of the second kind. Neural. Comput. Appl. 2022, 34, 1527–1540. [Google Scholar] [CrossRef]
  14. Shen, L.; Jia, S. Three-dimensional Gabor wavelets for pixel-based hyperspectral imagery classification. IEEE Trans. Geosci. Remote Sens. 2011, 49, 5039–5046. [Google Scholar] [CrossRef]
  15. Duan, P.; Kang, X.; Li, S.; Ghamisi, P.; Benediktsson, J.A. Fusion of Multiple Edge-Preserving Operations for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 10336–10349. [Google Scholar] [CrossRef]
  16. Zhong, P.; Wang, R. Modeling and classifying hyperspectral imagery by CRFs with sparse higher order potentials. IEEE Trans. Geosci. Remote Sens. 2011, 49, 688–705. [Google Scholar] [CrossRef]
  17. Mura, M.D.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of hyperspectral images by using extended morphological attribute profiles and independent component analysis. IEEE Trans. Geosci. Remote Sens. Lett. 2011, 8, 542–546. [Google Scholar] [CrossRef] [Green Version]
  18. Fang, L.; Li, S.; Duan, W.; Ren, J.; Benediktsson, J.A. Classification of hyperspectral images by exploiting spectral–spatial information of superpixel via multiple kernels. IEEE Trans. Geosci. Remote Sens. 2015, 53, 6663–6674. [Google Scholar] [CrossRef] [Green Version]
  19. Ghamisi, P.; Souza, R.; Benediktsson, J.A.; Zhu, X.; Rittner, L.; Lotufo, R.A. Extinction profiles for the classification of remote sensing data. IEEE Trans. Geosci. Remote Sens. 2016, 54, 5631–5645. [Google Scholar] [CrossRef]
  20. Audebert, N.; Saux, B.L.; Lefevre, S. Deep Learning for Classification of Hyperspectral Data: A Comparative Review. IEEE Trans. Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef] [Green Version]
  21. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  22. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef] [Green Version]
  23. Chen, Y.; Zhao, X.; Jia, X. Spectral–Spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  24. Mei, X.; Pan, E.; Ma, Y.; Dai, X.; Huang, J.; Fan, F.; Du, Q.; Zheng, H.; Ma, J. Spectral-Spatial Attention Networks for Hyperspectral Image Classification. Remote Sens. 2019, 11, 963. [Google Scholar] [CrossRef] [Green Version]
  25. Hang, R.; Qian, X.; Liu, Q. Cross-Modality Contrastive Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5532812. [Google Scholar] [CrossRef]
  26. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  27. Hang, R.; Li, Z.; Ghamisi, P.; Bhattacharyya, S.S. Hyperspectral Image Classification With Attention-Aided CNNs. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2281–2293. [Google Scholar] [CrossRef]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–16 October 2016; pp. 630–645. [Google Scholar]
  29. Huang, G.; Liu, Z.; Maaten, L.v.d.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 22–25 July 2017; pp. 4700–4708. [Google Scholar]
  30. Lee, H.; Kwon, H. Going Deeper With Contextual CNN for Hyperspectral Image Classification. IEEE Trans. Image Process. 2017, 26, 4843–4855. [Google Scholar] [CrossRef] [Green Version]
  31. Wang, W.; Dou, S.; Sun, L.; Jiang, Z. A Fast Dense Spectral–Spatial Convolution Network Framework for Hyperspectral Images Classification. Remote Sens. 2018, 10, 1068. [Google Scholar] [CrossRef] [Green Version]
  32. Li, Z.; Wang, T.; Li, W.; Du, Q.; Wang, C.; Liu, C.; Shi, X. Deep Multilayer Fusion Dense Network for Hyperspectral Image Classification. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2020, 13, 1258–1270. [Google Scholar] [CrossRef]
  33. Liang, L.; Zhang, S.; Li, J. Multiscale DenseNet Meets With Bi-RNN for Hyperspectral Image Classification. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2022, 15, 5401–5415. [Google Scholar] [CrossRef]
  34. Ma, W.; Yang, Q.; Wu, Y.; Zhao, W.; Zhang, X. Double-Branch Multi-Attention Mechanism Network for Hyperspectral Image Classification. Remote Sens. 2019, 11, 1307. [Google Scholar] [CrossRef] [Green Version]
  35. Li, R.; Zheng, S.; Duan, C.; Yang, Y.; Wang, X. Classification of hyperspectral image based on double-branch dual-attention mechanism network. Remote Sens. 2020, 12, 582. [Google Scholar] [CrossRef] [Green Version]
  36. Chen, Y.; Fan, H.; Xu, B.; Yan, Z.; Kalantidis, Y.; Rohrbach, M.; Yan, S.; Feng, J. Drop an octave: Reducing spatial redundancy in convolutional neural networks with octave convolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3435–3444. [Google Scholar]
  37. Tang, X.; Meng, F.; Zhang, X.; Cheung, Y.M.; Ma, J.; Liu, F.; Jiao, L. Hyperspectral Image Classification Based on 3-D Octave Convolution With Spatial–Spectral Attention Network. IEEE Trans. Geosci. Remote Sens. 2021, 59, 2430–2447. [Google Scholar] [CrossRef]
  38. Xu, Q.; Xiao, Y.; Luo, B.; Wang, D. CSA-MSO3DCNN: Multiscale octave 3D CNN with channel and spatial attention for hyperspectral image classification. Remote Sens. 2020, 12, 188. [Google Scholar] [CrossRef] [Green Version]
  39. Xu, Q.; Wang, D.; Luo, B. Faster Multiscale Capsule Network With Octave Convolution for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2021, 18, 361–365. [Google Scholar] [CrossRef]
  40. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN Feature Hierarchy for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2020, 17, 277–281. [Google Scholar] [CrossRef] [Green Version]
  41. Yu, D.; Li, Q.; Wang, X.; Xu, C.; Zhou, Y. A Cross-Level Spectral–Spatial Joint Encode Learning Framework for Imbalanced Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5411717. [Google Scholar] [CrossRef]
  42. Zhu, Q.; Deng, W.; Zheng, Z.; Zhong, Y.; Guan, Q.; Lin, W.; Zhang, L.; Li, D. A Spectral-Spatial-Dependent Global Learning Framework for Insufficient and Imbalanced Hyperspectral Image Classification. IEEE Trans. Cybern. 2021, 52, 11709–11723. [Google Scholar] [CrossRef]
  43. Liang, L.; Zhang, S.; Li, J.; Cui, Z. Hyperspectral Image Classification Via Double-Branch Multi-Scale Spectral-Spatial Convolution Network. In Proceedings of the IIEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 3588–3591. [Google Scholar]
  44. Dong, H.; Zhang, L.; Zou, B. Band Attention Convolutional Networks for Hyperspectral Image Classification. arXiv 2019, arXiv:1906.04379. [Google Scholar]
  45. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 16–20 June 2020; pp. 11534–11542. [Google Scholar]
  46. Shang, R.; Ren, J.; Zhu, S.; Zhang, W.; Feng, J.; Li, Y.; Jiao, L. Hyperspectral Image Classification Based on Pyramid Coordinate Attention and Weighted Self-Distillation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5544316. [Google Scholar] [CrossRef]
  47. Tu, B.; He, W.; He, W.; Qu, X.; Plaza, A. Hyperspectral Classification via Global-Local Hierarchical Weighting Fusion Network. IEEE J. Sel. Topics Appl. Earth Observ. Remote Sens. 2022, 15, 184–200. [Google Scholar] [CrossRef]
  48. Alkhatib, M.Q.; Al-Saad, M.; Aburaed, N.; Almansoori, S.; Zabalza, J.; Marshall, S.; Al-Ahmad, H. Tri-CNN: A Three Branch Model for Hyperspectral Image Classification. Remote Sens. 2023, 15, 316. [Google Scholar] [CrossRef]
  49. Xue, Z.; Yu, X.; Liu, B.; Tan, X.; Wei, X. HResNetAM: Hierarchical Residual Network With Attention Mechanism for Hyperspectral Image Classification. IEEE J. Sel. Topics Appl. Earth Obs. Remote Sens. 2021, 14, 3566–3580. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the network of the MOCNN approach for HSI classification.
Figure 1. The flowchart of the network of the MOCNN approach for HSI classification.
Remotesensing 15 01758 g001
Figure 2. The schematic of the overall structure of the 2D octave convolutional network.
Figure 2. The schematic of the overall structure of the 2D octave convolutional network.
Remotesensing 15 01758 g002
Figure 3. Schematic of the structure of 3D DenseNet.
Figure 3. Schematic of the structure of 3D DenseNet.
Remotesensing 15 01758 g003
Figure 4. The schematic of the structure of BAM. S and ⨂ denote the sigmoid activation function and the multiplication of elements, respectively.
Figure 4. The schematic of the structure of BAM. S and ⨂ denote the sigmoid activation function and the multiplication of elements, respectively.
Remotesensing 15 01758 g004
Figure 5. The schematic of the structure of ECA. Here, I stands for the input feature map of the ECA, C represents the number of channels of I, P denotes the set of features obtained from the input feature map I after AP, and ⨂ denotes the multiplication of elements.
Figure 5. The schematic of the structure of ECA. Here, I stands for the input feature map of the ECA, C represents the number of channels of I, P denotes the set of features obtained from the input feature map I after AP, and ⨂ denotes the multiplication of elements.
Remotesensing 15 01758 g005
Figure 6. Reference maps of the four HSIs. In this figure, sub-maps (a,c,e,g) refer to the false-color images of the UP, SV, IN, and ZY, whereas sub-maps (b,d,f,h) refer to the ground truth corresponding to them.
Figure 6. Reference maps of the four HSIs. In this figure, sub-maps (a,c,e,g) refer to the false-color images of the UP, SV, IN, and ZY, whereas sub-maps (b,d,f,h) refer to the ground truth corresponding to them.
Remotesensing 15 01758 g006
Figure 7. Map of the classification results from all approaches on UP. (a) Ground truth. (b) CDCNN [30]. (c) FDCNN [31]. (d) DBMA [34]. (e) DBDA [35]. (f) SSAN [24]. (g) TriCNN [48]. (h) 3DOC-CNN [37]. (i) HRAM [49]. (j) MOCNN.
Figure 7. Map of the classification results from all approaches on UP. (a) Ground truth. (b) CDCNN [30]. (c) FDCNN [31]. (d) DBMA [34]. (e) DBDA [35]. (f) SSAN [24]. (g) TriCNN [48]. (h) 3DOC-CNN [37]. (i) HRAM [49]. (j) MOCNN.
Remotesensing 15 01758 g007
Figure 8. Map of the classification results from all approaches on SV. (a) Ground truth. (b) CDCNN [30]. (c) FDSSC [31]. (d) DBMA [34]. (e) DBDA [35]. (f) SSAN [24]. (g) TriCNN [48]. (h) 3DOC-CNN [37]. (i) HRAM [49]. (j) MOCNN.
Figure 8. Map of the classification results from all approaches on SV. (a) Ground truth. (b) CDCNN [30]. (c) FDSSC [31]. (d) DBMA [34]. (e) DBDA [35]. (f) SSAN [24]. (g) TriCNN [48]. (h) 3DOC-CNN [37]. (i) HRAM [49]. (j) MOCNN.
Remotesensing 15 01758 g008
Figure 9. Map of the classification results from all approaches on IN. (a) Ground truth. (b) CDCNN [30]. (c) FDSSC [31]. (d) DBMA [34]. (e) DBDA [35]. (f) SSAN [24]. (g) TriCNN [48]. (h) 3DOC-CNN [37]. (i) HRAM [49]. (j) MOCNN.
Figure 9. Map of the classification results from all approaches on IN. (a) Ground truth. (b) CDCNN [30]. (c) FDSSC [31]. (d) DBMA [34]. (e) DBDA [35]. (f) SSAN [24]. (g) TriCNN [48]. (h) 3DOC-CNN [37]. (i) HRAM [49]. (j) MOCNN.
Remotesensing 15 01758 g009
Figure 10. Map of the classification results from all approaches on ZY. (a) ground truth. (b) CDCNN [30]. (c) FDSSC [31]. (d) DBMA [34]. (e) DBDA [35]. (f) SSAN [24]. (g) TriCNN [48]. (h) 3DOC-CNN [37]. (i) HRAM [49]. (j) MOCNN.
Figure 10. Map of the classification results from all approaches on ZY. (a) ground truth. (b) CDCNN [30]. (c) FDSSC [31]. (d) DBMA [34]. (e) DBDA [35]. (f) SSAN [24]. (g) TriCNN [48]. (h) 3DOC-CNN [37]. (i) HRAM [49]. (j) MOCNN.
Remotesensing 15 01758 g010
Figure 11. Classification performance under different percentages of training samples. (a) UP. (b) SV. (c) IN. (d) ZY.
Figure 11. Classification performance under different percentages of training samples. (a) UP. (b) SV. (c) IN. (d) ZY.
Remotesensing 15 01758 g011
Figure 12. Results of ablation experiments on different types of datasets.
Figure 12. Results of ablation experiments on different types of datasets.
Remotesensing 15 01758 g012
Table 1. Category information and corresponding experimental sample size for UP.
Table 1. Category information and corresponding experimental sample size for UP.
CategoryTotalTrainingValidationTest
C166311001006431
C218,64928028018,089
C3209932322035
C4306446462972
C5134521211303
C6502976764877
C7133020201290
C8368256563570
C99471515917
Total42,77664664641,484
Table 2. Category information and corresponding experimental sample size for SV.
Table 2. Category information and corresponding experimental sample size for SV.
CategoryTotalTrainingValidationTest
C1200921211967
C2372638383650
C3197620201936
C4139414141366
C5267827272624
C6395940403879
C7357936363507
C811,27111311311,045
C9620363636077
C10327833333212
C11106811111046
C12192720201887
C139161010896
C14107011111048
C15726873737122
C16180719191769
Total54,12954954953,031
Table 3. Category information and corresponding experimental sample size for IN.
Table 3. Category information and corresponding experimental sample size for IN.
CategoryTotalTrainingValidationTest
C1466630
C214281721721084
C3830100100630
C42372929179
C54832727429
C67305858614
C7284420
C84785858362
C9203314
C10972117117738
C1124552952951865
C125937272449
C132052525155
C141265152152961
C153864747292
C1693121269
Total10,249123812387773
Table 4. Category information and corresponding experimental sample size for ZY.
Table 4. Category information and corresponding experimental sample size for ZY.
CategoryTotalTrainingValidationTest
C1262553532519
C2130227271248
C3344269693304
C410,2432052059833
C5142529291367
C6148430301424
C7180837371734
C8149230301432
Total23,82148048022,861
Table 5. Network parameters settings for 2D octave and 3D DenseNet in the MOCNN model.
Table 5. Network parameters settings for 2D octave and 3D DenseNet in the MOCNN model.
ModelType/LayerFilter/OperationConfiguration
2D octave2Doc-conv1( k s p a , k s p a ), 32stride:1, padding:1, BN+mish
2Doc-conv2( k s p a , k s p a ), 64
2Doc-conv3( k s p a , k s p a ), 32
2Doc-conv4( k s p a , k s p a ), 16
3D DenseNetconv1(1, 1, k s p e ), 32stride:2, padding:0, BN+mish
conv2(1, 1, k s p e ), 16stride:1, padding:1, BN+mish
conv3(1, 1, k s p e ), 16
concat1concat(conv2, conv3), 32BN+mish
conv4(1, 1, k s p e ), 16stride:1, padding:1, BN+mish
concat2concat(conv2, conv3, conv4), 48BN+mish
conv5(1, 1, k s p e ), 16stride:1, padding:1, BN+mish
concat3concat(conv2, conv3, conv4, conv5), 64BN+mish
conv6(1, 1, k s p e 0 ), 16stride:1, padding:0, BN+mish
Table 6. Parameter settings on four different datasets inclusive of the learning rate, dropout, batch size, and epoch.
Table 6. Parameter settings on four different datasets inclusive of the learning rate, dropout, batch size, and epoch.
ClassLearning RateDropoutBatch SizeEpoch
UP0.000050.532250
SV0.00010.532400
IN0.000050.532400
ZY0.000050.564600
Table 7. Quantitative comparison of the CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], HRAM [49], and MOCNN on UP.
Table 7. Quantitative comparison of the CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], HRAM [49], and MOCNN on UP.
CategoryCDCNNFDSSCDBMADBDASSANTriCNN3DOC-CNNHRAMMOCNN
C192.1999.3597.6798.9092.8089.8695.2098.7599.21
C296.3799.7099.3099.7299.4599.71100.0099.6899.95
C372.2897.0393.1095.9769.7382.9589.1992.8797.25
C497.9496.6196.6297.5398.1596.7197.9197.3697.98
C599.1999.7099.3899.5197.0199.8794.5599.2995.32
C692.0499.6699.0798.5593.6695.1597.9999.89100.00
C790.4799.9599.1499.2785.4380.5382.5699.2696.20
C883.1890.7994.3092.6393.1792.9397.6291.8595.52
C999.0898.1597.0397.5990.7398.7998.8097.72100.00
OA92.86
± 2.15
98.33
± 0.93
98.00
± 0.39
98.37
± 0.49
94.94
± 0.61
95.41
± 0.86
97.39
± 0.53
98.17
± 0.81
98.92
± 0.55
AA91.41
± 3.40
97.88
± 0.91
97.29
± 0.53
97.74
± 0.75
91.13
± 1.35
92.95
± 0.92
94.87
± 0.47
97.41
± 1.01
97.94
±0.71
Ka90.48
± 2.91
97.79
± 1.23
97.35
± 0.52
97.85
± 0.65
93.27
± 0.72
93.91
± 0.95
96.54
± 0.65
97.57
± 1.33
98.57
± 0.41
Table 8. Quantitative comparison of the CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], HRAM [49], and MOCNN on SV.
Table 8. Quantitative comparison of the CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], HRAM [49], and MOCNN on SV.
CategoryCDCNNFDSSCDBMADBDASSANTriCNN3DOC-CNNHRAMMOCNN
C149.99100.00100.0097.4886.3299.9899.08100.00100.00
C279.97100.0099.9299.9699.5399.9299.7898.72100.00
C392.9693.6497.7698.2092.5199.4994.1695.9999.70
C493.9597.5594.2397.1399.7899.5097.6696.7898.66
C589.5996.7397.5198.4696.8495.5899.2899.4799.78
C698.1899.8198.7799.88100.00100.0097.6899.9998.70
C798.0299.9999.8099.9199.0399.9699.6099.93100.00
C880.4596.2594.6696.9990.8895.8195.7093.4999.83
C998.2099.7299.7799.4098.4799.8898.6899.4495.78
C1087.1799.4096.4798.3199.6697.6192.2898.6099.67
C1176.0897.0097.3496.8895.3299.9199.7198.16100.00
C1291.2899.8899.7199.7089.5196.2397.1499.9898.28
C1397.0199.8798.1599.3797.4395.8797.4399.4798.62
C1497.3598.1398.4098.4689.6095.8590.3697.6499.89
C1552.7196.1187.9387.1289.9686.5894.0295.2696.47
C1699.02100.0098.1399.9696.1099.1297.0699.9399.35
OA83.77
± 5.41
97.83
± 1.10
95.91
± 1.58
96.49
± 2.10
94.67
± 0.53
96.59
± 0.75
96.69
± 0.41
97.21
± 1.12
98.69
± 0.32
AA86.31
± 6.21
98.38
± 0.90
97.41
± 0.95
97.95
± 0.91
95.06
± 0.47
97.58
± 0.71
96.85
± 0.55
98.30
± 0.74
99.05
± 0.55
Ka81.95
± 2.20
97.58
± 1.23
95.46
± 2.05
96.10
± 2.33
94.07
± 0.65
96.20
± 0.84
96.31
± 0.72
96.88
± 1.26
98.54
± 0.71
Table 9. Quantitative comparison of the CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], HRAM [49], and MOCNN on IN.
Table 9. Quantitative comparison of the CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], HRAM [49], and MOCNN on IN.
CategoryCDCNNFDSSCDBMADBDASSANTriCNN3DOC-CNNHRAMMOCNN
C148.2796.0797.35100.0091.18100.00100.0074.2497.06
C280.8598.2695.5196.6496.7694.5997.4895.7697.42
C378.4596.9799.1399.1092.7298.3597.9998.2498.57
C474.1298.4596.2398.4695.6395.45100.0099.04100.00
C596.4497.9997.6498.7792.5397.8194.1397.0395.10
C696.2797.4198.6998.9299.3099.6997.3698.3698.74
C744.9195.8981.7082.30100.0099.2075.0070.7095.00
C889.30100.0099.8299.95100.00100.00100.0098.49100.00
C942.5095.6395.8297.20100.0063.33100.0060.00100.00
C1078.9596.1695.9294.4196.3197.6698.1596.9297.97
C1178.0299.0397.1099.1899.1698.2797.8699.1598.98
C1263.4098.6894.9296.4996.3192.2794.7995.1898.22
C1398.4598.6999.7998.7894.9799.5699.37100.00100.00
C1493.7599.1298.6999.3299.2998.9999.1999.3099.90
C1586.8698.1395.9497.9094.3394.9994.6798.69100.00
C1697.6495.9893.1891.6687.3295.1294.3794.0095.65
OA81.76
± 3.71
98.16
± 0.80
97.04
± 1.08
97.88
± 0.54
97.14
± 0.52
97.36
± 0.49
97.60
± 0.32
97.79
± 1.50
98.58
± 0.41
AA78.01
± 5.81
97.65
± 1.04
96.09
± 0.84
96.82
± 0.49
95.99
± 1.12
95.33
± 1.53
96.27
± 0.91
92.19
± 3.35
98.29
± 0.67
Ka79.08
± 4.36
97.91
± 1.01
96.62
± 1.24
97.58
± 1.07
96.74
± 0.81
96.99
± 0.55
97.27
± 0.55
97.48
± 1.01
98.39
± 0.51
Table 10. Quantitative comparison of the CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], HRAM [49], and MOCNN on ZY.
Table 10. Quantitative comparison of the CDCNN [30], FDSSC [31], DBMA [34], DBDA [35], SSAN [24], TriCNN [48], 3DOC-CNN [37], HRAM [49], and MOCNN on ZY.
CategoryCDCNNFDSSCDBMADBDASSANTriCNN3DOC-CNNHRAMMOCNN
180.1997.3496.8696.8899.8895.2896.1095.6797.66
285.5397.6396.6497.0390.4697.7599.2897.5897.04
381.9798.1097.0698.0997.3195.5292.0598.0199.49
489.2298.3898.5298.4696.6897.6698.6698.4797.67
555.6295.8192.4696.1094.1590.4991.2197.4199.93
688.0096.3795.7595.1394.0396.2791.2097.0998.03
794.7596.2798.8798.4999.6097.3698.4497.6297.98
861.6994.2994.3194.5486.5992.1995.1095.3398.95
OA84.14
± 3.73
97.43
± 0.83
96.96
± 0.44
97.50
± 0.55
96.06
± 0.95
95.98
± 0.66
96.31
± 0.45
97.59
± 0.36
98.16
± 0.63
AA86.31
± 6.21
96.77
± 0.62
95.93
± 0.68
96.84
± 0.73
94.84
± 1.02
95.31
± 0.81
95.25
± 0.41
97.14
± 0.41
98.34
± 0.62
Ka81.95
± 2.20
96.63
± 1.12
96.02
± 0.57
96.72
± 0.72
94.83
± 1.80
94.74
± 0.85
95.15
± 0.77
96.84
± 0.47
97.59
± 1.04
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, L.; Zhang, S.; Li, J.; Plaza, A.; Cui, Z. Multi-Scale Spectral-Spatial Attention Network for Hyperspectral Image Classification Combining 2D Octave and 3D Convolutional Neural Networks. Remote Sens. 2023, 15, 1758. https://doi.org/10.3390/rs15071758

AMA Style

Liang L, Zhang S, Li J, Plaza A, Cui Z. Multi-Scale Spectral-Spatial Attention Network for Hyperspectral Image Classification Combining 2D Octave and 3D Convolutional Neural Networks. Remote Sensing. 2023; 15(7):1758. https://doi.org/10.3390/rs15071758

Chicago/Turabian Style

Liang, Lianhui, Shaoquan Zhang, Jun Li, Antonio Plaza, and Zhi Cui. 2023. "Multi-Scale Spectral-Spatial Attention Network for Hyperspectral Image Classification Combining 2D Octave and 3D Convolutional Neural Networks" Remote Sensing 15, no. 7: 1758. https://doi.org/10.3390/rs15071758

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop