Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Next Article in Journal
Design Parameter Optimization of a Silicon-Based Grating Waveguide for Performance Improvement in Biochemical Sensor Application
Previous Article in Journal
Development of a Novel Guided Wave Generation System Using a Giant Magnetostrictive Actuator for Nondestructive Evaluation
Previous Article in Special Issue
Multisensor Analysis of Spectral Dimensionality and Soil Diversity in the Great Central Valley of California
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ISBDD Model for Classification of Hyperspectral Remote Sensing Imagery

1
School of Instrumentation Science and Opto-Electronics Engineering, Beihang University, Beijing 100191, China
2
School of Engineering, Newcastle University, Newcastle upon Tyne NE1 7RU, UK
3
School of Geographical and Earth Sciences, University of Glasgow, Glasgow G12 8QQ, UK
4
China Geological Survey, Beijing 100037, China
*
Authors to whom correspondence should be addressed.
Sensors 2018, 18(3), 780; https://doi.org/10.3390/s18030780
Submission received: 13 December 2017 / Revised: 25 January 2018 / Accepted: 3 February 2018 / Published: 5 March 2018
(This article belongs to the Special Issue Analysis of Multispectral and Hyperspectral Data)

Abstract

:
The diverse density (DD) algorithm was proposed to handle the problem of low classification accuracy when training samples contain interference such as mixed pixels. The DD algorithm can learn a feature vector from training bags, which comprise instances (pixels). However, the feature vector learned by the DD algorithm cannot always effectively represent one type of ground cover. To handle this problem, an instance space-based diverse density (ISBDD) model that employs a novel training strategy is proposed in this paper. In the ISBDD model, DD values of each pixel are computed instead of learning a feature vector, and as a result, the pixel can be classified according to its DD values. Airborne hyperspectral data collected by the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor and the Push-broom Hyperspectral Imager (PHI) are applied to evaluate the performance of the proposed model. Results show that the overall classification accuracy of ISBDD model on the AVIRIS and PHI images is up to 97.65% and 89.02%, respectively, while the kappa coefficient is up to 0.97 and 0.88, respectively.

1. Introduction

Hyperspectral remote sensing is widely used in water resources, air quality monitoring, and agricultural and soil properties research because of its high spectral resolution [1,2,3,4]. The high classification accuracy of hyperspectral images is the precondition of its effective application. However, the classification of hyperspectral images encounters many problems, such as the “curse of dimensionality”, nonlinear data and spatial heterogeneity. Many techniques have been proposed to handle these problems. The feature mining technique is proposed to reduce the dimension to decrease the impact of the “curse of dimensionality” by mining the dimension of the hyperspectral image [5]; the kernel function transformation technique works well when dealing with nonlinear data [6]; and spectral-spatial clustering considered both the spectral information and spatial correlativity to reduce the impact of the spatial heterogeneity problem [7]. Additionally, research shows that the classification accuracy is reduced if training samples contain interference such as mixed pixels in supervised classification [8,9].
In the classification of drug protein molecules, Dietterich et al. [10] proposed a multi-instance learning (MIL) method to handle the problem of training samples with interference. Later, the DD algorithm was proposed by Maron and Lozano-Perez to learn a simple description of a person from a series of images containing that person [11], and became a classical algorithm for the MIL method. A LiMu et al. applied the DD algorithm to classifying the high-resolution remotely sensed image and gained better result than the support vector machine classifier [12]. For image retrieval, Wada et al. combined the K nearest neighbor algorithm and the DD algorithm for content-based image retrieval. Their work improved the accuracy of image retrieval and retrieval efficiency because the new method reduces the impact of irrelevant information, which is regard as interference factors in the image [13]. In these research fields, the DD algorithm performed effectively in dealing with the interference problem. Bolton et al. emphasized the reasonable application of the DD algorithm to hyperspectral image for image typically sensed from a distance to allow for image formation errors and spectral mixing [14,15,16].
In the DD algorithm, only one feature vector is learned from the training bags of a type of ground cover to represent this type of ground cover. Then the DD classifier will classify each pixel in the hyperspectral image according the feature vectors which can represent different types of ground cover. However, only use one feature vector cannot represent one type of ground precisely, because the spectral characteristics of different pixels for one type of ground cover may have some interference in certain spectral bands. The interference may make different pixels of one type of ground cover slightly different in terms of their spectral characteristics [8]. In the present study, the instance space-based diverse density (ISBDD) model, which employs a novel training strategy, is proposed to deal with this problem. In the ISBDD model, the DD values of each pixel are computed according to the training bags of different ground covers, and the pixels are classified according to their DD values. Thus, the pixels in hyperspectral images will not be classified according to the learned feature vectors of different types of ground cover; instead, they will be classified directly according to the training bags of different types of ground cover. The proposed ISBDD model greatly increases the classification accuracy of hyperspectral image because of the new training strategy.

2. Materials and Methods

2.1. DD Algorithm

The DD algorithm was proposed to handle the problem of training samples with interference [11]. The most basic unit in the DD algorithm is called an instance, which is actually a pixel in a hyperspectral image. The two types of instances are positive and negative [17]. If an instance belongs to a certain type of ground cover, then it is called a positive instance for this type of ground cover, otherwise it is called a negative instance. An area in the image that contains several pixels (instances) is called a training bag. The two types of training bags are as follows: (i) a positive training bag, which contains at least one positive instance; and (ii) a negative training bag, which contains only negative instances [18]. The purpose of the DD algorithm is to learn a feature vector from the feature space to represent the training bags. The learning strategy of a feature vector is that the feature vector must be similar to positive bags and not to negative bags [19]. The similarity can be measured by a probability value computed from the distance between the feature vector and the training bags. The probability value is known as the DD value.
The mathematical description of the DD algorithm is as follows.
Assume that B i + is the i-th positive bag and B i j + is the j-th instance in the i-th positive bag, while B i j k + is the k-th attribute of the j-th instance. Similarly, B i is the i-th negative bag and B i j is the j-th instance in the i-th negative bag, and B i j k is the k-th attribute of the j-th instance. A multi-dimensional vector in the feature space is marked x , and the maximum DD value is represented by t max . Pr ( x = t | B 1 + , B 2 + , ...... , B n + , B 1 , ...... , B n ) is the probability of an instance that belongs to a certain type of ground cover. The learning purpose is to dig t max out and learn a representative feature vector through maximizing Pr ( x = t | B 1 + , B 2 + , ...... , B n + , B 1 , ...... , B n ) . Supposing each instance is subject to independent distribution, according to Bayesian theory [20]:
t max = max i Pr ( x = t | B i + ) i Pr ( x = t | B i )
Maron and Lozano-Perez used the noisy-or model to embody Equation (1):
Pr ( x = t | B i + ) = 1 j ( 1 Pr ( x = t | B i j + ) )
Pr ( x = t | B i ) = j ( 1 Pr ( x = t | B i j ) )
The similarity between x and an instance in the training bag B i j is defined in the form of Equation (4).
Pr ( x = t | B i j ) = exp ( | | B i j x | | 2 )

2.2. Classification of Hyperspectral Remotely Sensed Image Based on the DD Algorithm

The interference in hyperspectral remotely sensed images mainly contains mixed pixels, noise, large dispersion degree in one class and so on. In the imaging process of hyperspectral image sensors, mixed pixels are common in the resulting hyperspectral image because of the limitation of spatial resolution of sensors, especially at the border of two types of ground cover. Mixed pixels that contain the spectral characteristics of several types of ground cover may reduce the classification accuracy of the hyperspectral image. Also, noise such as measurement uncertainty also exists in hyperspectral images. In general, the supervised classification task involves training and classification processes [21]. In traditional supervised classification, the training samples with interference selected from the hyperspectral image is regard as a pure training sample. This will reduce the classification accuracy of hyperspectral image. In the DD algorithm, the influence of training samples with interference is considered in the training process. The training process of the DD algorithm is described in Figure 1. Assume that the curved surface shown in Figure 1 represents a feature space, and CA, CB are two types of ground covers. The positive bag of CA is represented by CA1, and the negative bag of CA is represented by CA2. Similarly, CB1 and CB2 are the positive and negative bags of CB, respectively. C is an instance, which is a potential feature vector of a certain type of ground cover. Now we want to judge whether instance C is more likely to represent CA or CB. In this case, the distance between C and CA1 is the same as the distance between C and CB1, and the distance of C and CA2 is greater than that of C and CB2. Therefore, C is more likely to represent CA.
Generally, c 1 , c 2 , ...... , c k are K types of ground covers and c i j is the j-th bag of the i-th ground cover. The spectral dimension of the hyperspectral image is represented by n. c i 1 , c i 2 , ...... , c i m are m bags of c i . In the training process, a feature vector is learned according to the positive bags and negative bags of the i-th type of ground cover. In the hyperspectral image, the negative bag of one type of ground cover is the positive bag of the other type of ground cover. The goal of the training process is to learn a feature vector, which can represent the training bags of a certain type of ground cover. The process can be described by Formula (5), as follows:
f ( c i 1 , c i 2 , ...... , c i m ) = f i
where f i is the learned feature vector of c i , and the f function represents the training process. In the classification process, a pixel in the hyperspectral image can be classified according to the feature vector.
The classification result of the DD model is determined by the feature vector, which was learned in the training process. However, only use one feature vector alone cannot represent all the pixels of a type of ground cover, and this will influence the performance of the DD model. In order to handle this problem, the ISBDD model is proposed.

2.3. ISBDD Model for Classification of Hyperspectral Remotely Sensed Image

The ISBDD model employs a new training strategy to carry out the training process. In the DD algorithm, only one feature vector is used to represent the training bags. However, only one feature vector alone cannot represent one type of ground cover well. In the new training strategy, we do not learn a feature vector from the training bags. Instead, we classify the unknown pixels according to DD values of the unknown pixel, which is computed from the training bags of all the types of ground cover.
Suppose that k types of ground cover and N pixels exist in a hyperspectral image. The vector of an input instance (pixel) is represented as x i R n ( i N ) , and the output label is represented as y { c 1 , c 2 , ...... , c k } , where c j represents the j-th type of ground ground cover. v i 1 , v i 2 , ...... , v i k are DD values of x i computed according to training bags of these k types of ground covers , where v i m is the DD value of pixel x i computed from the training bags of the m-th type of ground cover. The formula to compute v i m is as follows:
v i m = l Pr ( x i = t | B q + ) l Pr ( x i = t | B q )
where l is the number of training bags for the m-th type of ground cover.
The formula is also embodied through the noisy-or model as follows:
Pr ( x i = t | B q + ) = 1 j ( 1 Pr ( x i = t | B q j + ) )
Pr ( x i = t | B q ) = j ( 1 Pr ( x i = t | B q j ) )
After the training process, each pixel obtains k DD values, and the label of pixel x i is gained using Formula (9), as follows:
y = arg c i max { v i 1 , v i 2 , ...... , v i k }
The processes of the DD model and the ISBDD model are compared in Figure 2. In the training process of the DD model, the feature vector f i is learned, whereas in the ISBDD model, learning a feature vector is not required to represent the training bags. In the new model, we directly compute according to the training bags, and all the information provided by the training bags is utilized in the training process to deal with the noisy training samples. For example, a hyperspectral image that contains two types of ground cover needs to be classified. In the ISBDD model, firstly, the training samples of the two types of ground cover are selected according to the ground truth image or the ground survey image. Then, each pixel in the hyperspectral image which need to be classified is looped through, and the diverse density values of the pixel and the training sample bags of the two types of ground cover are computed according to Formula (6). Finally, the pixel is classified as a certain type of ground cover according to its diverse density values.

2.4. Experiment Description

The experiment for the classification of hyperspectral images was implemented on a PHI image and an AVIRIS image. The airborne hyperspectral image collected by PHI covers the Fanglu Tea plantation area in Jiangsu province of China, which is situated at (31°40′39″ N, 119°22′53″ E). We used an image with a size of 200   ×   150 pixels and 65 spectral channels in the experiment. The image was collected on October 2002. The area where the image covers contains 7 types of ground cover, namely paddy, caraway, wild-grass, pachyrhizus, tea, bamboo, and water. The spatial resolution of the PHI image is 2 m. The AVIRIS image named Indian Pines covers the agricultural demonstration zone in Northwest Indiana of the US. We use an image with a size of 145   ×   145 pixels and 224 spectral channels. It was collected on June 1992. The area where the image covers contains 16 types of ground cover, namely, alfalfa, corn-min (corn seeding), corn, grass/trees, grass/pasture, grass/pasture-moved (trimmed grass/pasture), hay-windrowed, oats, soybeans-notill (no-till soybeans), soybeans-min, soybean-clean (cleaned soybeans), wheat, woods, bldg-grass-tree-drives, stone-steel towers and corn-notill. The spatial resolution of the AVIRIS image was 20 m. We used the AVIRIS image to test the feasibility of ISBDD and then used the PHI image to test the applicability of ISBDD.
We call the hyperspectral image a data cube because it consists of two dimensions in the spatial dimension and one dimension in the spectral dimension. Figure 3 shows the data cube of the two images collected by AVIRIS and PHI, and Figure 4 shows the distribution of ground cover in the two images. Figure 4a is the ground truth image of the Indian Pines and Figure 4b is the ground survey image of the Fanglu Tea plantation. The two images can describe the distribution of ground cover. In Figure 4a, different color represents different types of ground cover, and the black color in the ground truth image of Indian Pines represent area where the ground cover is unknown.
The spectral characteristic of seven types of ground cover in the Indian Pines are shown in Figure 5. The training and testing samples are selected from the image, and their specific information is shown in Table 1. In this case, the training samples without interference for a certain type of ground cover indicate that the labels of all pixels in the training samples are consistent with the label of a certain type of ground over. The training samples with interference for a certain class indicate that the training samples contain pixels whose labels are inconsistent with the label of a certain type of ground cover. In the stage of sample selection, we choose the pixels which has the inconsistent label as the interference for a type of ground cover. We choose pixels with interference and pixels without interference based on the ground truth image of the Indian Pines and the ground survey image of the Fanglu Tea plantation.
The spectral characteristic of the seven types of ground cover in Fanglu Tea plantation area are shown in Figure 6. The training samples and testing samples are selected from the hyperspectral image, and the specific information is shown in Table 2.
The maximum likelihood (MLC) algorithm is a classical classification algorithm for remotely sensed images, and the support vector machine (SVM) method has been a popular and effective classification algorithm in recent years [22]. We compare the classifier based on MLC and SVM algorithm with the classifier based on ISBDD algorithm. To fully verify the ability of the ISBDD classifier, the DD classifier is utilized. The kernel function of the SVM classifier is the radial basis function. We select the tolerant penalty parameter of the SVM classifier. To verify the performance of the ISBDD model, pixels with interference, which do not belong to a certain type of ground cover, are introduced to the training samples. Thereby, we use the training samples with interference to train the ISBDD and DD classifiers. The training samples without interference are a subset of the training samples. The MLC and SVM classifiers are applied to the hyperspectral image using the training samples without interference and the training samples with interference.
The training samples of the two hyperspectral images are divided into five subsets, and the number of training samples in each subset is equal. Five rounds of experiment are conducted, and the final classification result is the average of the five results. The design of a randomized block experiment can reduce the impact of random factors and avoid the over-fitting of the results. The results are evaluated based on the overall accuracy, kappa coefficient and accuracies of each class.

3. Results

3.1. AVIRIS Image

We use the AVIRIS image to test the feasibility of ISBDD. Figure 7 shows the classified images of the Indian Pines. Figure 7a–f shows the images classified by the MLC classifiers, SVM classifiers, DD and ISBDD classifiers.
The classification performance can be measured by the classification accuracy of each type of ground cover, the overall accuracy, and the kappa coefficient. The classification accuracy of each type of ground cover can be used to evaluate the classification accuracy of each ground cover, and the overall accuracy can be used to evaluate the performance of a classifier on the entire image. The kappa coefficient can be used to evaluate the consistency of the classification results and the actual distribution. Table 3 shows the average classification accuracy comparison for the purpose of evaluating the performance of the four classifiers. The maximum accuracy is highlighted in bold.
Figure 7 and Table 3 show that the DD and ISBDD classifiers perform better than the MLC and SVM classifiers when the training samples contain interference. The ISBDD classifier performs best in the classification accuracy of each type of ground cover because it has the highest classification accuracy in 10 of the 16 types of ground cover. The capacity of the DD and ISBDD classifiers to deal with the interference problem benefits from their MIL framework. The ISBDD classifier performs better than the MLC and SVM classifiers even if its training samples contain interference, whereas the training samples of the MLC and SVM classifiers do not contain interference. However, the DD classifier is not effective in this situation because the ISBDD classifier employs a novel training strategy in the training process. The result proves that the use of ISBDD is feasible for handling the interference problem.

3.2. PHI Image

We use the AVIRIS image to test the applicability of ISBDD. The classified images of the Fanglu tea plantation are shown in Figure 8. Figure 8a–f shows the images classified by the four classifiers.
Table 4 shows the average classification accuracy comparison for the purpose of evaluating the performance of the four classifiers. The maximum precision is highlighted in bold.
Figure 8 and Table 4 show that the classification accuracy of MLC and SVM is the lowest when the training samples contain interference. However, when the training samples are pure, the classification accuracy of these classifiers is high, especially for the SVM. The ISBDD classifier performs best in the classification accuracy of each type of ground cover because it has the highest classification accuracy in three of the seven types of ground cover. The overall accuracy of ISBDD is also the highest, while the classification accuracy of DD is lower than that of the SVM (without interference). The results demonstrate the effectiveness of ISBDD.
The classification results of the two hyperspectral images demonstrate the feasibility and applicability of the ISBDD model when the training samples contain interference. When the training samples are not contain interference, the MLC classifier and the SVM classifier can achieve satisfactory classification accuracy. However, when the training samples contain interference, the advantage of the ISBDD model is apparent.

4. Discussion

4.1. Influence of Interference Intensity

The purpose of the DD and ISBDD model is to deal with problem of interference, wherein the proportion of pixels with interference can modulate the performance of the classifiers. Figure 9 shows the impact of the proportion of pixels with interference in the training samples on classification accuracies using the four classifiers for the Indian Pines image. The x-axis is the ratio of pixels with interference to pixels without interference in the training samples, indicating the intensity of interference. The y-axis shows the overall accuracy.
Figure 9 shows that the overall accuracy of the ISBDD classifier is higher than that of the other classifiers. In cases where the ratio is larger than 0.5, the overall accuracy of the MLC and SVM classifier declines sharply, whereas the overall accuracy of the DD and ISBDD classifier declines slowly. However, the decline range of overall accuracy from 0 to 0.9 of the ISBDD classifier is almost the same as that of the MLC classifier, and these results demonstrate the advantage of ISBDD in dealing with the interference problem.
Figure 10 illustrates the impact of the intensity of pixels with interference in training samples on classification accuracy when the four classifiers are used for the Indian Pines image. The x-axis shows the ratio of the pixels with interference to pixels without interference in the training samples. The y-axis shows the overall accuracy.
Figure 10 shows that the overall accuracy of the ISBDD classifier is higher than that of the other classifiers. The DD classifier does not perform better than the SVM or the MLC classifier when the ratio of pixels with interference and pixels without interference is smaller than 0.2. However, when the ratio is larger than 0.4, the overall accuracy of the MLC and SVM classifier declines sharply, whereas the overall accuracy of the DD and ISBDD classifier declines slowly. The result shows that the MLC classifier and the SVM classifier is more sensitive to interference than the ISBDD classifier.

4.2. Application Prospects and Future Work

An increasing number of satellites equipped with hyperspectral camera are being launched for the purpose of crop survey, military target detection, and so on. The research and application of pattern recognition, such as face detection, is also gaining popularity. All these fields are concerned with supervised classification, in which the interference problem is inevitable when selecting the training samples. The ISBDD can be used in supervised classification to handle the problem of training samples with interference. The training strategy of ISBDD provides the model the powerful capability to complete the classification task though the training samples that contain interference.
However, the high accuracy cost of the ISBDD classifier is that its computational complexity is higher than that of the DD classifier. Suppose that w is the width of a hyperspectral image, and h is the height of a hyperspectral image, then the w × h pixels exist completely in the hyperspectral image. Assuming that a total of N pixels exist in the positive bags and the image contains c types of ground cover, then in the training process, most computing resources are used to determine the DD value. Thus, the ratio of the computational complexity of the ISBDD classifier to that of the DD classifier can be described as follows:
R = w × h × c N
In general, the numerator of Formula (10) is greater than the denominator. Thus, the computational complexity of the ISBDD classifier is considerably higher. The efficiency of the ISBDD classifier should be improved because, in many situations, efficiency is crucial. One of the solutions for improving efficiency is to utilize the multi-thread technique and parallel computing, given that the computing process for each pixel in the ISBDD classifier is independent.

5. Conclusions

The ISBDD model, which employs a novel training strategy, is proposed in this paper. The major contribution of this work is in exploring a model with high classification accuracy to handle the problem of training samples with interference in the classification of hyperspectral image. In the proposed model, instead of learning a feature vector to represent the training bags, pixels are classified directly according to training bags. The ISBDD model is compared with several classical hyperspectral image classification models, including the diverse density model. Results demonstrated that the classifier based on the ISBDD model performs better than that based on other algorithms. Specifically, the overall accuracy of the ISBDD classifier is approximately 8% higher than that of the DD classifier. The overall accuracy of the proposed classifier is much higher than the SVM and MLC classifiers when the training samples contain interference. The results show that the ISBDD classifier can handle the interference problem effectively. In addition, it can be used in classification scenarios, especially when an accurate classification is highly required. However, the computational complexity of the ISBDD model is higher than that of the DD classifier because several DD values are computed for each pixel. Thus, the multi-thread technique and parallel computing can be utilized to increase the efficiency.

Acknowledgments

This work was supported by the National High Technology Research and Development Program (863 Program) (Grant No. 2012YQ05250 and No. 2016YFF0103604), National Key Technologies R&D Program (Grant No. 2016YFB0500505), National Natural Science Foundation of China (Grant No. 41402293), Program for Changjiang Scholars and Innovative Research Team (Grant No. IRT0705), China Scholarship Council (Ref No. 201606025034) and the UK Science and Technology Facilities Council (STFC) through the PAFiC project (Ref: ST/N006801/1).

Author Contributions

Na Li and Huijie Zhao conceived and designed the algorithm; Zhaopeng Xu and Xinchen Huang performed the experiments; Na Li and Zhaopeng Xu analyzed the results; Zhenhong Li, Jane Drummond and Daming Wang provided comments on the algorithm and experiments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Garaba, S.P.; Zielinski, O. Comparison of remote sensing reflectance from above-water and in-water measurements west of Greenland, Labrador Sea, Denmark Strait, and west of Iceland. Opt. Express 2013, 21, 15938–15950. [Google Scholar] [CrossRef] [PubMed]
  2. Filippi, A.M.; Archibald, R.; Bhaduri, B.L.; Bright, E.A. Hyperspectral agricultural mapping using support vector machine-based endmember extraction (SVM-BEE). Opt. Express 2009, 17, 23823–23842. [Google Scholar] [CrossRef] [PubMed]
  3. Babaeian, E.; Homaee, M.; Montzka, C.; Vereecken, H.; Norouzi, A.A. Towards retrieving soil hydraulic properties by hyperspectral remote sensing. Vadose Zone J. 2015, 14. [Google Scholar] [CrossRef]
  4. Klein, M.E.; Aalderink, B.J.; Padoan, R.; De Bruin, G.; Steemers, T.A. Quantitativen Hyperspectral Reflectance Imaging. Sensors 2008, 8, 5576–5618. [Google Scholar] [CrossRef] [PubMed]
  5. Jia, X.; Kuo, B.C.; Crawford, M.M. Feature mining for hyperspectral image classification. Proc. IEEE 2013, 101, 676–697. [Google Scholar] [CrossRef]
  6. Qi, B.; Zhao, C.; Youn, E.; Nansen, C. Use of weighting algorithms to improve traditional support vector machine based classifications of reflectance data. Opt. Express 2011, 19, 26816–26826. [Google Scholar] [CrossRef] [PubMed]
  7. Li, N.; Cao, Y. An improved classification approach based on spatial and spectral features for hyperspectral data. Spectrosc. Spectr. Anal. 2014, 34, 526–531. [Google Scholar]
  8. Li, N.; Zhao, H.J.; Huang, P.; Jia, G.R.; Bai, X. A novel logistic multi-Class supervised classification model based on multi-fractal spectrum parameters for hyperspectral Data. Int. J. Comput. Math. 2015, 92, 836–849. [Google Scholar] [CrossRef]
  9. Tao, D.; Jia, G.; Yuan, Y.; Zhao, H. A digital sensor simulator of the pushbroom offner hyperspectral imaging spectrometer. Sensors 2014, 14, 23822–23842. [Google Scholar] [CrossRef] [PubMed]
  10. Dietterich, T.G.; Lathrop, R.H.; Lozano-Pérez, T. Solving the multiple instance problem with axis-parallel rectangles. Artif. Intell. 1997, 89, 31–71. [Google Scholar] [CrossRef]
  11. Maron, O.; Lozano-Pérez, T. A framework for multiple-instance learning. Adv. Neural Inf. Process. Syst. 1998, 200, 570–576. [Google Scholar]
  12. Du, P.; Samat, A. Multiple instance learning method for high-resolution remote sensing image classification. Remote Sens. Inf. 2012, 27, 60–66. [Google Scholar]
  13. Wada, T.; Mukai, Y. Fast Keypoint Reduction for Image Retrieval by Accelerated Diverse Density Computation. In Proceedings of the 2015 IEEE International Conference on Data Mining Workshop, Atlantic City, NJ, USA, 14–17 November 2015; pp. 102–107. [Google Scholar]
  14. Bolton, J.; Gader, P. Application of Multiple-Instance Learning for Hyperspectral Image Analysis. IEEE Geosci. Remote Sens. Lett. 2015, 8, 889–893. [Google Scholar] [CrossRef]
  15. Bolton, J.; Gader, P. Multiple instance learning for hyperspectral image analysis. In Proceedings of the 2010 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Honolulu, HI, USA, 25–30 July 2010; pp. 4232–4235. [Google Scholar]
  16. Bolton, J.; Gader, P. Spatial multiple instance learning for hyperspectral image analysis. In Proceedings of the 2010 2nd Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing, Reykjavik, Iceland, 14–16 June 2010; Volume 8, pp. 1–4. [Google Scholar]
  17. Vanwinckelen, G.; Fierens, D.; Blockeel, H. Instance-level accuracy versus bag-level accuracy in multi-instance learning. Data Min. Knowl. Discov. 2016, 30, 313–341. [Google Scholar] [CrossRef] [Green Version]
  18. Xu, L.; Guo, M.Z.; Zou, Q.; Liu, Y.; Li, H.F. An Improved Diverse Density Algorithm for Multiple Overlapped Instances. In Proceedings of the Fourth International Conference on Natural Computation (ICNC ’08), Jinan, China, 18–20 October 2008. [Google Scholar]
  19. Fu, J.; Yin, J. Bag-level active multi-instance learning. In Proceedings of the Eighth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD), Shanghai, China, 26–28 July 2011; Volume 2, pp. 1307–1311. [Google Scholar]
  20. Chen, M.S.; Yang, S.Y.; Zhao, Z.J.; Fu, P.; Sun, Y.; Li, X.; Sun, Y.; Qi, X. Multi-points diverse density learning algorithm and its application in image retrieval. J. Jilin Univ. 2011, 41, 1456–1460. [Google Scholar]
  21. Ordóñez, C.; Cabo, C.; Sanz-Ablanedo, E. Automatic Detection and Classification of Pole-Like Objects for Urban Cartography Using Mobile Laser Scanning Data. Sensors 2017, 17, 1465. [Google Scholar] [CrossRef] [PubMed]
  22. Kozoderov, V.V.; Dmitriev, E.V. Testing different classification methods in airborne hyperspectral imagery processing. Opt. Express 2016, 24, A956–A965. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Training process of the DD algorithm.
Figure 1. Training process of the DD algorithm.
Sensors 18 00780 g001
Figure 2. Flow chart of the DD and ISBDD models.
Figure 2. Flow chart of the DD and ISBDD models.
Sensors 18 00780 g002
Figure 3. Data cube of images collected by AVIRIS and PHI.
Figure 3. Data cube of images collected by AVIRIS and PHI.
Sensors 18 00780 g003
Figure 4. Distribution of ground cover in the two images collected by AVIRIS and PHI.
Figure 4. Distribution of ground cover in the two images collected by AVIRIS and PHI.
Sensors 18 00780 g004
Figure 5. Spectral characteristic of 16 types of ground covers covered in the Indian Pines.
Figure 5. Spectral characteristic of 16 types of ground covers covered in the Indian Pines.
Sensors 18 00780 g005
Figure 6. Spectral characteristic of seven types of ground cover covered in the Fanglu Tea plantation.
Figure 6. Spectral characteristic of seven types of ground cover covered in the Fanglu Tea plantation.
Sensors 18 00780 g006
Figure 7. Classified images of the Indian Pines.
Figure 7. Classified images of the Indian Pines.
Sensors 18 00780 g007
Figure 8. Classified images of the Fanglu Tea plantation.
Figure 8. Classified images of the Fanglu Tea plantation.
Sensors 18 00780 g008
Figure 9. The impact of intensity of interference on classification accuracy for the Indian Pines.
Figure 9. The impact of intensity of interference on classification accuracy for the Indian Pines.
Sensors 18 00780 g009
Figure 10. The impact of intensity of interference on classification accuracy for the Fanglu tea plantation.
Figure 10. The impact of intensity of interference on classification accuracy for the Fanglu tea plantation.
Sensors 18 00780 g010
Table 1. Number of training samples and testing samples for the Indian Pines.
Table 1. Number of training samples and testing samples for the Indian Pines.
Class IDClass NameTraining SamplesTesting Samples
Without InterferenceWith Interference
1Alfalfa253218
2Corn-min283581
3Corn303854
4Grass/trees3140161
5Grass/pasture253080
6Grass/pasture-moved202512
7Hay-windrowed506564
8Oats202510
9Soybeans-notill2937155
10Soybeans-min3953184
11Soybean-clean384875
12Wheat304060
13Woods3848156
14Bldg-grass-tree-drives302560
15Stone-steel towers354516
16Corn-notill3038178
Table 2. Number of training samples and testing samples for the Fanglu Tea plantation.
Table 2. Number of training samples and testing samples for the Fanglu Tea plantation.
Class IDClass NameTraining SamplesTesting Samples
Without InterferenceWith Interference
1(W2)Water92120954
2(C4)Paddy195255976
3(V13)Caraway105138295
4(S2)Wild-grass105135382
5(V2)Pachyrhizus6682211
6(T7)Tea105135411
7(T6)Bamboo135180443
Table 3. Average classification accuracy comparison of the four classifiers.
Table 3. Average classification accuracy comparison of the four classifiers.
MethodMLCSVMMLC (Without Interference)SVM (Without Interference)DDISBDD
Alfalfa (%)85.56100.081.11100.00100.0100.0
Corn-min (%)56.2267.5566.7196.2277.2091.19
Corn (%)97.1498.5797.1498.5724.2975.71
Grass/trees (%)60.4075.3079.0699.7391.4195.70
Grass/pasture (%)81.00100.087.17100.00100.0100.0
Grass/pasture-moved (%)63.33100.011.67100.00100.0100.0
Hay-windrowed (%)99.8790.6799.8789.8777.7489.87
Oats (%)24.0092.008.00100.0088.00100.0
Soybeans-notill (%)32.3957.3228.3156.3477.3276.62
Soybeans-min (%)78.1688.7888.5785.3077.9694.90
Soybean-clean (%)100.0100.0100.00100.0093.21100.0
Wheat (%)95.67100.097.33100.00100.0100.0
Woods (%)98.1099.0598.3399.2993.33100.0
Bldg-grass-tree-drives (%)11.0034.337.3339.6750.6759.00
Stone-steel towers (%)100.0100.0100.0098.4029.6099.20
Corn-notill (%)40.9535.8118.8647.4360.5765.52
Overall accuracy (%)68.1777.7469.9284.7580.5589.02
Kappa coefficient0.650.760.670.840.790.88
Table 4. Average classification accuracy comparison of the four classifiers.
Table 4. Average classification accuracy comparison of the four classifiers.
MethodMLCSVMMLC (Without Interference)SVM (Without Interference)DDISBDD
Water (%)92.2495.6691.7294.3293.2797.72
Paddy (%)69.4580.0082.0998.0771.7499.78
Caraway (%)98.3798.10100.0099.4699.5298.85
Wild-grass (%)79.5872.0992.9894.5696.6592.20
Pachyrhizus (%)88.7290.0596.9797.9196.7895.26
Tea (%)95.0897.1398.3098.7499.3798.73
Bamboo (%)94.4096.6190.5292.6994.7296.89
Overall accuracy (%)85.7489.2090.8596.2689.4697.65
Kappa coefficient0.830.870.890.950.870.97

Share and Cite

MDPI and ACS Style

Li, N.; Xu, Z.; Zhao, H.; Huang, X.; Li, Z.; Drummond, J.; Wang, D. ISBDD Model for Classification of Hyperspectral Remote Sensing Imagery. Sensors 2018, 18, 780. https://doi.org/10.3390/s18030780

AMA Style

Li N, Xu Z, Zhao H, Huang X, Li Z, Drummond J, Wang D. ISBDD Model for Classification of Hyperspectral Remote Sensing Imagery. Sensors. 2018; 18(3):780. https://doi.org/10.3390/s18030780

Chicago/Turabian Style

Li, Na, Zhaopeng Xu, Huijie Zhao, Xinchen Huang, Zhenhong Li, Jane Drummond, and Daming Wang. 2018. "ISBDD Model for Classification of Hyperspectral Remote Sensing Imagery" Sensors 18, no. 3: 780. https://doi.org/10.3390/s18030780

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop