Land Use and Land Cover Classification Meets Deep Learning: A Review
Abstract
:1. Introduction
2. Typical DL Models
2.1. Convolutional Neural Networks
2.2. Recurrent Neural Networks
2.3. Generating Adversarial Networks
2.4. Autoencoder
2.5. Fully Convolutional Neural Networks
3. Datasets and Performance Metrics
3.1. Pixel-Level Datasets
- The Indian Pines [33,34] was created by NASA in 2015 and was the first public dataset to be proposed for land cover classification. The AVIRIS sensor (224 spectral bands) imaged Indian Pines, Indiana, and the sample dataset covered 145 by 145 pixels with 16 land cover classes and a spatial resolution of 20 m. The Indian Pines dataset is small in size but still representative of developing new algorithms and methods. For evaluating algorithms, a ratio of 80:20 or 70:30 is usually chosen to divide the training and test sets (Figure 7).
- Pavia University [34] was published in 2015, and the image was captured by a ROSIS sensor with a size of 610 × 340 pixels. There are nine different land cover classes, including pasture, bare soil, grassland, etc., containing a total of 103 spectral bands. A ratio of 80:20 or 70:30 is usually chosen to divide the training and test sets to evaluate the algorithms (Figure 8).
- LoveDA [35] is an adaptive ground cover dataset created by the RSIDEA team at Wuhan University in 2021 for land cover mapping. The 5987 images of this dataset are taken from the data collected by GoogleEarth in three areas, Wuhan, Nanjing, and Changzhou, and each image is 1024 × 1024 pixels in size with a spatial resolution of 0.3 m after correction processing. Due to the late publication of the LoveDA dataset, it has had less impact compared to some of the classic semantic segmentation datasets, such as Indian Pines and Pavia University (Figure 9).
- LandCoverNet [36] was created in 2020 as a public dataset for land cover classification for the world. These data are taken from Sentinel-1/2 and Landsat-8 multispectral satellite images acquired in 2018. This dataset has a total of 1980 samples with a size of 256 × 256 pixels and a spatial resolution of 10 m. It includes seven feature classes, namely, artificial bare ground, natural bare ground, cultivated vegetation, woody vegetation, semi-natural vegetation, water, and permanent snow/ice (Figure 10).
3.2. Patch-Level Datasets
- The UC Merced [39] dataset was proposed by researchers at the University of California, Merced, in 2010 and contains 21 scene categories. Each scene category has 100 images with a region of 256 × 256 pixels, where each pixel has a spatial resolution of 0.3 m. The 2100 images were collected from a variety of regions, including Los Angeles, Houston, and Miami, and cover a wide range of land-use types in the U.S. region. The vast majority of land use remote sensing classification experiments use the UC Merced dataset for data comparison, and again, it was one of the first datasets to be proposed. Most of the experiments divide the UC Merced dataset into an 80:20 or 50:50 ratio between training and test sets for algorithm evaluation (Figure 11).
- The AID [40] dataset was proposed in 2017, and was created by a research team from Wuhan University from images collected from Google Earth Images, covering many countries and regions around the world. The dataset contains 30 scene categories, each category contains 220 to 420 images, and the entire dataset has a total of 10,000 images, all of which are 600 × 600 pixels in size and have a spatial resolution of 8 m to 0.5 m. Because the images of the AID dataset are collected from multiple sensors, the algorithm is usually evaluated by dividing the training set and test set into 20:80 and 50:50 ratios (Figure 12).
- The NWPU-RESISC45 [41] is a dataset proposed by a research team from Northwestern Polytechnical University in 2017, with 45 scene categories, each with 700 images, all of which have a size of 256 × 256 pixels. Most of the images in this dataset have a spatial resolution of 30 m to 0.2 m, with lower spatial resolution for snow-covered mountain, lake, and island images. Because the NWPU-RESISC45 dataset contains a large number of scene categories, there exist some categories with high similarity between them, which makes it more challenging to develop new algorithms and methods (Figure 13).
- The EuroSAT [42] dataset is a large-scale dataset containing 27,000 images and was presented in 2019. It consists of images captured by the Sentinel-2 satellite and contains 13 spectral bands. A total of 10 scene categories are included, namely, cities, forests, farmland, grasslands, lakes, rivers, coasts, deserts, mountains, and industrial areas. Each category has 2000–3000 images, each with a region of 64 × 64 pixels. When evaluating the algorithm, a ratio of 80:20 is usually chosen to divide the training and test sets (Figure 14).
3.3. Performance Indicators
- Overall accuracy (OA) is the most used performance metric in LULC classification to measure the rate at which the model correctly predicts the samples in the dataset, which is represented by Equation (2).
- Average accuracy (AA) indicates the average of the rate of correct predictions by the model in each category sample, as represented by Equation (3).
- F1-score (F1) is the average value after calculating the precision rate and recall rate reconciliation; the larger the F1, the better the model performance, represented by Equation (4).
- Mean intersection and unity ratio (MIOU) represents the average of the prediction accuracies of all the categories in the dataset and is commonly used to evaluate semantic segmentation, As expressed by Equation (5).
4. Deep-Learning-Based LULC Classification
4.1. Convolutional Neural Networks
4.2. Generating Adversarial Networks
4.3. Recurrent Neural Networks
4.4. Autoencoder
5. LULC Challenges
- Data diversity: Remote sensing data may change according to different times, different regions, different sensors, different climates, etc., so our model needs to be constantly updated to adapt to the variability caused by data diversity.
- Category imbalance: When making LULC datasets, some categories will have too many or too few data samples due to factors such as differences in geographical distribution, human intervention, and labeling difficulties. Therefore, the model tends to favor the categories with more data when training these data, while the categories with fewer samples lead to poorer classification accuracy due to too few training samples.
- Sample labeling problem: The labeling of LULC training samples not only requires researchers to have a high level of expertise and be familiar with the characteristics of the vegetation, urban, and water categories, it also requires a lot of manpower and time to classify the category areas and classifications. Some features are highly similar to each other, e.g., swamps and wetlands, and researchers may make subjective errors in judgment that lead to relatively fuzzy boundary delineation between categories, thus reducing the classification accuracy of the model.
- Generalization ability of the model: A model may achieve high classification accuracy in a specific region; however, if it is changed to a region or area in other countries, the generalization ability of the model will be reduced due to different distribution of features, differences in the labeling of the samples, and differences in the characteristics of the categories, among other factors.
- Contextual modeling: Remote sensing data are spatially correlated, i.e., there are dependencies between data in nearby areas. However, traditional deep learning methods do not mine the contextual information between data well, making the model unable to learn and understand the data comprehensively, which reduces the classification accuracy of the model.
6. LULC Classification with Limited Samples
6.1. Transfer Learning
6.2. Data Augmentation
6.3. Active Learning
6.4. Weak Supervision Methods
6.4.1. Zero-Shot Learning
6.4.2. Few-Shot Learning
7. Prospects
- Cross-modal data fusion: With the rapid development of deep learning methods in the field of LULC classification, more research will be conducted in the future to fuse effective information from multiple modalities (e.g., texture, spectral, and temporal information, etc.) of remote sensing imagery and combine them with deep learning models to improve the accuracy and robustness of the network model classification.
- High-resolution data: The resolution of remote sensing image data will increase in the future, which puts more requirements on LULC classification. Therefore, one of the focuses of future research is to develop better algorithms and models to adapt to high-resolution data to capture the details and changes in feature characteristics more accurately.
- Expert knowledge: Although deep learning and other technologies have achieved excellent results in LULC classification, it is still very important to integrate the experience and knowledge of human experts. It can not only be converted into an interpretable form of the model to improve the explanation ability of the classification results but also correct the misclassification of the model to achieve higher robustness and accuracy, which is in line with the needs of practical applications.
- Transfer learning and adaptive learning: The combination of the two can help solve the problem of low generalization ability of models due to domain differences. Future research can explore how to narrow the gap between the source and target domains due to transfer learning and optimize the model using adaptive learning to enable higher adaptation on the target domain for higher classification performance and generalization ability.
8. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Chen, B.; Xu, B.; Zhu, Z.; Yuan, C.; Suen, H.P.; Guo, J.; Xu, N.; Li, W.; Zhao, Y.; Yang, J. Stable classification with limited sample: Transferring a 30-m resolution sample set collected in 2015 to mapping 10-m resolution global land cover in 2017. Sci. Bull 2019, 64, 3. [Google Scholar]
- Zhao, C. Advances of research and application in remote sensing for agriculture. Nongye Jixie Xuebao Trans. Chin. Soc. Agric. Mach. 2014, 45, 277–293. [Google Scholar]
- Schmitt, M.; Hughes, L.H.; Qiu, C.; Zhu, X.X. SEN12MS—A Curated Dataset of Georeferenced Multi-Spectral Sentinel-1/2 Imagery for Deep Learning and Data Fusion. arXiv 2019, arXiv:1906.07789. [Google Scholar] [CrossRef]
- Li, Y.; Xia, H.; Liu, Y.; Ji, K.; Huo, L.; Ni, C. Research on Morphological Indicator Extraction Method of Pinus massoniana Lamb. Based on 3D Reconstruction. Forests 2023, 14, 1726. [Google Scholar] [CrossRef]
- Feng, Q.; Liu, J.; Gong, J. UAV remote sensing for urban vegetation mapping using random forest and texture analysis. Remote Sens. 2015, 7, 1074–1094. [Google Scholar]
- Feng, Q.; Liu, J.; Gong, J. Urban flood mapping based on unmanned aerial vehicle remote sensing and random forest classifier—A case of Yuyao, China. Water 2015, 7, 1437–1455. [Google Scholar] [CrossRef]
- Gong, W.; Lian, J.; Zhu, Y. Capacitive flexible haptic sensor based on micro-cylindrical structure dielectric layer and its decoupling study. Measurement 2023, 223, 113785. [Google Scholar] [CrossRef]
- Chapelle, O.; Vapnik, V.; Bousquet, O.; Mukherjee, S. Choosing multiple parameters for support vector machines. Mach. Learn. 2002, 46, 131–159. [Google Scholar] [CrossRef]
- Xie, C.; Zhu, H.; Fei, Y. Deep coordinate attention network for single image super-resolution. IET Image Process. 2022, 16, 273–284. [Google Scholar] [CrossRef]
- Adam, E.; Mutanga, O.; Odindi, J.; Abdel-Rahman, E.M. Land-use/cover classification in a heterogeneous coastal landscape using RapidEye imagery: Evaluating the performance of random forest and support vector machines classifiers. Int. J. Remote Sens. 2014, 35, 3440–3458. [Google Scholar] [CrossRef]
- Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
- Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.-S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef]
- Kamilaris, A.; Prenafeta-Boldú, F.X. Deep learning in agriculture: A survey. Comput. Electron. Agric. 2018, 147, 70–90. [Google Scholar] [CrossRef]
- Deren, L.; Liangpei, Z.; Guisong, X. Automatic analysis and mining of remote sensing big data. Acta Geod. Et Cartogr. Sin. 2014, 43, 1211. [Google Scholar]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- Wu, K.; Jia, Z.; Duan, Q. The Detection of Kiwifruit Sunscald Using Spectral Reflectance Data Combined with Machine Learning and CNNs. Agronomy 2023, 13, 2137. [Google Scholar] [CrossRef]
- Chen, K.; Li, W.; Chen, J.; Zou, Z.; Shi, Z. Resolution-agnostic remote sensing scene classification with implicit neural representations. IEEE Geosci. Remote Sens. Lett. 2022, 20, 6000305. [Google Scholar] [CrossRef]
- Li, B.; Wang, Q.-W.; Liang, J.-H.; Zhu, E.-Z.; Zhou, R.-Q. SquconvNet: Deep Sequencer Convolutional Network for Hyperspectral Image Classification. Remote Sens. 2023, 15, 983. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef]
- Hang, R.; Liu, Q.; Hong, D.; Ghamisi, P. Cascaded recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5384–5394. [Google Scholar] [CrossRef]
- Ma, A.; Filippi, A.M.; Wang, Z.; Yin, Z. Hyperspectral image classification using similarity measurements-based deep recurrent neural networks. Remote Sens. 2019, 11, 194. [Google Scholar] [CrossRef]
- Fan, X.; Chen, L.; Xu, X.; Yan, C.; Fan, J.; Li, X. Land Cover Classification of Remote Sensing Images Based on Hierarchical Convolutional Recurrent Neural Network. Forests 2023, 14, 1881. [Google Scholar] [CrossRef]
- Zhao, L.; Ji, S. CNN, RNN, or ViT? An Evaluation of Different Deep Learning Architectures for Spatio-Temporal Representation of Sentinel Time Series. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 16, 44–56. [Google Scholar] [CrossRef]
- Sun, C.; Zhang, X.; Meng, H.; Cao, X.; Zhang, J.; Jiao, L. Dual-Branch Spectral-Spatial Adversarial Representation Learning for Hyperspectral Image Classification with Few Labeled Samples. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 1–15. [Google Scholar] [CrossRef]
- Dieste, Á.G.; Argüello, F.; Heras, D.B. ResBaGAN: A Residual Balancing GAN with Data Augmentation for Forest Mapping. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 6428–6447. [Google Scholar] [CrossRef]
- Jaiswal, G.; Rani, R.; Mangotra, H.; Sharma, A. Integration of hyperspectral imaging and autoencoders: Benefits, applications, hyperparameter tunning and challenges. Comput. Sci. Rev. 2023, 50, 100584. [Google Scholar] [CrossRef]
- Xu, C.; Gao, M.; Yan, J.; Jin, Y.; Yang, G.; Wu, W. MP-Net: An efficient and precise multi-layer pyramid crop classification network for remote sensing images. Comput. Electron. Agric. 2023, 212, 108065. [Google Scholar] [CrossRef]
- Yang, J.; Du, B.; Zhang, L. From center to surrounding: An interactive learning framework for hyperspectral image classification. ISPRS J. Photogramm. Remote Sens. 2023, 197, 145–166. [Google Scholar] [CrossRef]
- Baumgardner, M.F.; Biehl, L.L.; Landgrebe, D.A. 220 band aviris hyperspectral image data set: June 12, 1992 indian pine test site 3. Purdue Univ. Res. Repos. 2015, 10, 991. [Google Scholar]
- del Pais Vasco, U. Available online: http://www.ehu.es/ccwintco/index.php/Hyperspectral-Remote-Sensing-Scenes (accessed on 25 August 2012).
- Wang, J.; Zheng, Z.; Ma, A.; Lu, X.; Zhong, Y. LoveDA: A remote sensing land-cover dataset for domain adaptive semantic segmentation. arXiv 2021, arXiv:2110.08733. [Google Scholar]
- Alemohammad, H.; Booth, K. LandCoverNet: A global benchmark land cover classification training dataset. arXiv 2020, arXiv:2012.03111. [Google Scholar]
- Hughes, M.J.; Hayes, D.J. Automated detection of cloud and cloud shadow in single-date Landsat imagery using neural networks and spatial post-processing. Remote Sens. 2014, 6, 4907–4926. [Google Scholar] [CrossRef]
- Tong, X.-Y.; Xia, G.-S.; Lu, Q.; Shen, H.; Li, S.; You, S.; Zhang, L. Land-cover classification with high-resolution remote sensing images using transferable deep models. Remote Sens. Environ. 2020, 237, 111322. [Google Scholar] [CrossRef]
- Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; pp. 270–279. [Google Scholar]
- Xia, G.-S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A benchmark data set for performance evaluation of aerial scene classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar] [CrossRef]
- Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
- Helber, P.; Bischke, B.; Dengel, A.; Borth, D. Eurosat: A novel dataset and deep learning benchmark for land use and land cover classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2217–2226. [Google Scholar] [CrossRef]
- Sumbul, G.; Charfuelan, M.; Demir, B.; Markl, V. Bigearthnet: A large-scale benchmark archive for remote sensing image understanding. In Proceedings of the IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium, San Jose, CA, USA, 2–5 November 2019; pp. 5901–5904. [Google Scholar]
- Alhichri, H.; Alswayed, A.S.; Bazi, Y.; Ammour, N.; Alajlan, N.A. Classification of remote sensing images using EfficientNet-B3 CNN model with attention. IEEE Access 2021, 9, 14078–14094. [Google Scholar] [CrossRef]
- Zhang, P.; Bai, Y.; Wang, D.; Bai, B.; Li, Y. A meta-learning framework for few-shot classification of remote sensing scene. In Proceedings of the ICASSP 2021–2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 4590–4594. [Google Scholar]
- Moharram, M.A.; Sundaram, D.M. Land Use and Land Cover Classification with Hyperspectral Data: A comprehensive review of methods, challenges and future directions. Neurocomputing 2023, 536, 90–113. [Google Scholar] [CrossRef]
- Temenos, A.; Temenos, N.; Kaselimi, M.; Doulamis, A.; Doulamis, N. Interpretable deep learning framework for land use and land cover classification in remote sensing using SHAP. IEEE Geosci. Remote Sens. Lett. 2023, 20, 8500105. [Google Scholar] [CrossRef]
- Pei, H.; Owari, T.; Tsuyuki, S.; Zhong, Y. Application of a Novel Multiscale Global Graph Convolutional Neural Network to Improve the Accuracy of Forest Type Classification Using Aerial Photographs. Remote Sens. 2023, 15, 1001. [Google Scholar] [CrossRef]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. pp. 234–241. [Google Scholar]
- Shao, Z.; Zhou, W.; Deng, X.; Zhang, M.; Cheng, Q. Multilabel remote sensing image retrieval based on fully convolutional network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 318–328. [Google Scholar] [CrossRef]
- Shao, Z.; Yang, K.; Zhou, W. Performance evaluation of single-label and multi-label remote sensing image retrieval using a dense labeling dataset. Remote Sens. 2018, 10, 964. [Google Scholar]
- Ma, X.; Man, Q.; Yang, X.; Dong, P.; Yang, Z.; Wu, J.; Liu, C. Urban Feature Extraction within a Complex Urban Area with an Improved 3D-CNN Using Airborne Hyperspectral Data. Remote Sens. 2023, 15, 992. [Google Scholar] [CrossRef]
- Khan, S.D.; Basalamah, S. Multi-Branch Deep Learning Framework for Land Scene Classification in Satellite Imagery. Remote Sens. 2023, 15, 3408. [Google Scholar]
- Zhao, B.; Zhong, Y.; Xia, G.-S.; Zhang, L. Dirichlet-derived multiple topic scene classification model for high spatial resolution remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2015, 54, 2108–2123. [Google Scholar] [CrossRef]
- Tong, X.-Y.; Xia, G.-S.; Zhu, X.X. Enabling country-scale land cover mapping with meter-resolution satellite imagery. ISPRS J. Photogramm. Remote Sens. 2023, 196, 178–196. [Google Scholar] [CrossRef]
- Zagoruyko, S.; Komodakis, N. Learning to compare image patches via convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4353–4361. [Google Scholar]
- Tsai, Y.-H.; Hung, W.-C.; Schulter, S.; Sohn, K.; Yang, M.-H.; Chandraker, M. Learning to adapt structured output space for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7472–7481. [Google Scholar]
- Vu, T.-H.; Jain, H.; Bucher, M.; Cord, M.; Pérez, P. Advent: Adversarial entropy minimization for domain adaptation in semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2517–2526. [Google Scholar]
- Luo, Y.; Zheng, L.; Guan, T.; Yu, J.; Yang, Y. Taking a closer look at domain shift: Category-level adversaries for semantics consistent domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2507–2516. [Google Scholar]
- Wang, H.; Shen, T.; Zhang, W.; Duan, L.-Y.; Mei, T. Classes matter: A fine-grained adversarial approach to cross-domain semantic segmentation. In Proceedings of the European Conference on Computer Vision, Online, 23–28 August 2020; pp. 642–659. [Google Scholar]
- Wang, H.; Chang, W.; Yao, Y.; Yao, Z.; Zhao, Y.; Li, S.; Liu, Z.; Zhang, X. Cropformer: A new generalized deep learning classification approach for multi-scenario crop classification. Front. Plant Sci. 2023, 14, 1130659. [Google Scholar] [CrossRef]
- Singh, A.; Bruzzone, L. WIANet: A Wavelet-Inspired Attention-Based Convolution Neural Network for Land Cover Classification. IEEE Geosci. Remote Sens. Lett. 2022, 20, 5000305. [Google Scholar] [CrossRef]
- Mallat, S.G. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 674–693. [Google Scholar] [CrossRef]
- Zhou, Z.; Li, S.; Wu, W.; Guo, W.; Li, X.; Xia, G.; Zhao, Z. NaSC-TG2: Natural scene classification with Tiangong-2 remotely sensed imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 3228–3242. [Google Scholar] [CrossRef]
- Zhou, W.; Newsam, S.; Li, C.; Shao, Z. PatternNet: A benchmark dataset for performance evaluation of remote sensing image retrieval. ISPRS J. Photogramm. Remote Sens. 2018, 145, 197–209. [Google Scholar] [CrossRef]
- Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
- Radosavovic, I.; Kosaraju, R.P.; Girshick, R.; He, K.; Dollár, P. Designing network design spaces. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10428–10436. [Google Scholar]
- Ansith, S.; Bini, A. Land use classification of high resolution remote sensing images using an encoder based modified GAN architecture. Displays 2022, 74, 102229. [Google Scholar]
- Ma, A.; Yu, N.; Zheng, Z.; Zhong, Y.; Zhang, L. A supervised progressive growing generative adversarial network for remote sensing image scene classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5618818. [Google Scholar] [CrossRef]
- Miao, W.; Geng, J.; Jiang, W. Semi-supervised remote-sensing image scene classification using representation consistency siamese network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5616614. [Google Scholar] [CrossRef]
- Xu, C.; Shu, J.; Zhu, G. Adversarial Remote Sensing Scene Classification Based on Lie Group Feature Learning. Remote Sens. 2023, 15, 914. [Google Scholar] [CrossRef]
- Wang, J.; Liu, B.; Zhou, Y.; Zhao, J.; Xia, S.; Yang, Y.; Zhang, M.; Ming, L.M. Semisupervised multiscale generative adversarial network for semantic segmentation of remote sensing image. IEEE Geosci. Remote Sens. Lett. 2020, 19, 8003805. [Google Scholar] [CrossRef]
- Tang, Y.; Qiu, F.; Wang, B.; Wu, D.; Jing, L.; Sun, Z. A deep relearning method based on the recurrent neural network for land cover classification. GIScience Remote Sens. 2022, 59, 1344–1366. [Google Scholar] [CrossRef]
- Huang, X.; Lu, Q.; Zhang, L.; Plaza, A. New postprocessing methods for remote sensing image classification: A systematic study. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7140–7159. [Google Scholar] [CrossRef]
- Tao, C.; Lu, W.; Qi, J.; Wang, H. Spatial information considered network for scene classification. IEEE Geosci. Remote Sens. Lett. 2020, 18, 984–988. [Google Scholar] [CrossRef]
- Sohail, M.; Chen, Z.; Yang, B.; Liu, G. Multiscale spectral-spatial feature learning for hyperspectral image classification. Displays 2022, 74, 102278. [Google Scholar] [CrossRef]
- Zhang, X.; Sun, Y.; Jiang, K.; Li, C.; Jiao, L.; Zhou, H. Spatial sequential recurrent neural network for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 4141–4155. [Google Scholar] [CrossRef]
- Ibanez, D.; Fernandez-Beltran, R.; Pla, F.; Yokoya, N. Masked auto-encoding spectral–spatial transformer for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5542614. [Google Scholar] [CrossRef]
- Chen, Y.; Jiao, L.; Li, Y.; Zhao, J. Multilayer projective dictionary pair learning and sparse autoencoder for PolSAR image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 6683–6694. [Google Scholar] [CrossRef]
- Liu, G.; Li, Y.; Jiao, L.; Chen, Y.; Shang, R. Multiobjective evolutionary algorithm assisted stacked autoencoder for PolSAR image classification. Swarm Evol. Comput. 2021, 60, 100794. [Google Scholar] [CrossRef]
- Mughees, A.; Tao, L. Hyperspectral image classification based on deep auto-encoder and hidden Markov random field. In Proceedings of the 2017 13th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), Guilin, China, 29–31 July 2017; pp. 59–65. [Google Scholar]
- Mughees, A.; Tao, L. Hyper-voxel based deep learning for hyperspectral image classification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 840–844. [Google Scholar]
- Chen, C.; Zhang, J.; Li, T.; Yan, Q.; Xun, L. Spectral and Multi-Spatial-Feature Based Deep Learning for Hyperspectral Remote Sensing Image Classification. In Proceedings of the 2018 IEEE International Conference on Real-time Computing and Robotics (RCAR), Kandima, Maldives, 1–5 August 2018; pp. 421–426. [Google Scholar]
- Mughees, A.; Tao, L. Efficient deep auto-encoder learning for the classification of hyperspectral images. In Proceedings of the 2016 International Conference on Virtual Reality and Visualization (ICVRV), Hangzhou, China, 24–26 September 2016; pp. 44–51. [Google Scholar]
- Liu, J.; Wang, Y. Water body extraction in remote sensing imagery using domain adaptation-based network embedding selective self-attention and multi-scale feature fusion. Remote Sens. 2022, 14, 3538. [Google Scholar] [CrossRef]
- Soto, P.J.; Costa, G.A.; Feitosa, R.Q.; Ortega, M.X.; Bermudez, J.D.; Turnes, J.N. Domain-adversarial neural networks for deforestation detection in tropical forests. IEEE Geosci. Remote Sens. Lett. 2022, 19, 2504505. [Google Scholar] [CrossRef]
- Bai, L.; Du, S.; Zhang, X.; Wang, H.; Liu, B.; Ouyang, S. Domain adaptation for remote sensing image semantic segmentation: An integrated approach of contrastive learning and adversarial learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5628313. [Google Scholar] [CrossRef]
- Scott, G.J.; England, M.R.; Starms, W.A.; Marcum, R.A.; Davis, C.H. Training deep convolutional neural networks for land–cover classification of high-resolution imagery. IEEE Geosci. Remote Sens. Lett. 2017, 14, 549–553. [Google Scholar] [CrossRef]
- Yi, Y.; You, Y.; Li, C.; Zhou, W. EFM-Net: An Essential Feature Mining Network for Target Fine-Grained Classification in Optical Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5606416. [Google Scholar] [CrossRef]
- Haut, J.M.; Paoletti, M.E.; Plaza, J.; Li, J.; Plaza, A. Active learning with convolutional neural networks for hyperspectral image classification using a new Bayesian approach. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6440–6461. [Google Scholar] [CrossRef]
- Lei, Z.; Zeng, Y.; Liu, P.; Su, X. Active deep learning for hyperspectral image classification with uncertainty learning. IEEE Geosci. Remote Sens. Lett. 2021, 19, 5502405. [Google Scholar] [CrossRef]
- Li, A.; Lu, Z.; Wang, L.; Xiang, T.; Wen, J.-R. Zero-shot scene classification for high spatial resolution remote sensing images. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4157–4167. [Google Scholar] [CrossRef]
- Li, Y.; Zhu, Z.; Yu, J.-G.; Zhang, Y. Learning deep cross-modal embedding networks for zero-shot remote sensing image scene classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10590–10603. [Google Scholar] [CrossRef]
- Chen, Y.; Li, Y.; Mao, H.; Chai, X.; Jiao, L. A Novel Deep Nearest Neighbor Neural Network for Few-Shot Remote Sensing Image Scene Classification. Remote Sens. 2023, 15, 666. [Google Scholar] [CrossRef]
- Jiang, N.; Shi, H.; Geng, J. Multi-Scale Graph-Based Feature Fusion for Few-Shot Remote Sensing Image Scene Classification. Remote Sens. 2022, 14, 5550. [Google Scholar] [CrossRef]
Dataset | Data Sources | Number | Image Dimensions | Spatial Resolution | Number of Bands | Number of Categories | Year |
---|---|---|---|---|---|---|---|
SPARCS [37] | Landsat8 | 80 | 1000 × 1000 | 30 | 10 | 7 | 2014 |
LoveDA [35] | GoogleEarth | 5987 | 1024 × 1024 | 0.3 | 3 | 7 | 2021 |
Kennedy Space Center [34] | AVIRIS | 1 | 614 × 512 | 18 | 224 | 13 | 2014 |
Pavia University [34] | ROSIS | 1 | 610 × 340 | 1.3 | 103 | 9 | 2015 |
GID [38] | GF-2 | 150 | 6800 × 7200 | 1/4 | 4 | 15 | 2020 |
Indian Pines [33,34] | AVIRIS | 1 | 145 × 145 | 20 | 224 | 16 | 2015 |
LandCoverNet [36] | Sentinel-1/2 Lantsat-8 | 1980 | 256 × 256 | 10 | 10 | 7 | 2020 |
Dataset | Data Sources | Number | Image Dimensions | Spatial Resolution | Number of Bands | Number of Categories | Year |
---|---|---|---|---|---|---|---|
BigEarthNet [42,43] | Sentinel-2 | 590,326 | 120 × 120 | 10 | 13 | 43 | 2019 |
RESISC45 [41] | WorldView-2 | 31,500 | 256 × 256 | 2 | 3 | 45 | 2016 |
EuroSAT [42] | Sentinel-2 | 27,000 | 64 × 64 | 10/20/60 | 13 | 10 | 2019 |
AID [40] | Google Earth | 10,000 | 600 × 600 | 2 | 3 | 30 | 2017 |
UC MERCED [39] | Aerial imagery | 2100 | 256 × 256 | 0.3 | 3 | 21 | 2010 |
OPTIMAL-31 [44] | Google Earth | 1860 | 256 × 256 | - | 3 | 31 | 2019 |
RSD46-WHU [45] | GoogleEarth, Tianditu | 117,000 | 256 × 256 | 0.5–2 | 3 | 46 | 2017 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhao, S.; Tu, K.; Ye, S.; Tang, H.; Hu, Y.; Xie, C. Land Use and Land Cover Classification Meets Deep Learning: A Review. Sensors 2023, 23, 8966. https://doi.org/10.3390/s23218966
Zhao S, Tu K, Ye S, Tang H, Hu Y, Xie C. Land Use and Land Cover Classification Meets Deep Learning: A Review. Sensors. 2023; 23(21):8966. https://doi.org/10.3390/s23218966
Chicago/Turabian StyleZhao, Shengyu, Kaiwen Tu, Shutong Ye, Hao Tang, Yaocong Hu, and Chao Xie. 2023. "Land Use and Land Cover Classification Meets Deep Learning: A Review" Sensors 23, no. 21: 8966. https://doi.org/10.3390/s23218966