Parameter Analysis of Multiscale Two-Dimensional Fuzzy and Dispersion Entropy Measures Using Machine Learning Classification
Abstract
:1. Introduction
2. Materials and Methods
2.1. Datasets
2.2. Experimental Procedure
2.2.1. Data Pre-Processing
2.2.2. Feature Extraction
2.2.3. Fuzzy Entropy
2.2.4. Dispersion Entropy
2.2.5. Multiscale Approach
- Construct the coarse-grained images as
- Compute or of each coarse-grained image.
2.2.6. Classification
Naive Bayes
Decision Tree
Support Vector Machine
Multi-Layer Perceptron
K-Nearest Neighbour
2.2.7. Experimental Procedure
3. Results and Discussion
3.1. Epistroma
3.1.1. Classification Accuracy
- Classification Performance: As shown in Table 1, across all possible parameter and multiscale combinations, outperformed regarding average max classification accuracy for both image sizes, achieving accuracies of 98.84% and 100% for 50 × 50 px and 100 × 100 px images, respectively. However, achieved a higher average classification accuracy than for both image sizes, producing accuracies of 93.66% and 94.77% for 50 × 50 px and 100 × 100 px images, respectively.
- Parameter Optimization: Figure 2 shows a color map which displays the performance of different parameters with respect to each classifier for the and algorithms. In the case, for images of size 50 × 50 px, the value of m did not appear to have a significant influence on the classification performance, while a value of and larger values of r are recommended. Finally, for images of size 100 × 100 px, larger values for m, n, and r are recommended. In the case, for images of size 50 × 50 px, is recommended as all five classifiers performed optimally with this parameter value, while a c value of 3 or 6 is advised. Additionally, for images of size 100 × 100 px, larger values of m and c are favored.Additionally, in , parameter combinations of , , , , , and , , resulted in an average accuracy greater than 85% for all the classifiers and both image sizes. For images of size 50 × 50 px using , a parameter combination of and achieved an average classification accuracy greater than 85% for four out of the five classifiers, while a combination of and achieved an average accuracy greater than 85% for all the five classifiers. Similarly, for images of size 100 × 100 px, parameter combinations of and achieved an average accuracy greater than 85% for three out of the five classifiers. We should note that the 85% classification accuracy was chosen as a threshold for the choice of parameter combinations across different time scales, classifiers, and datasets for the purpose of providing a general application to datasets and classifiers that was not explored in our experiments. Additionally, we believe there exists many techniques in which we could increase our classification accuracy, such as by bagging which can improve the stability and accuracy of classification algorithms [39]. These techniques are not explored in our work as we are only investigating the influence that key parameters exhibit on the classification performance of features extracted by the and algorithms.
- Multiscale Analysis: Table 2 shows the values of , which resulted in the highest average classification accuracy for each classifier and image size. Results show that for all five classifiers, results in the highest average classification accuracy, indicating that complexity analysis provides a stronger textural description than irregularity analysis for the images found in this biomedical dataset.
3.1.2. Computation Time
3.2. KTH-TIPS
Classification Accuracy
- Classification Performance: As can be seen in Table 3, across all the possible parameter, texture, and multiscale combinations, outperformed in both average classification and max average classification accuracy. Additionally, the KNN classifier achieved the highest average and maximum average classification accuracy for both entropy techniques.
- Parameter Optimization: Figure 3 shows a color map that displays the performance of different parameters with respect to each classifier for the and algorithms. These results indicate that a majority of the classifiers (three out of five) achieve optimal classification performance for a parameter combination of , , and . Similarly, Figure 4 indicates that lower values for the parameters m and c produced a higher classification accuracy for textures extracted by the algorithm. Additionally, for , all parameter combinations resulted in an average accuracy greater than 85% for all the classifiers in the experiments containing the texture aluminium, while no parameter combination resulted in an average accuracy greater than 85% for any classifier when using .
- Texture Analysis: Figure 4 displays each texture vs. texture test alongside which entropy algorithm achieved the highest average classification accuracy, where a green cell represents an average classification accuracy greater than 85%, a yellow cell represents an accuracy between 70% and 85%, and a red cell represents an accurancy of less than 70%. Results indicate that performed extremely well in every test involving aluminium. Moreover, across the board, outperformed or matched the classification performance of on all the texture combinations. Furthermore, both entropy techniques performed poorly on a majority of the tests involving the texture corduroy.
- Multiscale Analysis: Table 4 displays the average classification accuracy for the multiscale extensions and across different scale factors. The results show that the texture images examined in this study contained complex structures across multiple spatial scales: for , the average classification accuracy increased for all the classifiers.
3.3. Main Findings and Significance of the Work
- Textural features extracted by resulted in better classification performance than those extracted by the as a majority.
- In , for images of size 50 × 50 px from the Epistroma dataset, parameter combinations of (1) and , and (2) and achieved an average classification accuracy greater than 85% for four out of the five classifiers and for all five classifiers, respectively. For images of size 100 × 100 px, parameter combinations of and achieved an average accuracy greater than 85% for three out of the five classifiers. Additionally, for images in the KTH-TIPS dataset, no parameter combination resulted in an average accuracy greater than 85%.
- In , for images in the Epistroma dataset, parameter combinations of (1) , , , , , and , and (2) , , and resulted in an average accuracy greater than 85% for both image sizes. Additionally, for images in the KTH-TIPS dataset, all parameter combinations resulted in an average accuracy greater than 85% for experiments containing the texture aluminium.
- The computation time of was invariant to changes in parameter values. Contrarily, larger values of m and c increased the computation time of exponentially. Furthermore, was computationally faster than for lower values of m and c. However, in most cases, this lowered the classification performance.
- The multiscale version of entropy measures led to the creation of a vector of entropy values. Our results reveal that, when the vector of entropy values is applied to the classifier, the subsequent results show improved classification accuracy. This shows that the texture of coarse-grained versions of images provides information for classification purposes.
- In most cases, the choice of classifier did not have a significant impact on the classification of the extracted features by both entropy algorithms.
4. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Cover, T.M. Elements of Information Theory; John Wiley & Sons: Hoboken, NJ, USA, 1999. [Google Scholar]
- Mohammad-Djafari, A. Entropy, information theory, information geometry and Bayesian inference in data, signal and image processing and inverse problems. Entropy 2015, 17, 3989–4027. [Google Scholar] [CrossRef]
- Shang, Y. Bounding extremal degrees of edge-independent random graphs using relative entropy. Entropy 2016, 18, 53. [Google Scholar] [CrossRef] [Green Version]
- Humeau-Heurtier, A. Entropy Analysis in Health Informatics. In Signal Processing Techniques for Computational Health Informatics; Springer: Berlin/Heidelberg, Germany; pp. 123–143.
- Furutani, N.; Takahashi, T.; Naito, N.; Maruishi, T.; Yoshimura, Y.; Hasegawa, C.; Hirosawa, T.; Kikuchi, M. Complexity of body movements during sleep in children with autism spectrum disorder. Entropy 2021, 23, 418. [Google Scholar] [CrossRef]
- Humeau-Heurtier, A. Evaluation of systems’ irregularity and complexity: Sample entropy, its derivatives, and their applications across scales and disciplines. Entropy 2018, 20, 794. [Google Scholar] [CrossRef] [Green Version]
- Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J.-Physiol.-Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [Green Version]
- Bandt, C.; Pompe, B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef]
- Chen, W.; Wang, Z.; Xie, H.; Yu, W. Characterization of surface EMG signal based on fuzzy entropy. IEEE Trans. Neural Syst. Rehabil. Eng. 2007, 15, 266–272. [Google Scholar] [CrossRef] [PubMed]
- Rostaghi, M.; Azami, H. Dispersion entropy: A measure for time-series analysis. IEEE Signal Process. Lett. 2016, 23, 610–614. [Google Scholar] [CrossRef]
- Silva, L.E.V.; Senra Filho, A.C.S.; Fazan, V.P.S.; Felipe, J.C.; Junior, L.M. Two-dimensional sample entropy: Assessing image texture through irregularity. Biomed. Phys. Eng. Express 2016, 2, 045002. [Google Scholar] [CrossRef]
- Ribeiro, H.V.; Zunino, L.; Lenzi, E.K.; Santoro, P.A.; Mendes, R.S. Complexity-entropy causality plane as a complexity measure for two-dimensional patterns. PLoS ONE 2012, 7, e40689. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Morel, C.; Humeau-Heurtier, A. Multiscale permutation entropy for two-dimensional patterns. Pattern Recognit. Lett. 2021, 150, 139–146. [Google Scholar] [CrossRef]
- Hilal, M.; Berthin, C.; Martin, L.; Azami, H.; Humeau-Heurtier, A. Bidimensional multiscale fuzzy entropy and its application to pseudoxanthoma elasticum. IEEE Trans. Biomed. Eng. 2019, 67, 2015–2022. [Google Scholar] [CrossRef]
- Azami, H.; da Silva, L.E.V.; Omoto, A.C.M.; Humeau-Heurtier, A. Two-dimensional dispersion entropy: An information-theoretic method for irregularity analysis of images. Signal Process. Image Commun. 2019, 75, 178–187. [Google Scholar] [CrossRef]
- Humeau-Heurtier, A. Texture feature extraction methods: A survey. IEEE Access 2019, 7, 8975–9000. [Google Scholar] [CrossRef]
- Backes, A.R.; Junior, J.J.D.M.S.; Kolb, R.M. A gravitational model for plant classification using adaxial epidermis texture. In Proceedings of the International Conference on Image Analysis and Processing, Genoa, Italy, 7–11 September 2015; pp. 89–96. [Google Scholar]
- Xu, D.; Chen, X.; Xie, Y.; Yang, C.; Gui, W. Complex networks-based texture extraction and classification method for mineral flotation froth images. Miner. Eng. 2015, 83, 105–116. [Google Scholar] [CrossRef]
- Zhang, Q.; Xiao, Y.; Chen, S.; Wang, C.; Zheng, H. Quantification of elastic heterogeneity using contourlet-based texture analysis in shear-wave elastography for breast tumor classification. Ultrasound Med. Biol. 2015, 41, 588–600. [Google Scholar] [CrossRef] [PubMed]
- Shi, X.; Cheng, H.D.; Hu, L.; Ju, W.; Tian, J. Detection and classification of masses in breast ultrasound images. Digit. Signal Process. 2010, 20, 824–836. [Google Scholar] [CrossRef]
- Humeau-Heurtier, A. The multiscale entropy algorithm and its variants: A review. Entropy 2015, 17, 3110–3123. [Google Scholar] [CrossRef] [Green Version]
- Baldini, G.; Hernez, R.; Ramos, J.L.; Amerini, I. Intrusion detection based on gray-level co-occurrence matrix and 2D dispersion entropy. Appl. Sci. 2021, 11, 5567. [Google Scholar] [CrossRef]
- EGFR colon TMA Stroma LBP Classification. Available online: http://fimm.webmicroscope.net/Research/Supplements/epistroma (accessed on 1 April 2021).
- The KTH-TIPS and KTH-TIPS2 Image Databases. Available online: https://www.csc.kth.se/cvap/databases/kth-tips/credits.html (accessed on 1 April 2021).
- Linder, N.; Konsti, J.; Turkki, R.; Rahtu, E.; Lundin, M.; Nordling, S.; Haglund, C.; Pietikäinen, M.; Ahonen, T.; Lundin, J. Identification of tumor epithelium and stroma in tissue microarrays using texture analysis. Diagn. Pathol. 2012, 7, 1–11. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Hayman, E.; Caputo, B.; Fritz, M.; Eklundh, J.O. On the significance of real-world conditions for material classification. In Proceedings of the European Conference on Computer Vision, Berlin, Germany, 11–14 May 2004; pp. 253–266. [Google Scholar]
- MATLAB, version 9.8.0.1323502 (2020a). In MATLAB; The MathWorks Inc.: Three Apple Hill Drive, Natick, MA, USA, 2010.
- Chen, W.; Zhuang, J.; Yu, W.; Wang, Z. Measuring complexity using fuzzyen, apen, and sampen. Med Eng. Phys. 2009, 31, 61–68. [Google Scholar] [CrossRef] [PubMed]
- Xie, H.B.; Chen, W.T.; He, W.X.; Liu, H. Complexity analysis of the biomedical signal using fuzzy entropy measurement. Appl. Soft Comput. 2011, 11, 2871–2879. [Google Scholar] [CrossRef]
- Azami, H.; Escudero, J. Amplitude-and fluctuation-based dispersion entropy. Entropy 2018, 20, 210. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Humeau-Heurtier, A.; Omoto, A.C.M.; Silva, L.E. Bi-dimensional multiscale entropy: Relation with discrete Fourier transform and biomedical application. Comput. Biol. Med. 2018, 100, 36–40. [Google Scholar] [CrossRef] [PubMed]
- Silva, L.E.; Duque, J.J.; Felipe, J.C.; Murta, L.O., Jr.; Humeau-Heurtier, A. Two-dimensional multiscale entropy analysis: Applications to image texture evaluation. Signal Process. 2018, 147, 224–232. [Google Scholar] [CrossRef]
- Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
- Zhang, H. The optimality of Naive Bayes. In Proceedings of the 17th International Florida Artificial Intelligence Research Society Conference, Miami Beach, Florida, 12–14 May 2004; Volume 2, pp. 562–567. [Google Scholar]
- Dumont, M.; Maree, R.; Wehenkel, L. Fast multi-class image annotation with random subwindows and multiple output randomized trees. In Proceedings of the VISAPP 2009—4th International Conference on Computer Vision Theory and Applications, Lisboa, Portugal, 5–8 February 2009; Volume 2, pp. 196–203. [Google Scholar]
- Wu, T.; Lin, C.; Weng, R. Probability estimates for multi-class classification by pairwise coupling. J. Mach. Learn. Res. 2004, 5, 975–1005. [Google Scholar]
- Kingma, D.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
- Bar-Hillel, A.; Hertz, T.; Shental, N. Learning Distance Functions using Equivalence Relations. In Proceedings of the 20th International Conference on Machine Learning, Washington, DC, USA, 21–24 August 2003; Volume 5, pp. 11–18. [Google Scholar]
- Opitz, D.; Maclin, R. Popular Ensemble Methods: An Empirical Study. J. Artif. Intell. Res. 1998, 11, 169–198. [Google Scholar] [CrossRef]
- Bianconi, F.; Bello-Cerezo, R.; Napoletano, P. Improved opponent color local binary patterns: An effective local image descriptor for color texture classification. J. Electron. Imaging 2017, 27, 011002. [Google Scholar] [CrossRef]
- Bello-Cerezo, R.; Fieguth, P.; Bianconi, F. LBP-motivated colour texture classification. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
- Kather, J.N.; Bello-Cerezo, R.; Di Maria, F.; van Pelt, G.W.; Mesker, W.E.; Halama, N.; Bianconi, F. Classification of tissue regions in histopathological images: Comparison between pre-trained convolutional neural networks and local binary patterns variants. In Deep Learners and Deep Learner Descriptors for Medical Applications; Springer: Berlin/Heidelberg, Germany, 2020; pp. 95–115. [Google Scholar]
Entropy Algorithm | Image Size | Average Accuracy | Average Max Accuracy |
---|---|---|---|
50 × 50 | 93.40 | 98.84 | |
100 × 100 | 94.15 | 100 | |
50 × 50 | 93.66 | 97.09 | |
100 × 100 | 94.77 | 96.80 |
50 × 50 | 100 × 100 | |||
---|---|---|---|---|
Classifier | ||||
Decision tree | ||||
Naive Bayes | ||||
SVM | ||||
MLP | ||||
KNN |
Average Accuracy | Max Average Accuracy | |||
---|---|---|---|---|
Classifier | ||||
Decision tree | 79.52 | 67.31 | 100 | 98.22 |
Naive Bayes | 74.99 | 56.26 | 100 | 95.09 |
SVM | 80.55 | 66.17 | 100 | 95.69 |
MLP | 73.78 | 55.40 | 100 | 97.32 |
KNN | 83.18 | 71.46 | 100 | 100 |
Classifier | ||
---|---|---|
Decision Tree | ||
Naive Bayes | ||
SVM | ||
MLP | ||
KNN |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Furlong, R.; Hilal, M.; O’Brien, V.; Humeau-Heurtier, A. Parameter Analysis of Multiscale Two-Dimensional Fuzzy and Dispersion Entropy Measures Using Machine Learning Classification. Entropy 2021, 23, 1303. https://doi.org/10.3390/e23101303
Furlong R, Hilal M, O’Brien V, Humeau-Heurtier A. Parameter Analysis of Multiscale Two-Dimensional Fuzzy and Dispersion Entropy Measures Using Machine Learning Classification. Entropy. 2021; 23(10):1303. https://doi.org/10.3390/e23101303
Chicago/Turabian StyleFurlong, Ryan, Mirvana Hilal, Vincent O’Brien, and Anne Humeau-Heurtier. 2021. "Parameter Analysis of Multiscale Two-Dimensional Fuzzy and Dispersion Entropy Measures Using Machine Learning Classification" Entropy 23, no. 10: 1303. https://doi.org/10.3390/e23101303