Enhancing Student Academic Success Prediction Through Ensemble Learning and Image-Based Behavioral Data Transformation
Abstract
:1. Introduction
- (1)
- This study introduces a novel feature integration approach by combining high-dimensional image features extracted by a CNN from student data transformed into images using methods such as Pixel Representation (PR), Sine Wave Transformation (SWT), Recurrence Plot (RP), and Gramian Angular Field (GAF) with low-dimensional numerical features extracted by an FCN from the original data. The feature-level fusion mechanism leverages concatenated features from both networks, enabling the model to automatically learn optimal combinations of complementary information from heterogeneous data sources for improved classification accuracy.
- (2)
- A feature-level fusion mechanism is introduced, where the output feature vectors from the CNN and FCN are concatenated to establish a unified feature structure. Instead of direct weighting, this method empowers the ensemble framework to learn an optimal combination of features through end-to-end training, capturing complex temporal and nonlinear relationships from image data while preserving essential numerical details from data, leveraging the complementary strengths of both networks and improving classification robustness.
- (3)
- To benchmark the proposed approach, a range of machine learning techniques, such as Stochastic Gradient Descent (SGD), Gradient Boosting (GB), Decision Tree (DT), Extra Tree (ET), K-Nearest Neighbor (KNN), Support Vector Machine (SVM), Naive Bayes (NB), Logistic Regression (LR), AdaBoost, and Random Forest (RF) are employed for performance comparison. Meanwhile, several deep learning methods, consisting of a Convolutional Neural Network (CNN), Fully Connected Network (FCN), MobileNet, and EfficientNetB4, are used for performance comparison. The evaluation metrics enable a thorough evaluation of the proposed framework against cutting-edge methods.
2. Related Work
2.1. Machine Learning with Student Success Prediction
2.2. Deep Learning with Image-Based Data Transformation Techniques
2.3. Ensemble Learning Approaches with Student Success Prediction
3. Materials and Methods
3.1. Dataset
3.2. Data Preprocessing
3.3. Image-Based Transformation Techniques
3.3.1. Pixel Representation
3.3.2. Sine Wave Transformation
3.3.3. Recurrence Plot
3.3.4. Gramian Angular Field
3.4. Machine Learning Models
3.5. Deep Learning Models
3.5.1. Conventional Neural Network
3.5.2. Fully Connected Network
3.5.3. MobileNet
3.5.4. EfficientNetB4
3.6. Proposed Methodology
Algorithm 1: The proposed EnCF algorithm |
Require: Training data , where is the input feature vector, and is the target classes. |
Require: Ensemble Model EnCF combining CNN with FCN. |
Ensure: Predicted target classes for test data. 1: Split into training set and test set . |
2: Use SMOTE to resample to handle class imbalance. |
3: Train CNN on image features from . |
4: Train FCN on raw features from . 5: Combine CNN and FCN outputs using a fully connected layer for final prediction. 6: Ensemble Prediction: 7: For each sample, in : 8: Compute image-based features using CNN. 9: Compute raw features using FCN. 10: Combine CNN and FCN outputs to predict class . 11: End for Return: Predicted target classes for test data. |
3.7. Evaluation Metrics
4. Results and Discussion
4.1. Experiment Setup
4.2. Experimental Results of 1D Raw Data
4.2.1. Results for Machine Learning Models with 1D Raw Data
4.2.2. Results for Deep Learning Models with 1D Raw Data
4.2.3. Results of All Models with 1D Raw Data
4.3. Results of Deep Learning with Four Feature Transformation Methods for 2D Data
4.4. Results of Ensemble Learning with Four Feature Transformation Methods for 2D Data
4.5. Results of All Models Using 1D and 2D Data
4.6. Comparison with Previous Studies
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
PR | Pixel Representation |
SWT | Sine Wave Transformation |
RP | Recurrence Plot |
GAF | Gramian Angular Field |
CNN | Convolutional Neural Network |
OULAD | Open University Learning Analytics Dataset |
FCN | Fully Convolutional Network |
EnCF | Ensemble CNN and FCN |
SMOTE | Synthetic Minority Over-sampling Technique |
RF | Random Forest |
SVM | Support Vector Machine |
ET | Extra Trees |
GB | Gradient Boosting |
DT | Decision Trees |
LR | Logistic Regression |
SGD | Stochastic Gradient Descent |
KNN | K-Nearest Neighbors |
NB | Naive Bayes |
ML | Machine Learning |
DL | Deep Learning |
ICT | Information and Communication Technology |
LMS | Learning Management Systems |
SIS | Student Information Systems |
EDM | Educational Data Mining |
GADF | Gradient Angle Difference Field |
LSTM | Long Short-Term Memory |
GRU | Gated Recurrent Units |
BLS | Broad Learning System |
WS | Wavelet Scalograms |
GASF | Gramian Angular Summation Field |
NLTK | Natural Language Toolkit |
TP | True Positive |
TN | True Negative |
FP | False Positive |
FN | False Negative |
References
- Hasan, R.; Palaniappan, S.; Mahmood, S.; Shah, B.; Abbas, A.; Sarker, K.U. Enhancing the teaching and learning process using video streaming servers and forecasting techniques. Sustainability 2019, 11, 2049. [Google Scholar] [CrossRef]
- Romero, C.; Ventura, S. Educational data mining and learning analytics: An updated survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2020, 10, e1355. [Google Scholar] [CrossRef]
- Du, X.; Yang, J.; Hung, J.-L.; Shelton, B. Educational data mining: A systematic review of research and emerging trends. Inf. Discov. Deliv. 2020, 48, 225–236. [Google Scholar] [CrossRef]
- Albreiki, B.; Zaki, N.; Alashwal, H. A systematic literature review of student’performance prediction using machine learning techniques. Educ. Sci. 2021, 11, 552. [Google Scholar] [CrossRef]
- Matz, S.C.; Bukow, C.S.; Peters, H.; Deacons, C.; Dinu, A.; Stachl, C. Using machine learning to predict student retention from socio-demographic characteristics and app-based engagement metrics. Sci. Rep. 2023, 13, 5705. [Google Scholar]
- Umer, R.; Susnjak, T.; Mathrani, A.; Suriadi, L. Current stance on predictive analytics in higher education: Opportunities, challenges and future directions. Interact. Learn. Environ. 2023, 31, 3503–3528. [Google Scholar] [CrossRef]
- Yang, Z.; Yang, J.; Rice, K.; Hung, J.-L.; Du, X. Using convolutional neural network to recognize learning images for early warning of at-risk students. IEEE Trans. Learn. Technol. 2020, 13, 617–630. [Google Scholar] [CrossRef]
- Kong, X.; Luo, C. A novel ConvLSTM with multifeature fusion for financial intelligent trading. Int. J. Intell. Syst. 2022, 37, 8855–8877. [Google Scholar] [CrossRef]
- Asif, R.; Merceron, A.; Ali, S.A.; Haider, N.G. Analyzing undergraduate students’ performance using educational data mining. Comput. Educ. 2017, 113, 177–194. [Google Scholar] [CrossRef]
- Yağcı, M. Educational data mining: Prediction of students’ academic performance using machine learning algorithms. Smart Learn. Environ. 2022, 9, 11. [Google Scholar] [CrossRef]
- Bujang, S.D.A.; Selamat, A.; Ibrahim, R.; Krejcar, O.; Herrera-Viedma, E.; Fujita, H.; Ghani, N.A.M. Multiclass prediction model for student grade prediction using machine learning. IEEE Access 2021, 9, 95608–95621. [Google Scholar] [CrossRef]
- Nayak, P.; Vaheed, S.; Gupta, S.; Mohan, N. Predicting students’ academic performance by mining the educational data through machine learning-based classification model. Educ. Inf. Technol. 2023, 28, 14611–14637. [Google Scholar] [CrossRef]
- Ram, M.S.; Srija, V.; Bhargav, V.; Madhavi, A.; Kumar, G.S. Machine Learning Based Student Academic Performance Prediction. In Proceedings of the 2021 Third International Conference on Inventive Research in Computing Applications (ICIRCA), Coimbatore, India, 2–4 September 2021; pp. 683–688. [Google Scholar]
- Balabied, S.A.A.; Eid, H.F. Utilizing random forest algorithm for early detection of academic underperformance in open learning environments. PeerJ Comput. Sci. 2023, 9, e1708. [Google Scholar] [CrossRef] [PubMed]
- Aljohani, N.R.; Fayoumi, A.; Hassan, S.-U. Predicting at-risk students using clickstream data in the virtual learning environment. Sustainability 2019, 11, 7238. [Google Scholar] [CrossRef]
- Waheed, H.; Hassan, S.-U.; Aljohani, N.R.; Hardman, J.; Alelyani, S.; Nawaz, R. Predicting academic performance of students from VLE big data using deep learning models. Comput. Hum. Behav. 2020, 104, 106189. [Google Scholar] [CrossRef]
- Huang, Q.; Zeng, Y. Improving academic performance predictions with dual graph neural networks. Complex Intell. Syst. 2024, 10, 3557–3575. [Google Scholar] [CrossRef]
- Ben Said, A.; Abdel-Salam, A.-S.G.; Hazaa, K.A. Performance prediction in online academic course: A deep learning approach with time series imaging. Multimed. Tools Appl. 2024, 83, 55427–55445. [Google Scholar] [CrossRef]
- Yang, C.-L.; Chen, Z.-X.; Yang, C.-Y. Sensor classification using convolutional neural network by encoding multivariate time series as two-dimensional colored images. Sensors 2019, 20, 168. [Google Scholar] [CrossRef]
- Li, J.; Wang, Q. Comparison of the representational ability in individual difference analysis using 2-D time-series image and time-series feature patterns. Expert Syst. Appl. 2023, 215, 119429. [Google Scholar] [CrossRef]
- Yin, J.; Zhuang, X.; Sui, W.; Sheng, Y.; Yang, Y. A new similarity measurement method for time series based on image fusion of recurrence plots and wavelet scalogram. Eng. Appl. Artif. Intell. 2024, 129, 107679. [Google Scholar] [CrossRef]
- Jin, X.-B.; Yang, A.; Su, T.; Kong, J.-L.; Bai, Y. Multi-channel fusion classification method based on time-series data. Sensors 2021, 21, 4391. [Google Scholar] [CrossRef] [PubMed]
- Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
- Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
- Schapire, R.E. The strength of weak learnability. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef]
- Chatzimparmpas, A.; Martins, R.M.; Kucher, K.; Kerren, A. StackGenVis: Alignment of data, algorithms, and models for stacking ensemble learning using performance metrics. IEEE Trans. Vis. Comput. Graph. 2020, 27, 1547–1557. [Google Scholar] [CrossRef]
- Seijo-Pardo, B.; Porto-Díaz, I.; Bolón-Canedo, V.; Alonso-Betanzos, A. Ensemble feature selection: Homogeneous and heterogeneous approaches. Knowl. Based Syst. 2017, 118, 124–139. [Google Scholar] [CrossRef]
- Xu, Q.; Wang, J.; Jiang, B.; Luo, B. Fine-grained visual classification via internal ensemble learning transformer. IEEE Trans. Multimed. 2023, 25, 9015–9028. [Google Scholar] [CrossRef]
- Saidani, O.; Umer, M.; Alshardan, A.; Alturki, N.; Nappi, M.; Ashraf, I. Student academic success prediction in multimedia-supported virtual learning system using ensemble learning approach. Multimed. Tools Appl. 2024, 83, 87553–87578. [Google Scholar] [CrossRef]
- Teoh, C.-W.; Ho, S.-B.; Dollmat, K.S.; Tan, C.-H. Ensemble-Learning techniques for predicting student performance on video-based learning. Int. J. Inf. Educ. Technol. 2022, 12, 741–745. [Google Scholar] [CrossRef]
- Shayegan, M.J.; Akhtari, R. A Stacking Machine Learning Model for Student Performance Prediction Based on Class Activities in E-Learning. Comput. Syst. Sci. Eng. 2024, 48, 1251–1272. [Google Scholar] [CrossRef]
- Al-Ameri, A.; Al-Shammari, W.; Castiglione, A.; Nappi, M.; Pero, C.; Umer, M. Student Academic Success Prediction Using Learning Management Multimedia Data With Convoluted Features and Ensemble Model. ACM J. Data Inf. Qual. 2024. [Google Scholar] [CrossRef]
- Evangelista, E. A hybrid machine learning framework for predicting students’ performance in virtual learning environment. Int. J. Emerg. Technol. Learn. (iJET) 2021, 16, 255–272. [Google Scholar] [CrossRef]
- Mary, T.A.C.; Rose, P.A.L. Ensemble Machine Learning Model for University Students’ Risk Prediction and Assessment of Cognitive Learning Outcomes. 2023 IJIET 2023, 13, 948–958. [Google Scholar]
- Hasan, R.; Palaniappan, S.; Mahmood, S.; Abbas, A.; Sarker, K.U. Dataset of Students’ Performance Using Student Information System, Moodle and the Mobile Application “eDify”. Data 2021, 6, 110. [Google Scholar] [CrossRef]
- Amankwah, M.G.; Camps, D.; Bethel, E.W.; Van Beeumen, R.; Perciano, T. Quantum pixel representations and compression for N-dimensional images. Sci. Rep. 2022, 12, 7712. [Google Scholar] [CrossRef]
- Naidu, V.; Divya, M.; Mahalakshmi, P. Multi-modal medical image fusion using multi-resolution discrete sine transform. Control Data Fusion E-J. 2017, 1, 13–26. [Google Scholar]
- Zhang, Y.; Hou, Y.; OuYang, K.; Zhou, S. Multi-scale signed recurrence plot based time series classification using inception architectural networks. Pattern Recognit. 2022, 123, 108385. [Google Scholar] [CrossRef]
- Jaleel, M.; Kucukler, O.F.; Alsalemi, A.; Amira, A.; Malekmohamadi, H.; Diao, K. Analyzing gas data using deep learning and 2-d gramian angular fields. IEEE Sens. J. 2023, 23, 6109–6116. [Google Scholar] [CrossRef]
- Sun, Z.; Wang, G.; Li, P.; Wang, H.; Zhang, M.; Liang, X. An improved random forest based on the classification accuracy and correlation measurement of decision trees. Expert Syst. Appl. 2024, 237, 121549. [Google Scholar] [CrossRef]
- Cervantes, J.; Garcia-Lamont, F.; Rodríguez-Mazahua, L.; Lopez, A. A comprehensive survey on support vector machine classification: Applications, challenges and trends. Neurocomputing 2020, 408, 189–215. [Google Scholar] [CrossRef]
- Saeed, U.; Jan, S.U.; Lee, Y.-D.; Koo, I. Fault diagnosis based on extremely randomized trees in wireless sensor networks. Reliab. Eng. Syst. Saf. 2021, 205, 107284. [Google Scholar] [CrossRef]
- Deng, S.; Su, J.; Zhu, Y.; Yu, Y.; Xiao, C. Forecasting carbon price trends based on an interpretable light gradient boosting machine and Bayesian optimization. Expert Syst. Appl. 2024, 242, 122502. [Google Scholar] [CrossRef]
- Wang, W.; Sun, D. The improved AdaBoost algorithms for imbalanced data classification. Inf. Sci. 2021, 563, 358–374. [Google Scholar] [CrossRef]
- Charbuty, B.; Abdulazeez, A. Classification based on decision tree algorithm for machine learning. J. Appl. Sci. Technol. Trends 2021, 2, 20–28. [Google Scholar] [CrossRef]
- Bailly, A.; Blanc, C.; Francis, É.; Guillotin, T.; Jamal, F.; Wakim, B.; Roy, P. Effects of dataset size and interactions on the prediction performance of logistic regression and deep learning models. Comput. Methods Programs Biomed. 2022, 213, 106504. [Google Scholar] [CrossRef]
- Qin, W.; Luo, X.; Zhou, M. Adaptively-accelerated Parallel Stochastic Gradient Descent for High-Dimensional and Incomplete Data Representation Learning. IEEE Trans. Big Data 2023, 10, 92–107. [Google Scholar] [CrossRef]
- Wang, A.X.; Chukova, S.S.; Nguyen, B.P. Ensemble k-nearest neighbors based on centroid displacement. Inf. Sci. 2023, 629, 313–323. [Google Scholar] [CrossRef]
- Wickramasinghe, I.; Kalutarage, H. Naive Bayes: Applications, variations and vulnerabilities: A review of literature with code snippets for implementation. Soft Comput. 2021, 25, 2277–2293. [Google Scholar] [CrossRef]
- Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A survey of convolutional neural networks: Analysis, applications, and prospects. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 6999–7019. [Google Scholar] [CrossRef]
- Petersen, P.; Voigtlaender, F. Equivalence of approximation by convolutional neural networks and fully-connected networks. Proc. Am. Math. Soc. 2020, 148, 1567–1581. [Google Scholar] [CrossRef]
- Ashwinkumar, S.; Rajagopal, S.; Manimaran, V.; Jegajothi, B. Automated plant leaf disease detection and classification using optimal MobileNet based convolutional neural networks. Mater. Today Proc. 2022, 51, 480–487. [Google Scholar] [CrossRef]
- Geetha, A.; Prakash, N. Classification of Glaucoma in Retinal Images Using EfficientnetB4 Deep Learning Model. Comput. Syst. Sci. Eng. 2022, 43, 1041–1055. [Google Scholar] [CrossRef]
- Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef] [PubMed]
Features | Description |
---|---|
Applicant Name | It represents the student’s name such as “Student 1”. |
Attempt Count | It indicates the number of attempts made for a specific module. It is categorized as “Low/Medium/High”. |
Prohibition | The student might have incomplete modules to address. It is categorized as “Yes/No”. |
CGPA | It shows the student’s overall grade point average. It is categorized as “Adequate/Excellent/Fair/Good/Poor/Very Good”. |
Remote Student | It represents whether the student is enrolled in an e-learning study mode. It is categorized as “Yes/No”. |
High Risk | It indicates the likelihood of failure in a particular subject. It is categorized as “Yes/No”. |
At Risk | It shows that a student who fails previous modules will be considered at risk. It is categorized as “Yes/No”. |
Term Exceeded | It represents the student’s advancement in their degree program. It is categorized as “Yes/No”. |
At Risk SSC | It indicates whether the student has been enrolled in the student success center due to academic deficiencies. It is categorized as “Yes/No”. |
Plagiarism history | It details the student’s previous instances of plagiarism in any module. It is categorized as “Low/Medium”. |
CW1 | The grades achieved by the student in their first coursework. It is categorized as “Fair/Fail/Adequate/Very Good/Good/Excellent”. |
Other Modules | It indicates the student’s enrollment in other modules during the current semester. It is categorized as “Low/Medium/High”. |
CW2 | The student’s score achieved in their second coursework. It is categorized as “Fair/Fail/Adequate/Very Good/Good/Excellent”. |
Online C | Time spent by the user on on-campus activities (measured in minutes). It is categorized as “Adequate/Excellent/Fair/Good/Poor/Very Good”. |
ESE | It refers to the marks obtained in the end-of-semester examination. It is categorized as “Fair/Fail/Adequate/Very Good/Good/Excellent”. |
Online O | Time spent by the user on off-campus activities (measured in minutes). It is categorized as “Adequate/Excellent/Fair/Good/Poor/Very Good”. |
Paused | The total count of times the video was stopped. |
Played | The total count of times the video was played. |
Likes | The total number of times the student has marked the video as liked. |
Segment | The total instances where a student has used the slider to play a specific section of the video. |
Result | It indicates whether the student has passed the exam. It is categorized as “Pass/Fail”. |
Class | Total |
---|---|
Pass | 264 |
Fail | 62 |
Total | 326 |
Frameworks | Pytorch, Sci-Kit Learn |
---|---|
Language | Python3.8 |
Ram | 256 G |
OS | Ubuntu 20.04 LTS |
CPU | Intel Xeon Platinum 8362 @ 2.80 GHz |
GPU | NVIDIA A40 48 G |
Model | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
RF | 0.9242 | 0.9308 | 0.9242 | 0.9169 |
SVM | 0.9091 | 0.9058 | 0.9091 | 0.9061 |
ET | 0.9091 | 0.9077 | 0.9091 | 0.9025 |
GB | 0.9091 | 0.9183 | 0.9091 | 0.8979 |
AdaBoost | 0.8485 | 0.8408 | 0.8485 | 0.8436 |
DT | 0.8333 | 0.8208 | 0.8333 | 0.8248 |
LR | 0.7727 | 0.8378 | 0.7727 | 0.7914 |
SGD | 0.6970 | 0.8142 | 0.6970 | 0.7269 |
KNN | 0.6061 | 0.7488 | 0.6061 | 0.6461 |
NB | 0.5455 | 0.7501 | 0.5455 | 0.5901 |
Model | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
1D CNN | 0.8491 | 0.8491 | 0.8491 | 0.8491 |
MobileNet | 0.8208 | 0.8255 | 0.8208 | 0.8213 |
EfficientNetB4 | 0.8302 | 0.8335 | 0.8302 | 0.8307 |
Model | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
ML Models | ||||
RF | 0.9242 | 0.9308 | 0.9242 | 0.9169 |
SVM | 0.9091 | 0.9058 | 0.9091 | 0.9061 |
ET | 0.9091 | 0.9077 | 0.9091 | 0.9025 |
GB | 0.9091 | 0.9183 | 0.9091 | 0.8979 |
AdaBoost | 0.8485 | 0.8408 | 0.8485 | 0.8436 |
DT | 0.8333 | 0.8208 | 0.8333 | 0.8248 |
LR | 0.7727 | 0.8378 | 0.7727 | 0.7914 |
SGD | 0.6970 | 0.8142 | 0.6970 | 0.7269 |
KNN | 0.6061 | 0.7488 | 0.6061 | 0.6461 |
NB | 0.5455 | 0.7501 | 0.5455 | 0.5901 |
DL Models | ||||
1D CNN | 0.8491 | 0.8491 | 0.8491 | 0.8491 |
MobileNet | 0.8208 | 0.8255 | 0.8208 | 0.8213 |
EfficientNetB4 | 0.8302 | 0.8335 | 0.8302 | 0.8307 |
Model | Feature Transformation | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|---|
2D CNN | PR | 0.8208 | 0.8271 | 0.8208 | 0.8176 |
2D FCN | PR | 0.8396 | 0.8410 | 0.8396 | 0.8384 |
2D CNN | GAF | 0.8679 | 0.8689 | 0.8679 | 0.8672 |
2D FCN | GAF | 0.7925 | 0.8034 | 0.7925 | 0.7868 |
2D CNN | SWT | 0.8868 | 0.8869 | 0.8868 | 0.8865 |
2D FCN | SWT | 0.8679 | 0.8679 | 0.8679 | 0.8676 |
2D CNN | RP | 0.9340 | 0.9340 | 0.9340 | 0.9339 |
2D FCN | RP | 0.8774 | 0.8820 | 0.8774 | 0.8777 |
Model | Feature Transformation | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|---|
2D CNN | PR | 0.8208 | 0.8271 | 0.8208 | 0.8176 |
2D FCN | PR | 0.8396 | 0.8410 | 0.8396 | 0.8384 |
2D EnCF | PR | 0.8396 | 0.8517 | 0.8396 | 0.8358 |
2D CNN | GAF | 0.8679 | 0.8689 | 0.8679 | 0.8672 |
2D FCN | GAF | 0.7925 | 0.8034 | 0.7925 | 0.7868 |
2D EnCF | GAF | 0.8585 | 0.8606 | 0.8585 | 0.8588 |
2D CNN | SWT | 0.8868 | 0.8869 | 0.8868 | 0.8865 |
2D FCN | SWT | 0.8679 | 0.8679 | 0.8679 | 0.8676 |
2D EnCF | SWT | 0.9057 | 0.9072 | 0.9057 | 0.9051 |
2D CNN | RP | 0.9340 | 0.9340 | 0.9340 | 0.9339 |
2D FCN | RP | 0.8774 | 0.8820 | 0.8774 | 0.8777 |
2D EnCF | RP | 0.9528 | 0.9529 | 0.9528 | 0.9528 |
Model | Feature Transformation | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|---|
ML Models | |||||
RF | / | 0.9242 | 0.9308 | 0.9242 | 0.9169 |
SVM | / | 0.9091 | 0.9058 | 0.9091 | 0.9061 |
ET | / | 0.9091 | 0.9077 | 0.9091 | 0.9025 |
GB | / | 0.9091 | 0.9183 | 0.9091 | 0.8979 |
AdaBoost | / | 0.8485 | 0.8408 | 0.8485 | 0.8436 |
DT | / | 0.8333 | 0.8208 | 0.8333 | 0.8248 |
LR | / | 0.7727 | 0.8378 | 0.7727 | 0.7914 |
SGD | / | 0.6970 | 0.8142 | 0.6970 | 0.7269 |
KNN | / | 0.6061 | 0.7488 | 0.6061 | 0.6461 |
NB | / | 0.5455 | 0.7501 | 0.5455 | 0.5901 |
DL Models | |||||
1D CNN | / | 0.8491 | 0.8491 | 0.8491 | 0.8491 |
MobileNet | / | 0.8208 | 0.8255 | 0.8208 | 0.8213 |
EfficientNetB4 | / | 0.8302 | 0.8335 | 0.8302 | 0.8307 |
2D CNN | PR | 0.8208 | 0.8271 | 0.8208 | 0.8176 |
2D FCN | PR | 0.8396 | 0.8410 | 0.8396 | 0.8384 |
2D EnCF | PR | 0.8396 | 0.8517 | 0.8396 | 0.8358 |
2D CNN | GAF | 0.8679 | 0.8689 | 0.8679 | 0.8672 |
2D FCN | GAF | 0.7925 | 0.8034 | 0.7925 | 0.7868 |
2D EnCF | GAF | 0.8585 | 0.8606 | 0.8585 | 0.8588 |
2D CNN | SWT | 0.8868 | 0.8869 | 0.8868 | 0.8865 |
2D FCN | SWT | 0.8679 | 0.8679 | 0.8679 | 0.8676 |
2D EnCF | SWT | 0.9057 | 0.9072 | 0.9057 | 0.9051 |
2D CNN | RP | 0.9340 | 0.9340 | 0.9340 | 0.9339 |
2D FCN | RP | 0.8774 | 0.8820 | 0.8774 | 0.8777 |
2D EnCF | RP | 0.9528 | 0.9529 | 0.9528 | 0.9528 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhao, S.; Zhou, D.; Wang, H.; Chen, D.; Yu, L. Enhancing Student Academic Success Prediction Through Ensemble Learning and Image-Based Behavioral Data Transformation. Appl. Sci. 2025, 15, 1231. https://doi.org/10.3390/app15031231
Zhao S, Zhou D, Wang H, Chen D, Yu L. Enhancing Student Academic Success Prediction Through Ensemble Learning and Image-Based Behavioral Data Transformation. Applied Sciences. 2025; 15(3):1231. https://doi.org/10.3390/app15031231
Chicago/Turabian StyleZhao, Shuai, Dongbo Zhou, Huan Wang, Di Chen, and Lin Yu. 2025. "Enhancing Student Academic Success Prediction Through Ensemble Learning and Image-Based Behavioral Data Transformation" Applied Sciences 15, no. 3: 1231. https://doi.org/10.3390/app15031231
APA StyleZhao, S., Zhou, D., Wang, H., Chen, D., & Yu, L. (2025). Enhancing Student Academic Success Prediction Through Ensemble Learning and Image-Based Behavioral Data Transformation. Applied Sciences, 15(3), 1231. https://doi.org/10.3390/app15031231