You are viewing a javascript disabled version of the site. Please enable Javascript for this site to function properly.
Go to headerGo to navigationGo to searchGo to contentsGo to footer
In content section. Select this link to jump to navigation

Metaheruistic optimization based machine-learning approach for analysis of academic students’ strategies

Abstract

In this technology world, education is also becoming one of the basic necessities of human life like food, shelter, and clothes. Even in day-to-day daily activities, the world is moving toward an automated process using technology developments. Some of the technology developments in day-to-day life activities are smartphone, internet activities, and home and office appliances. To cope with these advanced technologies, the persons must have basic educational qualification to understand and operate those appliances easily. Apart from this, the education helps the person to develop their personal growth in both knowledge and wealth. With the development of technologies, different Artificial Intelligence techniques have been applied on the datasets to analyze these factors and enhance the teaching method. But the current techniques were applied to one or two data models that analyze either their educational performance or demographic variable. But these models were not sufficient for analyzing all the factors that affects the education. To overcome this, a single optimized machine-learning approach is proposed in this paper to analyze the factors that affect the education. This analysis helps the faculty to enhance their teaching methodology and understand the student’s mentality toward education. The proposed Hybrid Cuckoo search-particle swarm optimization was implemented on three datasets to determine the factors that affect the education. These optimal factors are determined by identifying their relations to the final results of an individual person. All these optimal factors are combined and grades are grouped to analyze the proposed optimization process performance using regression neural network. The proposed optimization-based neural network was tested on three data models and its performance analysis showed that the proposed model can achieve higher accuracy of 99% that affects the individual education. This shows that the proposed model can help the faculty to enhance their attention to the students individually.

1Introduction

The world is utilizing technological advancements to move toward an automated procedure. People need to acquire a foundational education in order to deal with these new technologies. A person’s personal advancement in wealth and knowledge is facilitated by education. By introducing new ideas and effectively utilizing the resources at hand, it also aids in the development of nations. The current models used for estimating the academic performance were discussed below.

In [1] and [2], the decision tree models were used for estimating academic performance. Along with decision tree, the Naïve Bayes algorithm, and neural network are also used for estimating the student performance [3]. The deep learning algorithm is also used for estimating the dropout possibility of the students along with the traditional machine-learning algorithms [4]. In [5], the deep learning algorithms were implemented for estimating the student performance. The importance of preprocessing step in estimating the student performance was analyzed in [6]. In [7, 8] also, the preprocessing importance was analyzed using the correlation map before classification. In [9], the machine-learning algorithms like decision tree and multilayer perceptron network model were analyzed.

In [10] and [11], also the ML model performance for estimating the student performance was evaluated. A complete review of machine-learning models for estimating the student performance was presented in [12]. In [13], the stochastic gradient descent optimization algorithm was used to tune the network performance of multilayer perceptron in student performance prediction. In [14] also proposed a complete automate approach of student performance prediction using machine-learning technique. In [15] also analyzed the different types of ML models in estimating the student performance.

1.1Related works

A increasing body of research has focused on incorporating machine learning methods into meta-heuristics to address combinatorial optimisation issues. This integration intends to enhance the performance of the me-ta heuristics in terms of the quality of the solutions, the convergence rate, and the robustness of the search [21].

According to [20], forecasting student retention offers information on possibilities for deliberate student counselling. The review offers a research viewpoint on predicting student retention using machine learning through a number of significant discoveries, including the identification of the variables employed in prior studies and the prediction approaches Educational data mining is a group of data mining applications in the field of education, according to [22]. The analysis of student and teacher data is a focus of these programmes. The analysis could be put to use for classification or forecasting. Examined are machine learning techniques including Nave Bayes, ID3, C4.5, and SVM. [23] suggests a new machine learning-based model to forecast undergraduate students’ final test marks using the results of their midterm exam grades as the source data. An educational process and data mining plus machine learning (EPDM + ML) model was introduced in [24] and used to contextually analyse the performances of the teachers and provide recommendations based on information obtained from students’ evaluations of instruction (SET). Higher education institutions have begun integrating technology into their conventional teaching methods in [25] in an effort to improve learning and teaching. In this study, five machine learning algorithms were used to analyse two datasets in order to predict and categorise student performance.

1.2Shortcomings of the current techniques

Because the traditional machine-learning algorithms like SVM, ANN, and K-means techniques were applied to the educational dataset to observe the pattern in the set, and it helps to estimate the possible future outcomes. Even though the existing techniques can estimate the student performance but its performance was lagging in the following process:

The attribute analysis that affects the student performance was performed through the unsupervised learning or correlation analysis. Some of the models utilized thresholds for predicting the performance.

1.3Motivation of study

Demographic elements including a parent’s educational level, environment, and one’s own desire in education are among the many aspects that influence an individual’s education. On the datasets, a variety of artificial intelligence algorithms have been used to analyze these variables and improve the teaching strategy. These models were not enough for analysing every element that has an impact on education. In this paper, a single, optimized machine-learning strategy is suggested.

2Materials and methods

The survey analysis showed that the machine-learning algorithm is not only efficient in image classification or pattern recognition but it also plays a vital role in data mining applications like academic performance analysis to identify the relation between the inputs and outputs. Because the traditional machine-learning algorithms like SVM, ANN, and K-means techniques were applied to the educational dataset to observe the pattern in the set, and it also helps to estimate the possible future outcomes.

As the aforementioned both processes are depending on the user criteria, its performance will not be stable in predicting the results. Hence, in this, an optimized machine-learning approach is proposed for finding the factors that affect the student’s performance. This helps the faculty to counsel their students and improve their performance. The proposed method was tested on the three types of datasets as follows:

  • Student exam performance dataset

  • Student academic performance dataset

  • Portuguese student academic performance dataset.

The above three datasets have different input attributes and a common output attribute called student performance with different notations. These input attributes were processed by the proposed hybrid cuckoo-search and particle swarm optimization to find the optimal attribute that affects the student performance using regression neural network (RNN). Then, these factors are combined and tested using the RNN model for final evaluation. The working of the proposed model is explained briefly below.

3Proposed work

In this, the hybrid cuckoo search and particle swarm optimization were used to find the optimal factors that affect the student performance and was determined by solving the objective function. Here, the objective function is to minimize the regression neural network (RNN) classifier error rate. This process was performed on three types of datasets. These datasets were downloaded from the Kaggle website. Before performing the aforementioned process, the datasets undergone the following process:

  • Data collection

  • Preprocessing

  • Assigning values to categorical variables

  • Normalization

  • Data Exploration

Then, the preprocessed data will be subjected to the optimal attribute selection using hybrid CSO-PSO approach. The optimization process finds the factors that affect the student performance using the following objective function.

(1)
HCSO-PSOobjfunction=min(RNNErrorRate)

From this, the optimal factors from each model are collected, and the final performance was grouped as low, medium, and high. Then, the transformed dataset was trained using regression neural network, and its performance was measured in terms of accuracy, precision, recall, and F1-score.

3.2Dataset

In this paper, the three types of student performance data were collected from the kaggle websites [16–18]. The three datasets describe the following information:

  • In the first dataset [16], the student performance was analyzed in terms of student and parent behavior.

  • In the second dataset [17], the student exam performance was analyzed in terms of students reading pattern and parent educational qualification.

  • In the third dataset [18], the student performance was analyzed in terms of student and parent nature.

The data distribution in this three datasets is presented in Table 1.

Table 1

Data distribution in three datasets

VariablesDataset 1Dataset 2Dataset 4
Total number of records4801001650
Input attributes15732
Output attributesEducational stage:Test scoreScores G3
Low0 to 1000 to 20
Medium
High
Features analyzedDemographicDemographic InformationDemographic
Academic detailsStudent academic informationAcademic information
Student behaviorParentParent information
Parent responseStudent behavior
DemographicGenderGenderGender
NationalityPlaceAge
Place of birthLunchSchool
Address
Reason for school selection
Academic InformationStudent informationTest preparationNursery
GradeRead scoreHigher
SubjectWrite scoreGrade
Section idStudy time
SemesterAbsence
AbsenceHealth status
Subject materials:Educational support from
Revisedschool and parents
Announcements viewed
Discussion groups participated
Student behaviorRaised hand-Internet
Relationship
Alcohol consumption during workout and weekend
ParentParent response to the groupsParent educationParent education status
Parent satisfactionParent occupation
Parent cohabited
Family time

The above information in the dataset were either in categorical or numerical formats. But for an efficient data mining process, the data should be in a single format. In particular, the numeric data format helps to analyze the information in faster and efficiently. Hence, in this, the datasets were subjected to a preprocessing step

3.3Preprocessing

For an efficient and faster analysis, the datasets were subjected to a data transformation process.

In data transformation process, the binary variables like true or false, yes or no, or any two class variables like gender, college, or address nature are transformed to 1 and 0

(2)
V[binary]transformationV[0,1].

In categorical variables like educational stage, student nationality, and other categories are transformed into 1 to n categories based on number of categorical options.

(3)
V[categorical1ton]transformationV[1ton]

Once the datasets are preprocessed using the above Equations 2 and 3, the remaining data also combined together in each set to form a complete numerical dataset. Then, the transformed data was subjected to Z-score normalization to normalize the values between 0 and 1 in all categories. The Z-score normalization calculates the new values for all the variables using Equation 4.

(4)
Vnewz-scoreV-mean(V)std(V)

Then, the normalized values are subjected to the data exploration process for the first stage of data reduction.

3.4Data exploration

In this stage, the data distribution individually was analyzed which least impact on the target variable is called output variables in each set. The sample data exploration of the set in terms of gender and school is shown in Figs. 1 and 2.

Fig. 1

Dataset 1 exploration.

Dataset 1 exploration.
Fig. 2

Dataset 2 exploration.

Dataset 2 exploration.

In the above Fig. 1, the demographic variables like gender, subjects, guardian of the student were presented in figure format for dataset 1 with gender as base variable.

In Fig. 2, the demographic variables like gender, nationality, and lunch activities of the student were presented in figure format for dataset with gender as the base variable.

In Fig. 3, the demographic variables like gender, locality, guardian of the student were presented in figure format for a dataset with gender as base variable.

Fig. 3

Dataset 3 exploration.

Dataset 3 exploration.

In the data exploration part, the individual variable analysis was only carried out which gives an idea of the data distribution in the set. Based on this, the general variables like gender, nationality, and place of birth are removed from the sets. Then, the remaining variables were subjected to attribute selection using the proposed hybrid CSO-PSO.

3.5HCSO-PSO-based attribute selection

In this, the preprocessed variable from the set was subjected to optimization process to select the optimal attributes that are responsible for the student performance was determined by the proposed HCSO-PSO process. Both processes utilize the same objective function mentioned in Equation 2 was used. The boundary variables were only modified in the first stage of process using cuckoo search optimization. While in the second stage, the boundary variables are common to find the four optimal variables. The proposed method process is represented in the flow chart in Fig. 4.

Fig. 4

HCSO-PSO processes.

HCSO-PSO processes.

3.5.1First stage attribute selection using CSO

In this stage, the optimal six attributes from each set are determined using Table 2 Cuckoo search initialization parameters. Using Equation 1 and the steps mentioned in the [19] was used to find the six best attributes for evaluate the student performance.

Table 2

Initialization values for CSO

ParameterValue
Total no of nests20
Maximum iterations50
Attraction of nest0.25
Lower bound1 for all sets
Upper bound12 for set 1
6 for set 2
29 for set 3
No of output6

Apart from dataset 2, in datasets 1 and 3, the optimal attributes were selected from the set. In set 2, only six attributes are present; hence, all the six attributes were selected for the second stage process. In set 1, the student behavior and education were selected. In set 3, the student behavior like absences, study time, romantic, internet, and education g1, g2 was selected for set 3. These reduced sets were subjected to the second stage of processing using particle swarm optimization.

3.5.2Second stage attributes selection using PSO

In this stage, the optimal four attributes that affect the student performance were determined using Table 3 Particle swarm initialization parameters. Using Equation 1 and the swarm position update equations were used to find the four optimal attributes in all the sets.

Table 3

Initialization values for PSO

ParameterValue
Number of swarms50
Maximum iteration50
Inertia weight0.8
C11.5
C22.0
Lower bound1
Upper bound6
Optimal attributes4

The PSO begins its process by minimizing Equation 1 objective function. After the first iteration, the swarm update its position using Equation 5.

(5)
SwarmnewPosition=SwarmcurrentPosition+Swarmnewvelocity

Before position update, the swarm update its velocity using Equation 6 to find the optimal attributes that help to enhance the academic performance of students

(6)
velocitynew=Rand(0,1)*(LBS-SwarmcurrentPosition+(GBS-SwarmcurrentPosition
where LBS is the local best solution and GBS is the global best solution. With these equations, the swarm determines the optimal four attributes from each set. In set 1, the student performance like discussion, view, raising hand, and response was selected as optimal attributes. In set 2, apart from lunch and place of birth, all other attributes are selected. In set 3, the grade performance, study time and romantic were selected.

3.6Regression neural network

The reduce attributes from the hybrid CSO-PSO algorithm, a new dataset was framed with the reduced attribute set for analysis. Then, the new set is divided into training and testing using hold-out approach with a 0.3% value. This helps to split the dataset into 70% as training and 30% as testing.

The optimal four attributes will act as the first layer or input to the regression neural networks. In the corresponding test score, categorical variables were given as the outputs to the network.

The kernel function was multiplied with the inputs to determine the neuron weights using Equations 7 and 8.

(7)
kernel(x,xn)=e-kd/2σ2

(8)
kd=(x-xn)(x-xn)

In this layer, the kernel function in the step 2 is the denominator and the corresponding output is the numerator.

The inputs and outputs are related by dividing the numerator and denominator parts.

(9)
Y=kN=1yk*Kernel(x,xn)kN=1Kernel(x,xn)

Then, the transformed dataset was trained using a regression neural network, and its performance was measured in terms of accuracy, precision, recall, and F1-score.

4Evaluation metrics

In this, the proposed HCSO-PSO performance in attribute selection was evaluated using regression neural network training and testing model evaluation metrics. Because the RNN was modeled using the selected attributes to estimate the student performance. In this, the following metrics were used for evaluation, and it plays a major role in the evaluation process.

4.2Accuracy

Equation 10 is used for calculating the accuracy. This value evaluates the model in terms of the correct estimation of student performance.

(10)
Accuracy=TP+TN+FP+FNOverallRecords

The accuracy indicates the model types like underfitting, overfitting, and correct model. This accuracy will be calculated for both the training and testing to determine the model types.

4.2.1Under fitting model

When a model has lower accuracies in both the sets, then that model is an underfitting model as shown in Fig. 6.

4.2.2Overfitting model

When a model has higher accuracy in training and lower accuracy in testing means, the model is highly correlated to the training set, and it will act as an overfitting model as shown in Fig. 6.

4.2.3Accurate model

When the model accuracy is the same and higher for both training and testing means, it is a correct model for future estimation.

From Fig. 5, the following points were observed.

Fig. 5

Model types based on error.

Model types based on error.
Table 4

Model types based on error

ModelError nature
UnderfittingTraining and testing error is high
OverfittingTesting error >  training error
Correct modelTraining and testing error is optimal and somewhat equal

4.3Precision

Equation 11 is used for calculating the model precision. This value is used to identify the model performance in estimating the values accurately.

(11)
Precision=TPTP+FP

4.4Recall

Equation 12 is used for calculating the model precision. This value is used to identify the model performance in estimating the total number of times the model can find the values accurately.

(12)
Recall=TPTP+FN

4.5F1-score

F1 score is calculated from precision and recall and is given in Equation 13.

(13)
F1-score=2*Precision.RecallPrecision+Recall

Using Equations 10 to 13, the proposed HCSO-PSO-based regression neural network model is used in estimating the student performance was evaluated and the corresponding results are given below.

5Results and discussion

In this, the proposed method HCSO-PSO was analyzed on the three datasets to find the optimal attributes for estimating the student academic performance. Then, the optimal attributes were trained using the regression neural network and tested. Then, the network performance was evaluated using accuracy, precision, recall and F1-score.

5.1Dataset

In this, the sample screenshots of the three datasets are shown in Figs. 68.

Fig. 6

Dataset 1 sample.

Dataset 1 sample.
Fig. 7

Dataset 2 sample.

Dataset 2 sample.
Fig. 8

Dataset 3 sample.

Dataset 3 sample.

5.2Preprocessing and exploration

In the preprocessing step, the data were transformed and normalized the values 0 and 1 as shown in the sample Fig. 9 for set 2.

Fig. 9

Preprocessed and normalized set.

Preprocessed and normalized set.

5.3Data exploration

The results of data exploration are presented in Figs. 13 in the proposed work section.

5.4Attribute selection using HCSO-PSO

The optimization process performance will be evaluated using the convergence curve. The convergence curve is the graph drawn between the iterations and fitness function. This shows the optimization performance speed in finding the optimal attribute selection.

5.5First stage attribute selection using CSO

Figure 10 shows that the CSO finds the six optimal attributes after 30 iterations by maintaining the error rate as minimum below 0.2 for all the sets. Apart from dataset 2, in datasets 1 and 3, the optimal attributes were selected from the set. In set 2, only six attributes are present; hence, all the six attributes were selected for the second stage process. In set 1, the student behavior and education were selected. In set 3, the student behavior like absences, study time, romantic, internet, and education g1, g2 was selected for set 3. These reduced sets were subjected to the second stage of processing using particle swarm optimization.

Fig. 10

CSO convergence.

CSO convergence.

Figure 11 shows that the CSO find the six optimal attributes after 30 iterations by maintaining the error rate as minimum below 0.2 for all the sets. In set 1, the student performance like discussion, view, raising hand, and response was selected as optimal attributes. In set 2, apart from lunch and place of birth, all other attributes are selected. In set 3, the grade performance, study time, and romantic were selected. Then, the selected attributes were split as training and testing and then subjected to regression neural network modeling.

Fig. 11

PSO convergence.

PSO convergence.

Around 50 times of the iterations were performed, and various fitness function were used as inputs. The accuracy, precision and recall were then assessed as shown in Fig. 12. As a result, the existing methods achieves 95.65% and 77.98% of accuracy, where the proposed achieves 98.9% which is3.35% and 11.08% better than aforementioned methods. Then, the proposed modeled network performance was evaluated using evaluation metrics and the corresponding values are tabulated in Table 5.

Fig. 12

Comparison for various parameters.

Comparison for various parameters.

From Table 5, it was observed that the proposed HCSO-PSO-based RNN is the best model by having higher accuracy in training and testing. This proves that the proposed student’s behavioral and past performance is best for predicting their academic performance.

Table 5

Performance comparison

AlgorithmTesting (%)Training (%)Precision (%)Recall (%)Accuracy (%)
Proposed HCSO-PSO RNN98.798.996.296.498.9
Neural-Based Learning Algorithm (NBLA)85.6588.9888.9891.8995.65
Existing Method:80.97980.9870.7877.98
Fuzzy Learning Model (FLM)

6Discussion

This study’s findings have several ramifications for academic advisors. Academic advisors could start conversations about the types of study strategies students are using, the number of hours per week they devote to studying, and the ways they use self-regulation strategies like picking study locations and dealing with distractions because the participants benefited from the study skills they learned during the intervention. advisers can develop courses or teach study techniques to one or a small group of students, regardless of their work contexts and areas of expertise (e.g., professionals at face-to-face conferences, group facilitators, academic members with significant caseloads, or online advisers for distance learners). Teaching students how to learn just addresses a portion of the issue for college-bound students. This study suggests that in order for students to learn and successfully apply the concepts in their lessons, they need time. Advisors can and should be the ones to explain the use and benefit of these study techniques and to motivate students to keep using their newly acquired strategies. They are frequently better equipped than others on campus to see issues and follow up for a few weeks or even months.

6.1Implications and applications

Data mining methods are effective at locating hidden knowledge in huge datasets. When properly implemented in the educational sector, they can have significant positive effects on higher education institutions, students, education policymakers, and the prosperity of the nation. This work made a beneficial contribution to the field of educational data mining by finding the elements linked to academic probation. Policymakers, professionals, and educators can benefit from the findings by having a better understanding of the causes of the academic failure that can result in probation status. By taking into consideration certain other aspects besides high school performance, this can assist management in making better strategic decisions about the admission standards to universities. The study can also aid in determining the best strategy for pre-major and college admission requirements. The results of this study may be useful to academic advisors as well. According to substantial data, academic supervisors can, for instance, become more aware of the elements that have a negative impact on students’ performance and adjust their advising approaches. Students can also comprehend and steer clear of the issues that contribute to subpar academic achievement. Additionally, after a student enrols in an institution, the proposed model can be used to forecast their academic performance there. The connected educational stakeholders can take the necessary actions in accordance with the model’s predicted results. To avoid any barriers to academic growth, faculty should keep track of their advisees’ academic performance, meet with them twice a semester, and develop appropriate action plans. Students who are on supervision or who are at risk of being placed on probation should get ongoing follow-up sessions from them. Students will successfully exit the probationary period thanks to this. Additionally, parents should, whenever possible, be involved in their child’s academic circumstances. Parents who are well-informed could work together to create a much better learning environment for their children.

7Conclusions

In this technology world, education is also becoming one of the basic necessities of human life like food, shelter, and clothes. Even in day-to-day daily activities, the world is moving toward an automated process using technology developments. Some of the technology developments in day-to-day life activities are smartphone, internet activities, and home and office appliances. Demographic elements including parent’s educational level, environment, and one’s own desire in education are among the many aspects that influence an individual’s education. On the datasets, a variety of artificial intelligence algorithms have been used to analyze these variables and improve the teaching strategy. These models were not enough for analysing every element that has an impact on education. In this paper, a single, optimized machine-learning strategy is suggested. From the result evaluation, it was observed that the proposed HCSO-PSO-based RNN is the best model by having higher accuracy in training and testing. This proves that the proposed student behavioral and past performance is best for predicting their academic performance.

The results of this study indicate that some students may not be ready for the rigours of postsecondary education when they first enrol in higher education. The qualitative results strongly imply that research participants who were put on academic probation lacked basic academic skills, including the ability to study, communicate with their teachers, and complete required reading. They also employed scant study, self-control, and time management techniques. The negative effects of high school and family experiences, which do not adequately prepare students for success on their own in college, are also highlighted by the findings The study’s subjects had completed high school with honours. However, other people lacked the drive or abilities to seek assistance when they were struggling academically. They also lacked the study, time management, and other skills required to deal with academic obstacles. Future studies should look at the intricate relationships between low performance in college students and the framework academic advisers might create to address bad study habits, either through individual interventions or Learning Skills Courses (LSCs). These initiatives may serve as a basis for developing effective, differentiated initiatives to improve learning and expand the motivating and learning techniques employed by college.

Funding statement

The author received no specific funding for this study.

Conflicts of interest

The authors declare that they have no conflicts of interest to report regarding the present study.

References

[1] 

Kaunang F.J. , Rotikan R. , Students’ academic performance prediction using data mining. In 2018 Third International Conference on Informatics and Computing (ICIC), IEEE, pp. 1–5, 2018. Available: http://dx.doi.org/10.1109/IAC.2018.8780547.

[2] 

Saheed Y.K. , Oladele T.O. , Akanni A.O. , Ibrahim W.M. , Student performance prediction based on data mining classification techniques, Nigerian Journal of Technology 37: (4) ((2018) ), 1087–1091. Available: https://doi.org/10.4314/njt.374.1874.

[3] 

Alloghani M. , Al-Jumeily D. , Hussain A. , Aljaaf A.J. , Mustafina J. , Petrov E. , September: Application of machine learning on student data for the appraisal of academic performance. In 2018 11th International Conference on Developments in ESystems Engineering (DeSE), IEEE, pp. 157–162, 2018. Available: https://doi.org/10.1109/DeSE.2018.00038.

[4] 

Nagy M. , Molontay R. , Predicting dropout in higher education based on secondary school performance. In 2018 IEEE 22nd International Conference on Intelligent Engineering Systems (INES), IEEE, pp. 000389–000394, 2018. Available: https://doi.org/10.1109/INES.2018.8523888.

[5] 

Kim B.H. , Vizitei E. , Ganapathi V. , GritNet: Student performance prediction with deep learning, arXiv preprint arXiv:1804.07405, 2018. Available: https://doi.org/10.48550/arXiv.1804.07405.

[6] 

Sekeroglu B. , Dimililer K. , Tuncal K. , Student performance prediction and classification using machine learning algorithms. In Proceedings of the 2019 8th International Conference on Educational and Information Technology, pp. 7–11, 2019. Available: https://doi.org/10.1145/3318396.3318419.

[7] 

Hasan H.R. , Rabby A.S.A. , Islam M.T. , Hossain S.A. , Machine learning algorithm for student’s performance prediction. In 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), IEEE, pp. 1–7, 2019.

[8] 

Jain A. , Solanki S. , An efficient approach for multiclass student performance prediction based upon machine learning. In 2019 International Conference on Communication and Electronics Systems (ICCES), IEEE, pp. 1457–1462, 2019.

[9] 

Imran M. , Latif S. , Mehmood D. , Shah M.S. , Student academic performance prediction using supervised learning techniques, International Journal of Emerging Technologies in Learning 14: (14) ((2019) ).

[10] 

Hashim A.S. , Awadh W.A. , Hamoud A.K. , Student performance prediction model based on supervised machine learning algorithms., IOP Publishing, In IOP Conference Series: Materials Science and Engineering 928: (3) ((2020) ), 032019.

[11] 

Ha D.T. , Loan P.T.T. , Giap C.N. , Huong N.T.L. , An empirical study for student academic performance prediction using machine learning techniques, International Journal of Computer Science and Information Security (IJCSIS) 18: (3) ((2020) ), 21–28.

[12] 

Rastrollo-Guerrero J.L. , Gómez-Pulido J.A. , Durán-Domínguez A. , Analyzing and predicting students’performance by means of machine learning: A review, AppliedSciences 10: (3) ((2020) ), 1042.

[13] 

Ranjeeth S. , Latchoumi T.P. , Paul P.V. , Optimal stochastic gradient descent with multilayer perceptron based student’s academic performance prediction model, Recent Advances in Computer Science and Communications (Formerly: Recent Patents on Computer Science) 14: (6) ((2021) ), 1728–1741.

[14] 

Zeineddine H. , Braendle U. , Farah A. , Enhancing prediction ofstudent success: Automated machine learning approach, , Computers& Electrical Engineering 89: ((2021) ), 106903.

[15] 

Khan A. , Ghosh S.K. , Student performance analysis and prediction in classroom learning: A review of educational data mining studies, Education and Information Technologies 26: (1) ((2021) ), 205–240.

[16] 

Aljarah I. , Students’ Academic Performance Dataset, Kaggle, 2016. Available: https://www.kaggle.com/datasets/aljarah/xAPI-Edu-Data..

[17] 

Seshapanpu J. , Students performance in exams, Kaggle, 2018. Available: https://www.kaggle.com/datasets/spscientist/students-performance-in-exams.

[18] 

Sean D.S. , Student performance data set, Kaggle, 2020. Available: https://www.kaggle.com/datasets/larsen0966/student-performance-data-set.

[19] 

Yang X.S. , Deb S. , Cuckoo search via Lévy flights. In 2009 World Congress on Nature & Biologically Inspired Computing (NaBIC), IEEE, pp. 210–214, 2009.

[20] 

Cardona T. , Cudney E. , Hoerl R. , Snyder J. , Data mining and machine learning retention models in higher education, Journal of College Student Retention: Research, Theory & Practice 25: (1) ((2023) ), 51–75.

[21] 

Mamaghan K. , Maryam , Mohammadi M. , Meyer P. , Mohammad A. , El-Ghazali Talbi , Machine learning at the service of meta-heuristics for solving combinatorial optimization problems: A state-of-the-art, European Journal of Operational Research 296: (2) ((2022) ), 393–422.

[22] 

Pallathadka H. , Wenda A. , Ramirez-Asís E. , Asís-López M. , Flores-Albornoz J. , Phasinam K. , Classification and prediction of student performance data usingvarious machine learning algorithms, , Materials Today:Proceedings 80: ((2023) ), 3782–3785.

[23] 

Yağcí Mustafa , Educational data mining: Prediction of students’ academic performance using machine learning algorithms, Smart Learning Environments 9: (1) ((2022) ), 11.

[24] 

Okoye Kingsley , Arturo Arrona-Palacios , Claudia Camacho-Zuñiga , Joaquín Alejandro Guerra Achem , Jose Escamilla , Samira Hosseini , Towards teaching analytics: A contextual model for analysis of students’ evaluation of teaching through text mining and machine learning classification, Education and Information Technologies (2022), 1–43.

[25] 

Chung Jae Young , Sunbok Lee , Dropout early warning systems for high school students using machine learning, , Children and Youth Services Review 96: ((2019) ), 346–353.