Skip to content
BY 4.0 license Open Access Published by De Gruyter February 16, 2023

Evaluation and analysis of teaching quality of university teachers using machine learning algorithms

  • Ying Zhong EMAIL logo

Abstract

In order to better improve the teaching quality of university teachers, an effective method should be adopted for evaluation and analysis. This work studied the machine learning algorithms and selected the support vector machine (SVM) algorithm to evaluate teaching quality. First, the principles of selecting evaluation indexes were briefly introduced, and 16 evaluation indexes were selected from different aspects. Then, the SVM algorithm was used for evaluation. A genetic algorithm (GA)-SVM algorithm was designed and experimentally analyzed. It was found that the training time and testing time of the GA-SVM algorithm were 23.21 and 7.25 ms, both of which were shorter than the SVM algorithm. In the evaluation of teaching quality, the evaluation value of the GA-SVM algorithm was closer to the actual value, indicating that the evaluation result was more accurate. The average accuracy of the GA-SVM algorithm was 11.64% higher than that of the SVM algorithm (98.36 vs 86.72%). The experimental results verify that the GA-SVM algorithm can have a good application in evaluating and analyzing teaching quality in universities with its advantages in efficiency and accuracy.

1 Introduction

With the development of society, the scale of university education has been expanding, and the teaching quality of universities has been widely concerned as an important criterion for judging the excellence of universities. Evaluating teachers’ teaching quality can help people grasp the teaching effect of universities and help teachers, students, and school administrators better understand the shortcomings in the teaching process, thus promoting the efficiency of teachers’ teaching and school management. Aiming the evaluation of classroom teaching quality, Cheng and Yan [1] combined a generalized neural network (GRNN) with the fruit fly optimization algorithm to enhance the learning ability of GRNN. They found through experiments that the improved algorithm had smaller errors. Dong et al. [2] used a diversified evaluation method and applied mathematical fuzzy hierarchical analysis to evaluate teachers’ teaching quality scientifically and objectively. Jian [3] designed a set of evaluation indexes suitable for multimedia teaching quality, tested the evaluation results with the genetic algorithm (GA), and verified the method’s effectiveness through experiments. Bao and Yu [4] proposed a teaching quality evaluation method based on mobile edge computing for mixed online and offline sports teaching and found through experiments that the method had a small error and high efficiency. Evaluating and analyzing the teaching quality of university teachers with a reasonable and effective method can provide reliable data support and a theoretical basis for teaching and management work, thus promoting the steady development of university talent training [5]. With the development of technology, machine learning methods have appeared, and decision trees and association rules have been well applied in teaching quality evaluation, proving the reliability of machine learning methods. This work studied the usability of the support vector machine (SVM) method, aiming to further understand the value of the SVM method for evaluating teaching quality. To improve the performance of the SVM algorithm, its parameters were optimized by the GA to obtain a GA-SVM model. Moreover, an experimental analysis was performed to verify the effectiveness of the GA-SVM model. This study combined two machine learning algorithms and utilized the advantages of the GA to optimize the SVM algorithm, contributing to the further application of the SVM algorithm in the field of education. The results of this study verified the reliability of the optimized SVM method for evaluating teaching quality. The method can be applied in the teaching quality evaluation of more courses.

2 Evaluation index of teaching quality of university teachers

Evaluating teachers’ teaching quality can help people understand the quality of teachers and their work and find out the shortcomings in time to improve the quality of teaching teams and the teaching quality of universities [6]. Therefore, teaching quality evaluation and analysis must be objective, fair, and scientific, while many methods used at present have strong subjectivity and arbitrariness. The evaluation of the teaching quality of teachers in universities should use scientific, reasonable, valid, and reliable indicators; therefore, this study selected evaluation indicators according to the following principles:

  1. scientific: the chosen indicators should be consistent with the objective laws of teaching and learning;

  2. complete: the chosen indicators should be as comprehensive as possible;

  3. measurable: the chosen metrics should be measurable, explicit, and able to be grasped intuitively;

  4. orientated: the chosen indicators should be in line with modern teaching concepts and conducive to the comprehensive development cultivation of talents;

  5. valid: the chosen indicators should effectively evaluate the teachers’ lesson preparation and knowledge reserve.

Based on the above principles and the teaching content of university teachers, this study selected indicators from four aspects, as shown in Table 1.

Table 1

Teaching quality evaluation indicator system

First-level indicator Second-level indicator
Teaching attitude X1: standard mandarin, clear, and accurate expression
X2: well-prepared teaching and careful lesson preparation
X3: neatly written board, practical classroom materials
X4: teach by example, love, and dedication
Teaching method X5: active use of various teaching aids
X6: teaching is well-detailed and highlights important and difficult points
X7: be able to use modern teaching media scientifically
X8: active classroom atmosphere, focusing on communication and interaction with students
Teaching content X9: the teaching content is enriched, and the pace is reasonable
X10: consistent with syllabus and course plan
X11: focus on combining theory and practice
X12: focus on developing students’ abilities and emotions
Teaching effect X13: students can understand and learn what the teacher is teaching
X14: students have a strong sense of innovation
X15: students have a strong interest in learning
X16: students have strong problem-solving skills

Table 1 includes 16 indicators, and the value of every indicator is 100 points. All the 16 indicators are applicable to the teaching quality evaluation of teachers who teach different courses, which can be applied school-wide. The evaluation indicators are clear at a glance and easy for students to score.

3 SVM-based evaluation algorithm

The SVM algorithm is a kind of machine learning algorithm [7] that can solve small-sample, nonlinear problems [8] and has been commonly used in pattern recognition [9], computer vision [10], and image classification [11]. Suppose there is a sample set S as follows:

(1) S = { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x i , y i ) } , y i { 1 , 1 } .

The classification hyperplane is written as:

(2) ( w · x ) + b = 0 ,

where w is the weight and b is the bias. The problem of finding the optimal hyperplane is expressed as follows:

(3) min w , b 1 2 w 2 ,

(4) s . t . y i [ ( w · x ) + b ] 1 0 .

However, in practical problems, the data are mostly linearly inseparable [12]; thus, it is necessary to introduce penalty parameter C and relaxation factor ξ , then,

(5) min w , b 1 2 w 2 + C i = 1 l ξ i ,

(6) s . t . y i [ ( w · x ) + b ] 1 + ξ i 0 .

Then, the final optimization problem can be written as follows:

(7) max F ( a ) = i = 1 l a i 1 2 i , j = 1 l a i a j y i y j K ( x i , x j ) ,

where a is the Lagrangian multiplier, a = ( a i , , a l ) , and K ( x i , x j ) is the kernel function, the radial basis function (RBF) in this study is as follows:

(8) K ( x i , x j ) = exp ( γ x i x j 2 ) , γ = 1 2 σ 2 ,

where σ is the parameter of the kernel function.

The final classification function of the SVM algorithm can be written as follows:

(9) b ' = y i i = 1 l a i ' y i K ( x i , x j ) ,

(10) f ( x ) = sgn i = 1 l a i ' y i K ( x i , x j ) + b ' ,

where sgn represents the sign function and b ' represents the classification threshold value.

The performance of the SVM algorithm is mainly affected by two parameters [13]. In order to further improve the performance of the SVM algorithm, it was improved by using GA in this study. The GA is a stochastic search algorithm [14], which searches for the optimal solution by simulating the natural evolutionary process [15], and has good applications in data prediction [16], industrial control [17], parameter optimization [18], etc. The specific steps of optimizing parameters with the GA are as follows.

  1. Initial population: an initial population was established. The binary encoding method was used. Every individual contained two random numbers, i.e., parameters C and σ of the SVM algorithm.

  2. Fitness function: the fitness function of every individual, i.e., the minimum mean squared error of the SVM algorithm, was calculated.

  3. Selection: the operation was selected using the roulette algorithm [19], and the best individual entered the next generation.

  4. Crossover: individuals were randomly paired, and some of the chromosomes were exchanged based on a probability.

  5. Mutation: genetic values at some positions were changed based on probability.

  6. The computation was repeated for the fitness function until the termination condition was satisfied. Finally, the optimal parameters C and σ of the SVM algorithm were obtained.

The parameters optimized by the GA were input into the SVM model to obtain the GA-SVM model. The GA-SVM model can be applied to evaluate the teachers’ teaching quality. It needs to be trained using training data and then tested to obtain the evaluation results.

4 Experimental analysis

4.1 Experimental setup

The experimental environment was Windows 10. The development language was Python. The experimental platform was eclipse+pydev. The size of the initial population of the GA was 20, the crossover probability was set as 0.7, the variation probability was set as 0.03, the objective function was 0.0001, and the maximum number of iterations was 5,000. The SVM parameters of the GA output are as follows: C = 100, and σ = 1.25. The experimental data used included two types, and one was the standard test dataset for machine learning, coming from University of California, Irvine (UCI) [20]. Five datasets were selected from the database to analyze the performance of the algorithms. The five datasets are shown in Table 2.

Table 2

UCI datasets

Datasets Instances Number of classes Number of features
Iris 150 3 4
Wine 178 3 13
Glass 214 6 9
Diabetes 768 2 8
Heartstatlog 270 2 13

The other type came from the Guilin University. The evaluation indicator system proposed above was applied in Guilin University in the first semester of 2020. The evaluation data were collected for the experiment. The final evaluation grade for teachers’ teaching quality was given according to all the scores of the indicators and expert opinions. The grades included excellent, good, and poor, expressed by 1, 0 and -1. After excluding abnormal and missing data, a total of 380 data were obtained, 70% of which were used for training the SVM algorithm and the remaining 30% were used for testing. Every sample included 16 features, and the range of values was [0,100]. Some of the sample data are shown in Table 3.

Table 3

Sample data examples (unit: point)

Sample number 1 2 3 4 5 6 7
X1 95 90 70 85 60 92 84
X2 95 89 70 85 60 85 82
X3 95 88 71 90 60 98 81
X4 90 87 65 86 65 96 80
X5 90 95 65 87 65 93 84
X6 85 98 60 78 65 97 83
X7 94 80 60 78 60 94 75
X8 92 95 72 79 60 91 79
X9 94 94 73 80 60 95 78
X10 89 93 68 81 77 86 76
X11 88 92 69 81 76 89 74
X12 87 89 70 85 74 85 75
X13 97 86 70 85 72 89 77
X14 98 91 65 86 73 89 72
X15 95 92 65 87 68 94 71
X16 95 95 70 88 68 96 69
Y Excellent Excellent Poor Good Poor Excellent Good

To facilitate the experiment, the sample data were normalized, and the range was reduced to [0,1]. The calculation formula is as follows:

(11) x i ' = x i x min x max x min .

Then, cross-validations were performed ten times to compare the effectiveness of the SVM algorithm and the improved GA-SVM algorithm in evaluating the teaching quality.

5 Experimental results

The SVM and GA-SVM algorithms were compared with the Naive Bayes (NB) algorithm [21], the K-neural network (KNN) algorithm [22], and the decision tree (DT) algorithm [23], and the results are shown in Table 4.

Table 4

Comparison of accuracy between different machine learning algorithms

Datasets NB (%) KNN (%) DT (%) SVM (%) GA-SVM (%)
Iris 64.27 70.21 75.36 88.64 97.33
Wine 60.18 71.26 74.82 86.32 95.67
Glass 62.36 72.64 76.41 85.71 98.03
Diabetes 63.49 70.88 77.29 86.93 97.29
Heartstatlog 65.77 73.59 75.87 84.71 96.84

It was seen from Table 4 that when classifying different datasets, the accuracy of the NB algorithm was below 70%, the lowest, and the accuracy of the KNN and DT algorithms was between 70 and 80%, slightly higher than the NB algorithm; the SVM algorithm had an accuracy above 80%, and its accuracy for the Iris dataset was 88.64%, which was 20% higher than the NB algorithm and 10% higher than the DT algorithm. These results demonstrated that the SVM algorithm had the best classification performance among these machine learning algorithms. Finally, the comparison between the SVM and GA-SVM algorithms showed that when classifying the Iris dataset, the accuracy of the GA-SVM was 97.33, 8.69% higher than the SVM algorithm, verifying that the improved algorithm was effective.

The comparison of the testing time is shown in Table 5.

Table 5

Comparison of the testing time between different machine learning algorithms (unit: s)

Datasets NB KNN DT SVM GA-SVM
Iris 0.12 0.09 0.08 0.07 0.04
Wine 0.14 0.11 0.09 0.06 0.05
Glass 0.37 0.35 0.33 0.29 0.21
Diabetes 1.76 1.35 1.07 0.92 0.78
Heartstatlog 0.68 0.55 0.46 0.37 0.23

It is seen from Table 5 that these algorithms all suggested similar features, i.e., the larger the sample size in the dataset is, the longer the testing time is. These algorithms spent the longest time in classifying the Diabetes algorithm and the shortest time in classifying the Iris dataset. The comparison of these algorithms showed that the testing time of the NB, KNN, and DT algorithms was longer, indicating that the SVM was more efficient in computation. The comparison between the SVM and GA-SVM algorithms demonstrated that when classifying the Diabetes dataset, the testing time of the GA-SVM algorithm was 0.78 s, which was 15.22% shorter than that of the SVM algorithm (0.9 s), indicating that the improvement of the SVM algorithm effectively improved the computational efficiency of the SVM algorithm.

First, the time required by the two algorithms for teaching quality evaluation was compared on the actual datasets, as shown in Figure 1.

Figure 1 
               Comparison of the running time between algorithms.
Figure 1

Comparison of the running time between algorithms.

It was seen from Figure 1 that the running time of the SVM algorithm was slightly longer. For the training of the algorithm, the running time of the SVM algorithm was 27.84 ms, while the time of the GA-SVM algorithm was 23.21 ms, which was 16.63% shorter than the SVM algorithm. For the testing of the algorithm, the running time of the SVM algorithm was 9.88 ms, and the running time of the GA-SVM algorithm was 7.25 ms, which was 26.62% shorter than the SVM algorithm. In the experiments, the difference in the running time between the two algorithms was less significant due to the small amount of data used, but the above results verified that the GA-SVM algorithm was more efficient.

The evaluation results of the two algorithms were compared, and some of the results are shown in Table 6.

Table 6

Comparison of partial evaluation results

Real evaluation results Evaluation results of SVM algorithm Evaluation results of GA-SVM algorithm
Evaluation value Grade Evaluation value Grade Evaluation value Grade
1 Excellent 1.112 Excellent 0.998 Excellent
1 Excellent 0.016 Good 0.215 Good
0 Good −0.864 Poor 0.124 Good
1 Excellent 1.213 Excellent 1.011 Excellent
−1 Difference −0.862 Poor −0.997 Poor
−1 Difference −1.364 Poor −1.012 Poor
0 Good −0.979 Poor 0.016 Good

Bold: The result of algorithm evaluation is different from that of real evaluation.

It was seen from Table 6 that among the seven samples, the error between the evaluation value of the SVM algorithm and the actual value was large, and three samples were wrongly evaluated. For example, the sample with an “excellent” result was evaluated as “good,” and the sample with a “good” result was evaluated as “poor.” In contrast, the evaluation value of the GA-SVM algorithm was closer to the real value, i.e., more accurate, and only one sample was wrongly evaluated. The results verified that the GA-SVM algorithm was reliable in evaluating the teachers’ teaching quality.

Finally, the evaluation accuracy of the two methods was compared, as shown in Figure 2.

Figure 2 
               Comparison of the evaluation accuracy between the two methods.
Figure 2

Comparison of the evaluation accuracy between the two methods.

It was seen from Figure 2 that in the ten times of cross-validation, there was a large gap in the accuracy between the two methods, and the accuracy of the SVM algorithm ranged from 80 to 90%, with a maximum of 89.91% and a minimum of 82.78%, the stability of the evaluation results obtained by the SVM algorithm was also not very high, i.e., there were fluctuations, while the accuracy of the GA-SVM algorithm always stayed above 95%, with a maximum of 98.93% and a minimum of 97.89%. The average accuracy of the GA-SVM algorithm was 11.64% higher than that of the SVM algorithm (98.36 vs 86.72%), verifying the effectiveness of the GA-SVM algorithm in evaluating the teachers’ teaching quality. The above results demonstrated that the GA-SVM algorithm could accurately evaluate the teaching quality and be further promoted and applied in practice.

6 Conclusion

This work studied the evaluation and analysis of teaching quality in universities with the SVM algorithm. The SVM algorithm was improved by optimizing the parameters of the SVM algorithm after selecting evaluation indicators. Then, the GA-SVM algorithm was applied to the dataset of the actual evaluation results of teaching quality. The experimental results showed that the GA-SVM algorithm had a shorter running time and obtained results closer to the actual situation. The accuracy rate of the GA-SVM algorithm was 11.64% higher than the SVM algorithm, which verified the method’s reliability. This work provides some theoretical bases for better evaluation of the teaching quality of university teachers.

  1. Conflict of interest: The author states no conflict of interest.

References

[1] Cheng J, Yan X. The quality evaluation of classroom teaching based on FOA-GRNN. Proc Comput Sci. 2017;107:355–60.10.1016/j.procs.2017.03.117Search in Google Scholar

[2] Dong QW, Wang SM, Han FJ, Zhang RD. Innovative research and practice of teachers’ teaching quality evaluation under the guidance of ‘innovation and entrepreneurship’. Proc Comput Sci. 2019;154:770–6.10.1016/j.procs.2019.06.123Search in Google Scholar

[3] Jian Q. Multimedia teaching quality evaluation system in colleges based on genetic algorithm and social computing approach. IEEE Access. 2019;7:1.10.1109/ACCESS.2019.2957447Search in Google Scholar

[4] Bao L, Yu P. Evaluation method of online and offline hybrid teaching quality of physical education based on mobile edge computing. Mob Netw Appl. 2021;2021:1–11.10.1007/s11036-021-01774-wSearch in Google Scholar

[5] Duong T, Do TD, Nguyen NP. Exploiting faculty evaluation forms to improve teaching quality: An analytical review. Science & Information Conference. vol. 2015; 2015. p. 457–62.10.1109/SAI.2015.7237182Search in Google Scholar

[6] Teachers AC, Adams T, Aguilar E, Berg E, Cismowski L, Cody A, et al. A coherent system of teacher evaluation for quality teaching. Educ Policy Anal Arch. 2015;23:26.10.14507/epaa.v23.2006Search in Google Scholar

[7] Sun W, Chang C, Long Q. Bayesian non-linear support vector machine for high-dimensional data with incorporation of graph information on features. 2019 IEEE International Conference on Big Data (Big Data). vol. 2019, 2019; p. 4874–82.10.1109/BigData47090.2019.9006473Search in Google Scholar

[8] Singh C, Walia E, Kaur KP. Enhancing color image retrieval performance with feature fusion and non-linear support vector machine classifier. Opt - Int J Light Electron Opt. 2018;158:127–41.10.1016/j.ijleo.2017.11.202Search in Google Scholar

[9] Ghimire D, Lee J. Geometric feature-based facial expression recognition in image sequences using multi-class adaboost and support vector machines. Sensors. 2016);13:7714–34.10.3390/s130607714Search in Google Scholar PubMed PubMed Central

[10] Bui DT, Tran AT, Klempe H, Pradhan B, Revhaug I. Spatial prediction models for shallow landslide hazards: a comparative assessment of the efficacy of support vector machines, artificial neural networks, kernel logistic regression, and logistic model tree. Landslides. 2016;13:361–78.10.1007/s10346-015-0557-6Search in Google Scholar

[11] Cotrim W, Felix LB, Minim V, Campos RC, Minim L. Development of a hybrid system based on convolutional neural networks and support vector machines for recognition and tracking color changes in food during thermal processing. Chem Eng Sci. 2021;240:116679.10.1016/j.ces.2021.116679Search in Google Scholar

[12] Nie F, Zhu W, Li X. Decision Tree SVM: An extension of linear SVM for non-linear classification. Neurocomputing. 2020;401:153–9.10.1016/j.neucom.2019.10.051Search in Google Scholar

[13] Kondo H, Asanuma Y. Optimization of hyperparameters and feature selection for random forests and support vector machines by artificial bee colony algorithm. Trans Jpn Soc Artif Intell. 2019;34:G-I36_1–11.10.1527/tjsai.G-I36Search in Google Scholar

[14] Zhang Y, Liu M. An improved genetic algorithm encoded by adaptive degressive ary number. Soft Comput. 2018;22:6861–75.10.1007/s00500-017-2981-0Search in Google Scholar

[15] Farsi MA. Genetic algorithms: Principles and application in RAMS. Nature-Inspired Comput Paradig Syst. 2021;2021:25–46.10.1016/B978-0-12-823749-6.00001-5Search in Google Scholar

[16] Yang HJ, Hu X. Wavelet neural network with improved genetic algorithm for traffic flow time series prediction. Opt - Int J Light Electron Opt. 2016;127:8103–10.10.1016/j.ijleo.2016.06.017Search in Google Scholar

[17] Volkanovski A, Mavko B, Boševski T, Chaushevski A, Čepin M. Genetic algorithm optimisation of the maintenance scheduling of generating units in a power system. Reliab Eng Syst Safe. 2017;93:779–89.10.1016/j.ress.2007.03.027Search in Google Scholar

[18] Hraiech SE, Chebbi AH, Affi Z, Romdhane L. Genetic algorithm coupled with the Krawczyk method for multi-objective design parameters optimization of the 3-UPU manipulator. Robotica. 2020;38:1138–54.10.1017/S0263574719001292Search in Google Scholar

[19] Cerf R. The quasispecies regime for the simple genetic algorithm with roulette wheel selection. Adv Appl Probab. 2017;49:903–26.10.1017/apr.2017.26Search in Google Scholar

[20] http://archive.ics.uci.edu/ml/datasets.php.Search in Google Scholar

[21] Edelson M, Kuo TT. Generalizable prediction of COVID-19 mortality on worldwide patient data. JAMIA Open. 2022;5:1–9.10.1093/jamiaopen/ooac036Search in Google Scholar PubMed PubMed Central

[22] Wu H, Zhao YP, Tan HJ. A hybrid of fast K-nearest neighbor and improved directed acyclic graph support vector machine for large-scale supersonic inlet flow pattern recognition. P I Mech Eng G-J AER. 2022;136:109–22.10.1177/09544100211008601Search in Google Scholar

[23] Pradeepika A, Sabitha R. Examination of diabetes mellitus for early forecast using decision tree classifier and an innovative dependent feature vector based Naive Bayes classifier. ECS Trans. 2022;107:12937–52.10.1149/10701.12937ecstSearch in Google Scholar

Received: 2022-05-27
Revised: 2022-10-17
Accepted: 2022-11-05
Published Online: 2023-02-16

© 2023 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 16.5.2024 from https://www.degruyter.com/document/doi/10.1515/jisys-2022-0204/html
Scroll to top button