Essay Scoring Model Based on Gated Recurrent Unit Technique

Authors

  • Eluwa J.  Department of Computer Science, Babcock University, Ilishan-Remo, Ogun State, Nigeria
  • Kuyoro S.  Department of Computer Science, Babcock University, Ilishan-Remo, Ogun State, Nigeria
  • Awodele O.  Department of Computer Science, Babcock University, Ilishan-Remo, Ogun State, Nigeria
  • Ajayi A.  Department of Computer Science, Babcock University, Ilishan-Remo, Ogun State, Nigeria

DOI:

https://doi.org//10.32628/IJSRSET229257

Keywords:

Deep Neural Network (DNN), Global Vectorization (GloVe), Hyper Text Mark-up Language (HTML), Machine Learning (ML), Natural Language Processing (NLP).

Abstract

Educational evaluation is a major factor in determining students’ learning aptitude and academic performance. The scoring technique that relies solely on human labour is time consuming, costly, and logistically challenging as this rating is usually based on the opinion of “biased” human. Several studies have considered using machine learning techniques with feature extraction based on Term Frequency (TF) - Part of Speech (POS) Tagging without consideration to global vectorization (GloVe). These solutions require the process of selecting deterministic features that are directly related to essay quality which is time-consuming and needs a great deal of linguistic knowledge. Gated Recurrent Unit (a variation of Recurrent Neural Network) deep learning technique with focus on morphological analysis of essays for content-based assessment has therefore shown the capability of addressing the challenges posed by other AES techniques by building more abstract and complete linkages among features. Deep learning algorithms such as Multilayer Perceptron (MLP), Long Short-Term Memory (LSTM), and Gated Recurrent Unit (GRU) were used to learn the model with performance evaluation on metrics such as validation accuracy, training time, loss function, and Quadratic Weighted Kappa. The performance results showed that MLP, LSTM and GRU had average Quadratic Weighted Kappa (QWK) values of 0.65, 0.86 and 0.88 respectively with each algorithm having an average training time of 61.4, 62.68 and 67.86 seconds respectively. The loss functions for MLP, LSTM and GRU were 0.296, 0.24 and 0.126. This meant that GRU had the best estimate of the difference between the actual and forecasted scores. MLP, LSTM, and GRU had average validation accuracy of 0.48, 0.537, and 0.511 respectively. GRU was shown to be the optimal classifier and was used in the development of the essay scoring model.

References

  1. Shetty, S. & Siddiqa, A. (2019). Deep Learning Algorithms and Applications in Computer Vision. International Journal of Computer Sciences and Engineering. https://doi.org/10.26438/ijcse/v7i7.195201.
  2. Lim, C., Bong, C., Wong, W. & Lee, N. (2021). A Comprehensive Review of Automated Essay Scoring (AES) Research and Development. Pertanika Journal of Science & Technology. 29 (3): 1875 – 1899. https://doi.org/10.47836/pjst.29.3.27.
  3. Ramesh, D. & Sanampudi, S.K. (2021). An Automated Essay Scoring System: A system literature review. https://doi.org/10.1007/s10462-021-10068-2.
  4. Srivastava, K., Dhanda, N., & Shrivastava, A. (2020). An Analysis of Automatic Essay Grading Systems. International Journal of Recent Technology and Engineering (IJRTE). ISSN: 2277-3878, 8(6).
  5. Page, E. B, (1966). "The Imminence of Grading Essays by Computer". Phi Delta Kappan, 48:238-243.
  6. Hearst, M. A. (2000). The debate on automated essay grading. IEEE Intelligent Systems and their applications. 15(5), 22-37.
  7. Williams, R. (2001). Automated essay grading: An evaluation of four conceptual models. In New horizons in university teaching and learning: Responding to change. Centre for Educational Advancement, Curtin University. 173-184.
  8. Valenti, S., Neri, F., & Cucchiarelli, A. (2003). An overview of current research on automated essay grading. Journal of Information Technology Education: Research, 2(1), 319-330.
  9. Dong & Zhang (2016). Automatic Features for Essay Scoring – An Empirical Study. Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. 1072–1077.
  10. Taghipour, K., Ng, H. (2016). A neural approach to automated essay scoring. In: Proceedings of the 2016 conference on empirical methods in natural language processing. 1882–1891. https://doi.org/10.18653/v1/D16-1193.
  11. Wang Z., Liu J., & Dong R. (2018). Intelligent Auto-grading System. Proceedings of CCIS.
  12. Chen and Li (2018). Relevance-Based Automated Essay Scoring via Hierarchical Recurrent Model. In: 2018 International Conference on Asian Language Processing (IALP). 378–383. https://doi. org/ 10. 1109/ IALP. 2018. 86292 56.
  13. Dasgupta, T., Naskar, A., Dey, L., & Saha, R. (2018). Augmenting Textual Qualitative Features in Deep Convolution Recurrent Neural Network for Automatic Essay Scoring. Proceedings of the 5th Workshop on Natural Language Processing Techniques for Educational Applications. 93–102.
  14. Liang G, On B, Jeong D, Kim H & Choi G. (2018). Automated Essay Scoring: A Siamese Bidirectional LSTM Neural Network Architecture. Symmetry, 10(12), 682–. https://doi.org/10.3390/sym10120682.
  15. Ruseti, S., Dascalu, M., Johnson, A., McNamara, D., Balyan, R., McCarthy, K., & Trausan-Matu, S. (2018). Scoring summaries using recurrent neural networks. In: International Conference on Intelligent Tutoring Systems. 191–201.
  16. Chen, Z. & Zhou, Y. (2019). Research on Automatic Essay Scoring of Composition Based on CNN and OR. In: 2019 2nd International Conference on Artificial Intelligence and Big Data (ICAIBD). https://doi.org/10.1109/ICAIBD.2019.88370 07.
  17. Cai C. (2019). Automatic essay scoring with recurrent neural network. In: Proceedings of the 3rd International Conference on High Performance Compilation, Computing and Communications. https://doi.org/10.1145/3318265.3318296.
  18. Zupanc, K., & Bosnic, Z. (2017). Automated essay evaluation with semantic analysis. Knowledge-Based Systems, 118–132.
  19. Xia, L., Liu, J., Zhang, Z. (2019). Automatic essay scoring model based on two-layer bidirectional Long and Short-Term Memory Network. In: Proceedings of the 2019 3rd International Conference on Computer Science and Artificial intelligence. https://doi.org/10.1145/3374587.3374596. 133-137.
  20. Uto, M. & Okano, M. (2020). Robust Neural Automated Essay Scoring Using Item Response Theory. In: Artificial Intelligence in Education. AIED 2020. (12163). https://doi.org/10.1007/978-3-030-52237-7_44.
  21. Zhu, W. & Sun, Y. (2020). Automated essay scoring system using multi-model Machine Learning. Computer Science & Information Technology. 109-117. https://doi.org/10.5121/csit.2020.101211.
  22. Kuyoro, S., Eluwa, J., Awodele, O. & Ajayi, A. (2021). Characterization of Essay Content for Content-Based Assessment Using Morphological Classification Technique. International Journal of Scientific and Engineering Research. 12(1). ISSN:2229-5518.

Downloads

Published

2022-04-30

Issue

Section

Research Articles

How to Cite

[1]
Eluwa J., Kuyoro S., Awodele O., Ajayi A., " Essay Scoring Model Based on Gated Recurrent Unit Technique, International Journal of Scientific Research in Science, Engineering and Technology(IJSRSET), Print ISSN : 2395-1990, Online ISSN : 2394-4099, Volume 9, Issue 2, pp.323-330, March-April-2022. Available at doi : https://doi.org/10.32628/IJSRSET229257