IDEAS home Printed from https://ideas.repec.org/a/gam/jsusta/v15y2023i7p6229-d1116123.html
   My bibliography  Save this article

Evaluating Student Knowledge Assessment Using Machine Learning Techniques

Author

Listed:
  • Nuha Alruwais

    (Department of Computer Science and Engineering, College of Applied Studies and Community Services, King Saud University, P.O. Box 22459, Riyadh 11495, Saudi Arabia)

  • Mohammed Zakariah

    (Department of Computer Science, College of Computer Science and Information, King Saud University, P.O. Box 11442, Riyadh 11574, Saudi Arabia)

Abstract

The process of learning about a student’s knowledge and comprehension of a particular subject is referred to as student knowledge assessment. It helps to identify areas where students need additional support or challenge and can be used to evaluate the effectiveness of instruction, make important decisions such as on student placement and curriculum development, and monitor the quality of education. Evaluating student knowledge assessment is essential to measuring student progress, informing instruction, and providing feedback to improve student performance and enhance the overall teaching and learning experience. This research paper is designed to create a machine learning (ML)-based system that assesses student performance and knowledge throughout the course of their studies and pinpoints the key variables that have the most significant effects on that performance and expertise. Additionally, it describes the impact of running models with data that only contains key features on their performance. To classify the students, the paper employs seven different classifiers, including support vector machines (SVM), logistic regression (LR), random forest (RF), decision tree (DT), gradient boosting machine (GBM), Gaussian Naive Bayes (GNB), and multi-layer perceptron (MLP). This paper carries out two experiments to see how best to replicate the automatic classification of student knowledge. In the first experiment, the dataset (Dataset 1) was used in its original state, including all five properties listed in the dataset, to evaluate the performance indicators. In the second experiment, the least correlated variable was removed from the dataset to create a smaller dataset (Dataset 2), and the same set of performance indicators was evaluated. Then, the performance indicators using Dataset 1 and Dataset 2 were compared. The GBM exhibited the highest prediction accuracy of 98%, according to Dataset 1. In terms of prediction error, the GBM also performed well. The accuracy of optimistic forecasts on student performance, denoted as the performance indicator ‘precision’, was highest in GBM at 99%, while DT, RF, and SVM were 98% accurate in their optimistic forecasts for Dataset 1. The second experiment’s findings demonstrated that practically no classifiers showed appreciable improvements in prediction accuracy with a reduced feature set in Dataset 2. It showed that the time required for related learning objects and the knowledge level corresponding to a goal learning object have less impact.

Suggested Citation

  • Nuha Alruwais & Mohammed Zakariah, 2023. "Evaluating Student Knowledge Assessment Using Machine Learning Techniques," Sustainability, MDPI, vol. 15(7), pages 1-25, April.
  • Handle: RePEc:gam:jsusta:v:15:y:2023:i:7:p:6229-:d:1116123
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2071-1050/15/7/6229/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2071-1050/15/7/6229/
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Dario Sansone, 2019. "Beyond Early Warning Indicators: High School Dropout and Machine Learning," Oxford Bulletin of Economics and Statistics, Department of Economics, University of Oxford, vol. 81(2), pages 456-485, April.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Delogu, Marco & Lagravinese, Raffaele & Paolini, Dimitri & Resce, Giuliano, 2024. "Predicting dropout from higher education: Evidence from Italy," Economic Modelling, Elsevier, vol. 130(C).
    2. Filmer,Deon P. & Nahata,Vatsal & Sabarwal,Shwetlena, 2021. "Preparation, Practice, and Beliefs : A Machine Learning Approach to Understanding Teacher Effectiveness," Policy Research Working Paper Series 9847, The World Bank.
    3. Hazal Colak Oz & Çiçek Güven & Gonzalo Nápoles, 2023. "School dropout prediction and feature importance exploration in Malawi using household panel data: machine learning approach," Journal of Computational Social Science, Springer, vol. 6(1), pages 245-287, April.
    4. Isphording, Ingo E. & Raabe, Tobias, 2019. "Early Identification of College Dropouts Using Machine-Learning: Conceptual Considerations and an Empirical Example," IZA Research Reports 89, Institute of Labor Economics (IZA).
    5. Maria do Carmo Nicoletti & Osvaldo Luiz de Oliveira, 2020. "A Machine Learning-Based Computational System Proposal Aiming at Higher Education Dropout Prediction," Higher Education Studies, Canadian Center of Science and Education, vol. 10(4), pages 1-12, December.
    6. Ashesh Rambachan & Amanda Coston & Edward Kennedy, 2022. "Robust Design and Evaluation of Predictive Algorithms under Unobserved Confounding," Papers 2212.09844, arXiv.org, revised May 2024.
    7. McKenzie, David & Sansone, Dario, 2019. "Predicting entrepreneurial success is hard: Evidence from a business plan competition in Nigeria," Journal of Development Economics, Elsevier, vol. 141(C).
    8. Liyang Tang, 2020. "Application of Nonlinear Autoregressive with Exogenous Input (NARX) neural network in macroeconomic forecasting, national goal setting and global competitiveness assessment," Papers 2005.08735, arXiv.org.
    9. Montorsi, Carlotta & Fusco, Alessio & Van Kerm, Philippe & Bordas, Stéphane P.A., 2024. "Predicting depression in old age: Combining life course data with machine learning," Economics & Human Biology, Elsevier, vol. 52(C).
    10. Miguel Angel Valles-Coral & Luis Salazar-Ramírez & Richard Injante & Edwin Augusto Hernandez-Torres & Juan Juárez-Díaz & Jorge Raul Navarro-Cabrera & Lloy Pinedo & Pierre Vidaurre-Rojas, 2022. "Density-Based Unsupervised Learning Algorithm to Categorize College Students into Dropout Risk Levels," Data, MDPI, vol. 7(11), pages 1-18, November.
    11. Bacon, Victoria R. & Kearney, Christopher A., 2020. "School climate and student-based contextual learning factors as predictors of school absenteeism severity at multiple levels via CHAID analysis," Children and Youth Services Review, Elsevier, vol. 118(C).

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jsusta:v:15:y:2023:i:7:p:6229-:d:1116123. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.