Academia.eduAcademia.edu

Outline

Ontology-Based Multiple Choice Question Generation

2014, The Scientific World Journal

https://doi.org/10.1155/2014/274949

Abstract

With recent advancements in Semantic Web technologies, a new trend in MCQ item generation has emerged through the use of ontologies. Ontologies are knowledge representation structures that formally describe entities in a domain and their relationships, thus enabling automated inference and reasoning. Ontology-based MCQ item generation is still in its infancy, but substantial research efforts are being made in the field. However, the applicability of these models for use in an educational setting has not been thoroughly evaluated. In this paper, we present an experimental evaluation of an ontology-based MCQ item generation system known as OntoQue. The evaluation was conducted using two different domain ontologies. The findings of this study show that ontology-based MCQ generation systems produce satisfactory MCQ items to a certain extent. However, the evaluation also revealed a number of shortcomings with current ontology-based MCQ item generation systems with regard to the education...

FAQs

sparkles

AI

What factors influence the quality of ontology-based MCQs generated by OntoQue?add

The evaluation noted an average quality score of 8.8 out of 10, with distracter plausibility being a significant factor affecting quality. Specifically, rule violations identified include issues with distracters being too easy or not plausible, particularly in the HistOnto ontology.

How do different ontologies affect the educational significance of generated MCQs?add

The study found that MCQs generated from the SemQ ontology were considered more educationally significant than those from the HistOnto ontology, which had nearly half classified as irrelevant. This was attributed to the kind of domain knowledge represented, with factual knowledge being less useful.

What were the main methodologies used for MCQ generation in this study?add

The study utilized class-based, individual-based, and property-based strategies for MCQ generation in the OntoQue system. Each strategy involved deriving stems and distracters from RDF statements within ontological structures.

What recommendations were made for improving ontology-based MCQ generation systems?add

The paper recommends incorporating cognitive levels and learning objectives to enhance educational relevance and developing strategies to create plausible distracters. Additionally, it suggests integrating linguistic knowledge to improve item language and structure.

Which evaluation metrics were used to assess the quality of generated MCQs?add

Key metrics included the total number of worthy MCQ items, modification measures, and adherence to MCQ design rules, with non-revision rates signaling item quality.

References (30)

  1. M. Al-Yahya, "OntoQue: a question generation engine for edu- cational assesment based on domain ontologies, " in Proceedings of the 11th IEEE International Conference on Advanced Learning Technologies (ICALT '11), pp. 393-395, Athens, Ga, USA, July 2011.
  2. N. Knouf, "Transformation of bibTeX into an OWL ontology, " 2003, http://oaei.ontologymatching.org/2004/Contest/301/on- to.html.
  3. A. Papasalouros, K. Kanaris, and K. Kotis, "Automatic genera- tion of multiple choice questions from domain ontologies, " in Proceedings of the IADIS e-Learning Conference, pp. 427-434, Amsterdam, The Netherlands, July 2008.
  4. J. C. Brown, G. A. Frishkoff, and M. Eskenazi, "Automatic question generation for vocabulary assessment, " in Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing, pp. 819-826, Strouds- burg, Pa, USA, October 2005.
  5. R. Mitkov, L. A. Ha, and N. Karamanis, "A computer-aided environment for generating multiple-choice test items, " Natural Language Engineering, vol. 12, no. 2, pp. 177-194, 2006.
  6. J. H. Wolfe, "An Aid to Independent Study through Automatic Question Generation (AUTOQUEST), " October 1975.
  7. T. Hirashima, "Metadata editor of exercise problems for intel- ligent e-learning, " in Proceedings of the Workshop Notes of Applications of Semantic Web Technologies for E-Learning at (ISWC '04), pp. 99-101, Hirosjima, Japan, 2004.
  8. D. Gates, "Generating look-back strategy questions from expos- itory texts, " in Proceedings of the Workshop on the Question Generation Shared Task and Evaluation Challenge, Arlington, Va, USA, 2008.
  9. A. N. Kumar, "Generation of problems, answers, grade, and feedback-case study of a fully automated tutor, " ACM Journal on Educational Resources in Computing, vol. 5, no. 3, 2005.
  10. S. Ou, C. Orasan, D. Mekhaldi, and L. Hasler, "Automatic ques- tion pattern generation for ontology-based question answer- ing, " in Proceedings of the 21th International Florida Artificial Intelligence Research Society Conference (FLAIRS '08), pp. 183- 188, Menlo Park, Calif, USA, May 2008.
  11. E. Hovy, L. Gerber, U. Hermjakob, C. Lin, and D. Ravichandran, "Toward Semantics-Based Answer Pinpointing, " 2001.
  12. C. Fellbaum, WordNet, MIT Press, 1998.
  13. A. Corbett and J. Mostow, "Automating comprehension ques- tions: lessons from a reading tutor, " in Proceedings of the Work- shop on the Question Generation Shared Task and Evaluation Challenge, Arlington, Va, USA, 2008.
  14. R. Mitkov, L. A. Ha, and N. Karamanis, "A computer-aided environment for generating multiple-choice test items, " Natural Language Engineering, vol. 12, no. 2, pp. 177-194, 2006.
  15. N. Silveira, "Towards a framework for question generation, " in Proceedings of the Workshop on the Question Generation Shared Task and Evaluation Challenge, Arlington, Va, USA, 2008.
  16. L. Stanescu, C. S. Spahiu, A. Ion, and A. Spahiu, "Question generation for learning evaluation, " in Proceedings of the Inter- national Multiconference on Computer Science and Information Technology (IMCSIT '08), pp. 509-513, Wisla, Poland, October 2008.
  17. B. Cecila Alves, J. Mark Gierl, and Lai Hollis, "Using automated item generation to promote principled test design and devel- opment, " in Proceedings of the American Educational Research Association (AERA '10), Denver, Colo, USA, 2010.
  18. M. J. Gierl, J. Zhou, and C. Alves, "Developing a taxonomy of item model types to promote assessment engineering, " Journal of Technology, Learning, and Assessment, vol. 7, no. 2, pp. 1-50, 2008.
  19. E. Holohan, M. Melia, D. McMullen, and C. Pahl, "Adaptive E- learning content generation based on semantic web technol- ogy, " in Proceedings of the International Workshop on Applica- tions of Semantic Web Technologies for E-learning, Amsterdam, The Netherlands, 2005.
  20. M. Tosic and M. Cubric, "SeMCQ-Protégé plugin for auto- matic ontology-driven multiple choice question tests genera- tion, " in Proceedings of the 11th International Protégé Conference, Amsterdam, The Netherlands, 2009.
  21. H. Knublauch, R. Fergerson, N. Noy, and M. Musen, "The Protégé OWL plugin: an open development environment for semantic web applications, " in The Semantic Web-ISWC 2004, vol. 3298, pp. 229-243, Springer, Berlin, Germany, 2004.
  22. B. Žitko, S. Stankov, M. Rosić, and A. Grubišić, "Dynamic test generation over ontology-based knowledge representation in authoring shell, " Expert Systems with Applications, vol. 36, pp. 8185-8196, 2009.
  23. I. Leshcheva, D. Gorovaya, and D. Leshchev, "Ontology-based assessment technique, " in Proceedings of the 2nd International Workshop on Semantic Web Applications in Higher Education (SemHE '10), Southampton, UK, 2010.
  24. C. D. Bailey, J. N. Karcher, and B. Clevenger, "A comparison of the quality of multiple-choice questions from CPA exams and textbook test banks, " The Accounting Educators' Journal, vol. 10, no. 2, 1998.
  25. P. Buitelaar, P. Cimiano, P. Haase, and M. Sintek, "Towards linguistically grounded ontologies, " in Proceedings of the 6th European Semantic Web Conference on The Semantic Web: Research and Applications, pp. 111-125, Heraklion, Crete, Greece, 2009.
  26. L. R. Aiken, "Testing with multiple-choice items, " Journal of Research & Development in Education, vol. 20, no. 4, pp. 44-58, 1987.
  27. L. C. Jacobs and C. I. Chase, Developing and Using Tests Effectively: A Guide For Faculty, Jossey-Bass, San Francisco, Calif, USA, 1992.
  28. K. Woodford and P. Bancroft, "Multiple choice questions not considered harmful, " in Proceedings of the 7th Australasian Conference on Computing Education(ACE '05), A. Young and D. Tolhurst, Eds., vol. 42, pp. 109-116, Australian Computer Society, Darlinghurst, Australia, 2005.
  29. G. Francopoulo, M. George, N. Calzolari et al., "Lexical Markup Framework (LMF), " in Proceedings of the Language Resources and Evaluation Conference (LREC '06), Genoa, Italy, 2006.
  30. N. E. Gronlund, Measurement and Evaluation in Teaching, Macmillan, 1985.