Academia.eduAcademia.edu

Outline

A Similarity-based Theory of Controlling MCQ Difficulty

Abstract

—Several attempts have already been made to automate the generation of assessment questions. These attempts were mainly technical and lacked a theoretical backing. We explore psychological and educational theories to support the development of principled methods to generate questions and control their properties. We present a similarity-based theory to control the difficulty of multiple-choice questions and show its practicality and consistency with the psychological and educational theories.

FAQs

sparkles

AI

What key factors influence the difficulty of multiple-choice questions (MCQs)?add

The paper shows that difficulty correlates with the similarity between the key answer and distractors, with distractors requiring deeper knowledge for effective elimination.

How does knowledge structure affect question answering ability?add

Higher mastery students exhibit cohesive knowledge structures, enabling better discrimination among similar concepts, unlike low mastery students with fragmented knowledge.

What role do distractors play in MCQ quality and assessment accuracy?add

Effective distractors should be plausible yet recognizable by knowledgeable students, promoting accurate assessment of student mastery levels.

What are the implications of psychological theories on MCQ question design?add

Psychological theories, like spreading activation, affirm that knowledge organization influences how students retrieve information and solve questions effectively.

How can similarity measures enhance question generation from ontologies?add

Utilizing similarity measures such as Lin's supports the selection of quality distractors, improving the probability of valid assessment outcomes.

References (51)

  1. L. Aiken. Writing multiple-choice items to measure higher-order educational objectives. Educational and Psychological Measurement, 42(3):803-806, 1982.
  2. M. Al-Yahya. Ontoque: A question generation engine for educational assessment based on domain ontologies. In 11th IEEE International Conference on Advanced Learning Technologies, 2011.
  3. T. Alsubait, B. Parsia, and U. Sattler. Next generation of e-assessment: automatic generation of questions. International Journal of Technology Enhanced Learning, 4(3/4):156-171, 2012.
  4. J. R. Anderson. A spreding activation theory of memory. Journal of Verbal Learning and Verbal Behavior, 22:261-295, 1983.
  5. J. R. Anderson. Cognitive psychology and its implications. New York: W. H. Freeman, 4 edition, 1995.
  6. F. Baader, D. Calvanese, D. L. McGuinness, D. Nardi, and P. F. (eds.) Patel-Schneider. The Description Logic Handbook: Theory, Implemen- tation and Applications. Cambridge University Press, second edition edition, 2007.
  7. D. A. Balota. Automatic semantic activation and episodic memory encoding. Journal of Verbal Learning and Verbal Behavior, 22:88-104, 1983.
  8. D. A. Balota and R. F. Lorch. Depth of automatic spreading Mediated priming effects in pronunciation but not in lexical decision. Journal of Experimental Psychology: Learning, Memory, Cognition, 12:336-345, 1986.
  9. B. S. Bloom and D. R. Krathwohl. Taxonomy of educational objectives: The classification of educational goals by a committee of college and university examiners. Handbook 1. Cognitive domain. New York: Addison-Wesley, 1956.
  10. J. Brown, G. Firshkoff, and M. Eskenazi. Automatic question generation for vocabulary assessment. In Proceedings of HLT/EMNLP, pages 819- 826, Vancuver, Canada, 2005.
  11. J. Carneson, G. Delpierre, and K. Masters. Designing and Managing Multiple Choice Questions. 1996.
  12. M.T.H. Chi, R. Glaser, and E. Rees. Expertise in problem-solving. In R.J. Sternberg (Ed.), Advances in the psychology of human intelligence (Volume 1). Hillsdale, NJ: Erlbaum, 1982.
  13. M.T.H. Chi and R.D. Koeske. Network representation of a childs dinosaur knowledge. Developmental Psychology, 19:2939, 1983.
  14. G. Chung, D. Niemi, and W. L. Bewley. Assessment applications of ontologies. In Paper presented at the Annual Meeting of the American Educational Research Association, 2003.
  15. A. Collins and E. Loftus. A spreading-activation theory of semantic processing. Psychological review, 82(6):407-428, 1975.
  16. A. Collins and M. Quillian. Retrieval time from semantic memory. Journal of Verbal Learning and Verbal Behavior, 8:240-248, 1969.
  17. W. Crawford. Item difficulty as related to the complexity of intellectual processes. Journal of Educational Measurement, 5(2):103-107, 1968.
  18. M. Cubric and M. Tosic. Semcq protg plugin for automatic ontology- driven multiple choice question tests generation. In 11th Intl. Protg Conference, Poster and Demo Session, 2009.
  19. B. B. Davis. Tools for Teaching. San Francisco, CA: Jossey-Bass, 2001.
  20. C. dAmato, S. Staab, and N. Fanizzi. On the inuence of description logics ontologies on conceptual similarity. In EKAW '08 Proceedings of the 16th international conference on Knowledge Engineering: Practice and Patterns, 2008.
  21. H. Fisher-Hoch and S. Hughes. What makes mathematics exam questions difficult. British Educational Research Association, University of Lancaster, England, 1996.
  22. H. Fisher-Hoch, S. Hughes, and T. Bramley. What makes gcse exami- nation questions difficult? outcomes of manipulating difficulty of gcse questions. 1994.
  23. C. Gobbo and M. T. H. Chi. How knowledge is structured and used by expert and novice children. Cognitive Development, 1:221-237, 1986.
  24. GRESampleQuestions. Best sample questions. retrieved march 10, 2012, from http://www.bestsamplequestions.com/gre-questions/analogies/.
  25. L. Guttman. Image theory for the structure of quantitative variates. Psychometrica, 18(4):277-296, 1953.
  26. T.M. Haladyna and S.M. Downing. How many options is enough for a multiple choice test item? Educational & Psychological Measurement, 53(4):999-1010, 1993.
  27. M. Heilman. Automatic Factual Question Generation from Text. PhD thesis, Language Technologies Institute, School of Computer Science, Carnegie Mellon University, 2011.
  28. E. Holohan, M. Melia, D. McMullen, and C. Pahl. The generation of e-learning exercise problems from subject ontologies. In Proceedings of the Sixth IEEE International Conference on Advanced Learning Technologies, pages 967-969, 2006.
  29. A. Hoshino and H. Nakagawa. Real-time multiple choice question generation for language testing: a preliminary study. In Proceedings of the Second Workshop on Building Educational Applications using Natural Language Processing, pages 17-20, Ann Arbor, US, 2005.
  30. K. Janowicz. Sim-dl: Towards a semantic similarity measurement theory for the description logic alcnr in geographic information retrieval. In SeBGIS 2006, OTM Workshops 2006, pages 1681-1692, 2006.
  31. J. Kehoe. Basic item analysis for multiple-choice tests. Practical Assessment, Research & Evaluation, 4(10), 1995.
  32. K. Lehmann and A. Turhan. A framework for semantic-based similarity measures for elh-concepts. JELIA 2012, pages 307-319, 2012.
  33. D. Lin. An information-theoretic definition of similarity. In In: Proc. of the 15th International Conference on Machine Learning, page 296304, San Francisco, CA, 1998. Morgan Kaufmann.
  34. C. Liu, C. Wang, Z. Gao, and S. Huang. Applications of lexical information for algorithmically composing multiple-choice cloze items. In Proceedings of the Second Workshop on Building Educational Ap- plications using Natural Language Processing, pages 1-8, Ann Arbor, US, 2005.
  35. E. F. Loftus. Activation of semantic memory. American Journal of Psychology, 86:331-337, 1974.
  36. J. Lowman. Mastering the Techniques of Teaching (2nd ed.). San Francisco: Jossey-Bass, 1995.
  37. M. Miller, R. Linn, and N. Gronlund. Measurement and Assessment in Teaching, Tenth Edition. Pearson, 2008.
  38. R. Mitkov, L. An Ha, and N. Karamani. A computer-aided environment for generating multiple-choice test items.cambridge university press. Natural Language Engineering, 12(2):177-194, 2006.
  39. R. Mitkov and L.A. Ha. Computer-aided generation of multiple-choice tests. In Proceedings of the First Workshop on Building Educational Ap- plications using Natural Language Processing, pages 17-22, Edmonton, Canada, 2003.
  40. R. Mitkov, L.A. Ha, A. Varga, and L. Rello. Semantic similarity of distractors in multiple-choice tests: Extrinsic evaluation. In Proceedings of the Workshop on Geometrical Models of Natural Language Semantics, pages 49-56, Athens, Greece, March 2009. Association for Computa- tional Linguistics.
  41. A. Papasalouros, K. Kotis, and K. Kanaris. Automatic generation of multiple-choice questions from domain ontologies. In IADIS e-Learning 2008 conference, Amsterdam, 2008.
  42. M. Paxton. A linguistic perspective on multiple choice questioning. Assessment & Evaluation in Higher Education, 25(2):109-119, 2001.
  43. R. Rada, H. Mili, E. Bicknell, and M. Blettner. Development and application of a metric on semantic nets. In In: IEEE Transaction on Systems, Man, and Cybernetics, volume 19, page 1730, 1989.
  44. P. Resnik. Using information content to evaluate semantic similarity in a taxonomy. In In Proceedings of the 14th international joint conference on Artificial intelligence (IJCAI95), volume 1, pages 448-453, 1995.
  45. G. M. Seddon. The properties of Bloom's taxonomy of educational objectives for the cognitive domain. Review of Educational Research, 48(2):pp. 303-323, 1978.
  46. F. Sharifian and R. Samani. Hierarchical spreading of activation. In In F. Sharifian (Ed.) Proc. of the Conference on Language, Cognition, and Interpretation, pages 1-10, 1997.
  47. J. T. Sidick, G. V. Barrett, and D. Doverspike. Three-alternative multiple- choice tests: An attractive option. Personnel Psychology, 47:829-835, 1994.
  48. R. Smith. An empirical investigation of complexity and process in multiple-choice items. Journal of Educational Measurement, 7(1):33- 41, 1970.
  49. J. W.Pellegrino, N. Chudowsky, and R. Glaser, editors. Knowing what Students Know: The Science and Design of Educational Assessment. NATIONAL ACADEMY PRESS, Washington, DC, 2001.
  50. Z. Wu and MS. Palmer. Verb semantics and lexical selection. In In: Proceedings of the 32nd. Annual Meeting of the Association for Computational Linguistics (ACL 1994), page 133138, 1994.
  51. B. Zitko, S. Stankov, M. Rosic, and A. Grubisic. Dynamic test generation over ontology-based knowledge representation in authoring shell. Expert Systems with Applications: An International Journal, 36(4):8185-8196, 2008.