Elsevier

Journal of Web Semantics

Volume 34, October 2015, Pages 40-54
Journal of Web Semantics

A novel approach to generate MCQs from domain ontology: Considering DL semantics and open-world assumption

https://doi.org/10.1016/j.websem.2015.05.005Get rights and content

Abstract

Ontologies are structures, used for knowledge representation, which model domain knowledge in the form of concepts, roles, instances and their relationships. This knowledge can be exploited by an assessment system in the form of multiple choice questions (MCQs). The existing approaches, which use ontologies expressed in the Web Ontology Language (OWL) for MCQ generation, are limited to simple concept related questions — “What is C?” or “Which of the following is an example of C?” (where C is a concept symbol) — or analogy type questions involving roles. There are no efforts in the literature which make use of the terminological axioms in the ontology such as existential, universal and cardinality restrictions on concepts and roles for MCQ generation. Also, there are no systematic methods for generating incorrect answers (distractors) from ontologies. Distractor generation process has to be given much importance, since the generated distractors determine the quality and hardness of an MCQ. We propose two new MCQ generation approaches, which generate MCQs that are very useful and realistic in conducting assessment tests, and the corresponding distractor generating techniques. Our distractor generation techniques, unlike other methods, consider the open-world assumption, so that the generated MCQs will always be valid (falsity of distractors is ensured). Furthermore, we present a measure to determine the difficulty level (a value between 0 and 1) of the generated MCQs. The proposed system is implemented, and experiments on specific ontologies have shown the effectiveness of the approaches. We also did an empirical study by generating question items from a real-world ontology and validated our results with the help of domain experts.

Introduction

Automated assessment systems serve as a method to measure the level of learning as well as to provide a faster solution for large scale assessments. Many tests like TOEFL, IELTS, GRE ​and GMAT are dependent on online assessment systems to make the assessment task easier. Such systems mainly use multiple choice questions rather than subjective questions for conducting the test.

Using Multiple Choice Questions (MCQs) for assessments has both merits and demerits. They are preferred for assessing broad range of knowledge. This is mainly because they require less administrative overhead and provide instant feedback to test takers. However, studies by Barbara Gross  [1] and Sidick et al.  [2] show that, developing effective objective type questions is time consuming and requires domain expertise to generate good quality MCQs. So, there is a need for an automated method for MCQ generation from a given knowledge source.

Recently, a handful of studies [3], [4], [5], [6], [7], [8], [9] explored the use of structured domain knowledge in the form of description logic based ontologies to automatically generate MCQs. This would enable online assessment systems to utilize existing knowledge bases for the assessment of learner’s knowledge and skills. But, there are challenges involved in generating MCQs from these ontologies. Some of the challenges that the existing approaches tried to address are: (i) How to frame interesting and good quality questions from ontologies? (ii) How to generate proper incorrect answers (distractors) for the framed question? (iii) How to control the difficulty level of the generated questions? Although the previous efforts were not in vain, there are substantial shortcomings in fully exploiting the formalized knowledge in an ontology for MCQ generation. In this paper, we show that, with a better understanding of the semantics of a given ontology (expressed in Web Ontology Language), the three challenges can be addressed more elegantly.

Challenge 1. Framing interesting and good quality questions. In the literature, the approaches that use ontologies have the limitation that they generate simple concept related questions–“What is C?” or “Which of the following is an example of C?” (where C is a concept symbol)–or analogy type questions involving roles. These questions are very basic  [10] and do not contain any domain related specifics. In other words, the approaches which generate such questions, do not appropriately make use of the axiomatized knowledge in an ontology. Furthermore, restrictions (existential, universal and cardinality) on concepts and roles in ontologies are not utilized properly for question generation in any of the current approaches.

Consider a movie ontology with statements, With respect to this, we can frame a question about the instance braveHeart: “Choose a movie directed by an actor?” Our approach in this paper is an effort in this direction.

Challenge 2. Proper distractor generation. Under the closed-world assumption (CWA), we can choose any instance which is different from the instance braveHeart as a distractor for the question in the example above. But, Web Ontology Language (OWL) adheres to the open-world assumption (OWA): statements which are not logical consequences of a given knowledge base are not necessarily considered false. Therefore, not all distractors which are generated under CWA can be guaranteed as true distractors.

We observed that most of the existing MCQ generation techniques [7] randomly select instances which do not belong to the class of the correct answer as distractors. The incorrectness of the distractors cannot be ensured by this random selection method, which in turn made it necessary to manually check the correctness of the question items before making use of them. We address this issue by proposing a systematic method to generate only those question items which are valid under OWA.

Challenge 3. Control the difficulty level of the generated MCQ. MCQs of varying difficulty level are necessary to assess the depth of knowledge of a learner (student). We introduce a measure to find out the difficulty level of the generated MCQs based on the similarity-based theory suggested by Alsubait et al.  [11].

In this paper, we propose two approaches (i) node-label-set based approach (ii) edge-label-set based approach to generate (two) interesting types of MCQs. We adopt description logic specifications of the ontology to generate the so called label-sets (node-label-sets and edge-label-sets). A measure to estimate the difficulty level of generated MCQs is also proposed by means of these label-sets. We study the feasibility of our approaches by implementing them and generating MCQs from some sample ontologies. In  Appendix A, we list some of the MCQs, which are generated from Geographical Entity ontology. To validate our new approaches and difficulty measure, we generated question items from a real-world ontology and got them evaluated by domain experts. Statistics of our empirical evaluation validate our arguments and are detailed in Section  6. The new notations and abbreviations that we introduced in this paper are listed in  Appendix B along with their meaning.

Section snippets

Related work

Papasalouros et al.  [5] suggested 11 strategies based on classes, properties and terminologies of ontologies for framing MCQs and the corresponding distracting answers. Their MCQ generation methods lack proper theoretical support for when to use which strategy, and the stem of all the generated questions remains the same (“Choose the correct sentence”).

Cubric and Tosic  [4] and M. Cubric  [6] generated MCQs of knowledge level (“Which of the following definition describes the concept C?”),

Preliminaries

In this section, we describe: MCQ, the Description Logic (DL) SHIQ  based ontologies (SHIQ  ontologies) and an example ontology (Harry-Potter-Book ontology).

Proposed MCQ generation approaches

Once we get the label-sets of all the instances (node-label-sets) and all the pairs of instances (edge-label-sets) in a given ontology, we can generate MCQs based on the following two approaches.

Experiments

In order to support the approaches presented in this paper, we implemented our techniques in Java 1.7 using the Jena framework (2.11.0) as a portable library. We use the Jena Ontology API8 to explore concepts and to look up restrictions in the ontologies.

We consider four ontologies for our experimentation.

  • 1.

    Harry-Potter-Book ontology9 is developed by our

Empirical evaluation

In this section, we describe the empirical evaluation done to (1) check the usefulness of the question items which are generated from our two MCQ generation approaches and (2) verify the effectiveness of the proposed measure for estimating question difficulty.

Conclusion and future work

We proposed the generation techniques of two new categories of multiple choice question types based on the node-label-sets and edge-label-sets of the instances in an ontology. We introduced a technique called Label-set-Reduction to make the label-sets suitable for generating MCQs by converting them to a reduced form (called Reduced-node-label-sets). We suggested a systematic method to find the distractors for the generated MCQs. The similarity of these distractors with the correct answer is

Acknowledgments

This research has been funded by IIT Madras and the Ministry of Human Resource Development, Government of India. We express our gratitude to the participants of our evaluation process: Dr. S. Gnanasambabdan (Director of Plant Protection, Quarantine & Storage), Ministry of Agriculture, Gov. of India; Dr. S. Nazreen Hassan (Asst. Professor), Mr. J. Delince and Mr. J.M. Samraj, Department of Social Sciences AC & RI, Killikulam, Tamil Nadu, India; Ms. Deepthi. S (Deputy Manager), Vegetable and

References (22)

  • M. Schmidt-Schauß et al.

    Attributive concept descriptions with complements

    Artificial Intelligence

    (1991)
  • D. Barbara~Gross

    Tools for Teaching

    (1993)
  • J.T. Sidick et al.

    Three-alternative multiple-choice tests: An attractive option

    Personnel Psychol.

    (1994)
  • T. Alsubait, B. Parsia, U. Sattler, Mining ontologies for analogy questions: A similarity-based approach, in: Vol. 849...
  • M. Cubric et al.

    Towards automatic generation of e-assessment using semantic web technologies

  • A. Papasalouros et al.

    Automatic Generation of Multiple Choice Questions from Domain Ontologies

    (2008)
  • M. Tosic, M. Cubric, SeMCQ- protege plugin for automatic ontology- driven multiple choice question tests generation,...
  • M. Al-Yahya

    Ontology-based multiple choice question generation

    Scient. World J.

    (2014)
  • K. Zoumpatianos et al.

    Automated transformation of swrl rules into multiple-choice questions

  • B. A˚itko et al.

    Dynamic test generation over ontology-based knowledge representation in authoring shell

    Syst. Appl.

    (2009)
  • Cited by (36)

    • AI-based learning content generation and learning pathway augmentation to increase learner engagement

      2023, Computers and Education: Artificial Intelligence
      Citation Excerpt :

      Another semi-automatic approach based on Linked Open Data is proposed by Bühmann et al. (2015) which uses entity summarisation techniques for content selection and RDF verbalization for question generation. Vinu et al. (2015) propose a method that uses terminological axioms in the ontology such as existential, universal and cardinality restrictions on concepts and roles for MCQ generation. Apart from the fact that the aforementioned ontology based approaches are semi-automatic, one more disadvantage of using ontologies for automatic question generation is that creating ontologies requires significant human efforts for the domains which are not defined in the available open ontologies such as DBPedia.

    • Assessment Methods and Tools for Pharmacy Education

      2018, Pharmacy Education in the Twenty First Century and Beyond: Global Achievements and Challenges
    • Generating Multiple Choice Questions for Computing Courses Using Large Language Models

      2023, Proceedings - Frontiers in Education Conference, FIE
    View all citing articles on Scopus
    View full text