Welcome to the testmaster site for
Student Assessment Research

and discussion


Home
testmaster
My Vision
Published Work
TDLB Awards
3D Assessment
MCQ's
Control
Higher Level Thinking
Security Issues
Action at Last?
My Letter to Gov
Gov. Action?
Comment

Contact Me?

Multi Choice Question Formats

Contents of this Section

NOTE: I would like to draw the reader's attention to the significant work of Robert Wood - Assessment and Testing: A survey of research, Published 1993 - University of Cambridge Local Examinations Syndicate, and from which much of the argument and detail of this work is drawn.



Discussion.

Some writers claim that MCQ's are unsuitable for testing higher level abilities, although Vernon (Vernon P.E. 1964) stated that it is by no means certain that conventional examinations can test higher level thinking either. 

It has been claimed that Multiple Choice Questions (MCQ's) promote convergent thinking ie. that there is only one correct answer to a question, and that answer is known. This criticism is more substantial than the first, especially if MCQ's are used to test higher order cognitive skills (see Frederiksen 1984). 

I believe that open ended questions may discourage an answer, not because of any shortage of knowledge or understanding, but because of the difficulty involved in the construction of an answer. Mathews (Mathews 1967) expressed similar opinions regarding open-ended questions and, it is the validity of open-ended questions as an assessment technique is being questioned this time. 

The essay question is often the favoured 'traditional' examination item at all levels. For those who advocate essay questions above all other, I would direct them to a comprehensive discussion of essay questions, and the problems they raise regarding the nature of extended writing, it's constructs (content, organisation and style etc.), the construction of questions, marking issues, marker error, analytic vs. holistic (impression) marking and the influence of handwriting given by Wood (Wood R.(1) 1993) 

The fact is however, that much of the time we are testing basic knowledge and understanding, and regardless of the criticism of MCQ's, they are a valuable assessment tool. MCQ's involve thinking, and little writing. They encourage the candidate to answer. 

Wood (Wood R.(2) 1993) claims that it is now possible to be prescriptive about a number of issues concerning the construction of MCQ test items. These issues include:- 

 

1. Number of alternatives.

The practical difficulties involved in writing multiple distractors, and the results of research by a variety of authors, suggests that a four choice, three distractor format is amongst the most effective.

 

2. 'None of These' responses.

The use of this as a response tends to increase the level of complexity without , increasing reliability or validity.

 

3. Scoring Systems.

 Students who do not know an answer will often guess. 

Guessing is the main concern by those who use MCQ's to test students, as the guessed answer does not guarantee knowledge of the subject. On balance, research suggests that to instruct students not to guess answers is inadvisable. Linn (Linn M.C. 1987), showed that females are more reluctant to guess than males, and Grandy (Grandy J. 1987) showed that females are more likely to leave items blank. 

Akeroyd's (Akeroyd F.M. 1982) "Dual Response" system may be useful in rewarding partial information. For a four-choice test item, candidates score: 

  • 1 mark for identifying the correct answer with 1 response;
  • .5 marks for 2 responses, 1 of which is correct;
  • .25 marks for an omission ie. no response;
  • 0 mark for an incorrect answer OR combination of 2 incorrect responses.

The .25 mark for omissions is said to eliminate fluctuations of scores due to random guessing.

Akeroyd claims that this system is a good way of learning about student response behaviour, but appears to be untried to date. An adaptation of this, for computer aided assessment use, may be possible.

 

4. Answer until correct.

This gives immediate feedback to the student. The scoring is dependent upon the number of alternatives ie. if there are three, then a correct selection first time scores 2, second time scores 1 and third scores 0.

 

5. Changing Answers.

 If a student is capable of recognising an incorrect answer after reflection, then they should be able to change it. If the student does not know the answer in the first place, they are unlikely to attempt to change the answer given.

 

6. Confidence Weighting.

This system involves asking students to award a weighting to their answer depending upon their confidence in that answer being correct.

Research in this technique has failed to establish its ability to provide useful data (Wood R. (3)1990).

 

7. Differential Weighting.

This system involves weighting responses in a variety of ways. Research in this area has failed to identify a method that can offer reliable differentiation across all levels of response. Echternacht (Echternacht G.J. 1976) found that his specially constructed, subjectively weighted test items were even less accurate than number-right scores.

  

The Cloze Testing Technique

These tests are often "grouped together" with multi choice tests as they appear similar. Cloze tests involve the student inputting a word in order to complete a meaningful sentence. Cloze tests may offer a selection of correct words, correct words and additional incorrect "distractors" (cloze-elide) for the student to position correctly within the sentence(s), or words may not be provided at all. 

Criticism of cloze is that it is "unauthentic", and Carroll (Carroll et.al. 1959) found that:- 

"the ability to restore texts is somewhat independent of competence in a language at it is ordinarily defined. That is to say, we observed many people who are perfectly competent and literate in a language but who do not show facility in a special task of guessing what a missing letter or word in a text might be."

  • Generally, the paper based cloze test is not popular among students or teachers.

    Bibliography 

    Akeroyd F.M.

     

    Progress in multiple choice scoring methods, 1977/81. 1982 Journal of Further and Higher Education, 6, 87-90
    Carroll J.B.

    Carton A.S.

    Wild C.

    Write Your Own Language Tests. 1959 Oxford: Pergamon Press
    Echternacht G.J. Reliability and validity of item option weighting schemes 1976 Educational and Psychological Measurement 36, 301-10
    Frederiksen N. The real test bias: Influences of testing on teaching and learning. 1984 American Psychology 39 193-202
    Grandy J. Characteristics of examinees who leave questions unanswered on the GRE general test under rights-only scoring. 1987 ETS Research Report, 87-38 Princeton, NJ: Educational Testing service
    Linn M.C. Gender differences in National Attainment of Educational Progress science items: What does 'I don't know' really mean? 1987 Journal of Research in Science Teaching 24. 267-78
    Mathews J.C. The Nuffield Foundation Sciences Teaching Project VII: o-Level chemistry examinations. 1967 School Science Review, 49, 21-30
    Vernon P.E. The Certificate of Secondary Education: An Introduction to Objective-type Examinations. 1964 Examinations Bulletin No 4 London Secondary Schools Examination Council.
    Wood R. Assessment and Testing: A survey of research.

    Ch.5 (1)

    Ch 4 (2)

    Ch 4 (3)

    1993 University of Cambridge Local Examinations Syndicate
  •  

    1