Processing Test Marks
Eric Scharf



This "web-paper" presents the mathematical basis for the processing of the marks for tests or examinations comprising multiple choice questions. The gathering of marks as well as the actual marking of multiple choice tests can be automated. However, the marking process normally requires a viable model that relates the number of questions a student gets right to the number of questions to which the student actually knows the answer. The implications of using popular marking schemes are examined; it is shown - using simple algebra - that such schemes can differ widely in terms of the actual overall mark that a student receives for a multiple choice test. Without trying to be unduly prescriptive, I would suggest that the averaging marking scheme, using the number of questions actually attempted, together with 3 or 4 choices per question, is likely to be the fairest approach in most educational cases.

What you will see here

  1. Introduction, Challenges and Background
  2. Logistics for this Paper
  3. Basic Assumptions
  4. Processing Options
  5. Raw Data - No Further Processing
  6. Average or Statistical Approach
  7. Putting our equations to the test
  8. Subtraction Approach
  9. Future Challenges
  10. Conclusions
  11. Reference & Links
  12. Symbols and Abbreviations

1. Introduction, Challenges and Background

It is helpful to the reader if we say at the outset what a multiple choice test (MCT) actually is, and indeed, what the advantages and disadvantages are of deploying this kind of test in an educational context. Essentially, we shall start to steer towards the rationale for this "web paper".

1.1 Multiple Choice Tests

A multiple choice test (MCT) consists of a series of multiple choice questions (MCQs). We shall assume in this web paper that each multiple choice question (MCQ) consists of a number of possible answers or choices, only one of which is actually the right answer.

The following is an example of a typical multiple choice question.

Which one of the four places given below is the capital of Austria?
  1. Vaduz
  2. Salzburg
  3. Vienna
  4. Munich

1.2 Uses, Advantages and Challenges of Multiple Choice Tests

Multiple Choice Tests have two main applications, namely: (1) Self learning (auto didactic) and (2) Assessment under examination conditions. For both applications, the speed of marking by the computer can be of tremendous benefit, in that it allows almost immediate feedback to the student as regards his or her knowledge of the subject. The concept of "raw data" will be explained in detail in section 5 below.

1.3 Why this paper?

In the humble experience of the author, the implications of using specific processing options for the raw marks from a MCT are not always understood. The nature of the transformation ("black box") between the raw MCT marks and the final official marks for a student is of course crucial. The different types of transformations used can result in a wide variation in the target marks for a student. The author of this paper therefore believes that clarification regarding the possible options is long overdue!

Indeed, the initial experiences of the author, regarding the lack of understanding about multiple choice tests (MCTs) in the educational sector, led to a paper [ref1] in which the author was the lead contributor. This paper provides an important initial basis for the present web paper.

Further investigations by the author revealed an earlier and seemingly forgotten article [ref3] dating from 1988. This latter paper refers to a "Formula Scoring" method, which in fact is the equivalent of the "Average Strategy" to be developed below. Unfortunately, the paper [ref3] does not give a clear presentation of the "Formula Scoring" method and the implications of its deployment for multiple choice tests (MCTs); it is hoped that the present paper can go some way to provide firm mathematical clarification of these issues. Other processing approaches for MCTs include those used by the exams of Australian Mathematics Competition [ref4] which have what appears to be an ad hoc approach to processing the output marks from the MCTs.

Both the author's initial experiences and subsequent investigations led to the present web paper, which, it is hoped may help to clarify the issues surrounding the processing of the raw marks given by MCTs.

2. Logistics for this Paper

This paper - as mentioned - derives in part from a previous refereed publication [ref1] by the author. This present web paper, however, should be viewed as living document, whose contents may be amended from time to time to accomodate the latest information.

In this paper I shall repeat myself in various places with the aim of emphasising and explaining relevant points. In this respect I beg to differ with those academics and others who express a devotion to the hapax legomenon (or ‘απαξ λεγομενον if you wish), for it is by repetition that we learn so many things including the one (or more!) languages that we speak and use in daily life.

I have found that the popular browsers listed in the browser table on the "read me" page display the contents of this page - including the equations - correctly when executing in the machine and operating system environment which is also given on the "read me" page. However, I cannot, of course, guarantee that this will be the case for all browser, machine and operating system environments.

For future reference, if you reached this page by another route, you can also access this page with the following shortcut:

3. Basic Assumptions

In the mathematical developments in sections 3 and 4, we shall make the following assumptions, which should not normally affect the generality of the arguments.

Two further points should be made at this juncture.

4. Processing Options for Multiple Choice Tests

We now look at three main options. These may not be the only options, but are frequently used. They are ways of dealing with wrong ("none right") answers in order to get at the number K of answers that a student actually knows as opposed to the number R of answers a student gets right. Note that the term "wrong" is used to refer to the questions that a student does not get right as opposed to the questions for which a student does not know the answer.

  1. The raw data option ignores wrong answers.
  2. The averaging or statistical option, to obtain K, subtracts from R a term based on the number of wrong ("none-right") questions and the assumed statistical information about the occurence of wrong questions.
  3. The subtractive option, to obtain K, subtracts from R a given proportion or factor (e.g. 1) of wrong ("none-right") questions.

The averaging (or statistical) approach and the subtractive approach can both be subdivided into two further approaches taking into account the number

Hence, the three main processing options have so far given us FIVE possible candidates for the processing of the marks produced by multiple choice tests. We shall look at each candidate in turn.

5. Raw Data - No Further Processing

Here the output marks produced directly by the MCT are regarded as raw data and are not subject to further processing apart from scaling. We could of course assume that the same number (M) of marks (where M is 1 or more) is allocated to each question for a right answer, and scaling is merely a matter of multiplying the number (R) of right questions by M. In the context of our mathematical analysis, it is easiest here to assume, without loss of generality, 1 mark per right question.

Here we repeat for clarity the following distinctions which were alluded to in section 1.2 above.

In this "Raw Data" case, students can gain marks by guessing (i.e. by luck), and, on average, 1/A of the guesses will be correct, where of course we assume that there are A possible choices for a question, only one choice of which is right. This will be so, even if the students have absolutely no knowledge of the actual topic being examined, since wrong results are not penalised.

In other words, this simple approach does not attempt to distinguish between answers resulting from genuine knowledge, and those resulting from lucky guesses. This scenario is usually different from a conventional written paper, where to gain any marks at all, some correct response – written or pictorial – is normally required of the student. In other words, a written paper is more likely to elicit a student's thought processes than a multiple choice test. The present web paper is a testimony to the fact that I certainly do not give up entirely on the merits of multiple choice tests, but that, at the same time, I would advocate great caution, for the devil can reside in the detail!

Remember also, in the context of "Raw Marks", that the computer conveniently gathers the marks for us; what we do – or ask the computer to do - with those marks is up to us as the assessors and could have a significant effect on a student’s marks in the MCT. This should be a significant message of the subsequent sections of this "web paper".

A typical exam rubric for this version of an MCT could include the following statements. “There is no penalty for a wrong answer. You are therefore urged to attempt every question. If you feel you do not know the answer to a question you should select the answer that you feel is the most likely.”

However, the raw data approach is well suited to self learning, in other words the auto-didactic context, where the student is examining himself or herself.

6. Average or Statistical Approach

How can we devise an approach that aims to determine the number of questions that a student actually knows as opposed to the number of questions that a student gets right?

6.1 Deriving a Formula

The total number (R) of questions that a student gets right, is the sum of the number (K) of questions that the student actually knows, plus the number (C) of questions that the student gets right by chance.


We are after determining the number (K) of questions the answers to which a student knows, and not the number (R) of questions that a student happens to get right and which is what an MCT gives us in the first instance as raw data. In other words we need to consider:

K = R − C [2]

Of course, to increase our understanding of C, we could ask what causes a student to get by chance the right answers to - possibly - just some of the questions attempted. Put the opposite way, what causes a student to get the wrong answer to any particular question attempted? In this case, for example, students:

  1. typically might take a guess at the questions they do not know, in the hope of producing a right answer.
  2. may mistakenly think they know the answer.
  3. may by accident – perhaps by being distracted – put down the wrong answer.

We now consider the general situation, where the assessor (marker) has absolutely no knowledge (not even the three items just considered above) about a student’s motive for getting C questions right by chance. We try to find a plausible model for C.

This can be done by assuming that all questions carry equal weight and that for each question there are "A" possibilities, typically 3 or 4, of which only one possibility is right. In addition, we let Q be the total number of questions attempted. In this case, the number (C) of questions that are right by chance can be obtained by saying that the proportion (1/A) of the (Q-K) questions that a student does not know but has attempted, will - on average - contribute to those questions that appear as being right in the test. This gives:

C   =   Q − K

Equation [3] is our basic model for relating C to what we wish to find, namely K, the number of questions a student actually knows. While there may be other candidate options to determine C, such as for example a root mean square criterion, the model in equation [3] is attractively simple and - as we shall see - useable.

We can now use elementary algebra to combine equations [1] and [3] to give us:

R   =   K   +   Q − K

or, rearranging,

K   =   R   −   Q − R
A − 1

Equation [5] shows that the (Q-R) wrong results will penalise the student. Essentially, in equation [5], in order to obtain K we reduce the number R of right questions, which the test gives us, by a penalty factor based on the number (Q-R) of wrong answers divided by the number (A-1) of wrong possibilities for each question. Equation [5] thus reflects the fact that the more possibilities (i.e. the higher the number A of choices per question) there are, the more difficult it is to guess correctly the right answer. (Note, as an aside, that in this equation, negative values of K can be forced to a floor value of zero, which is then the minimum mark a student can achieve.)

6.2 Deriving the Averaging Formula by Another Way

We have said that the zero penalty option is not ideally suited to an examination context, since it will, on average, give a student Q/A marks - even if that student has no knowledge whatsoever of the exam subject. So we can subtract this from the number R of right answers which the computer has given us. We then need to rescale the resultant number by (A/(A-1)) to get back to our original scale of 1 to Q marks. This gives equation [6] below.

K   =   (   R   −   Q
 ) (  A
A − 1

After simplifying equation [6], we obtain equation [7] which is actually equation [5] above!

K   =   R   −   Q − R
A − 1

6.3 Comments on Averaging Approach

With this average or statistical approach, a student attempting a given number X of questions, all of which are answered correctly, may get more marks than his colleague who attempts Y questions, where Y > X, but where still only X questions are answered correctly. This is fair because the “Y” student has used more chances. This is illustrated in table 1 in section 5.7 below.

When using this approach, a possible rubric for an examination paper could read: “You will only be assessed on questions which you have attempted and to which you have provided a legible answer. It is recommended that you avoid guessing an answer to a question.”

6.4 Averaging with Maximum Penalty

Here we consider all questions, even those that are not attempted. In this case, we substitute for Q, the total number of questions attempted, the value T, representing the total number of questions offered. In all cases other than the case where the students gets all offered equations right, this variant with T will penalise a student to a greater extent than staying with Q.

With the Maximum Penalty strategy, the number of questions that a student has not answered are added to the number of questions a student has answered wrongly. Effectively, we take equation [5] and replace Q, the total number of questions attempted, by T, the total number of questions actually set. This gives the following equation, which is the limiting case of equation [9] when all the questions have been attempted. The subscript "AT" refers to the combination of the averaging approach with the consideration of the total number T of questions offered.

KAT   =   R   −   T − R
A − 1

Another form of the Maximum Penalty Equation can be derived, by assuming M marks per question, and putting S = KAT M and SMAX = T M. This gives the following equation:

  =   S − (SMAX  / A)
SMAX − (SMAX  / A)

Equations 8 and 9 are given for completeness, but their presence here should not detract from the main mathematical or algebraic development in this web paper. Hence the symbols used in equations 8 and 9, unless used elsewhere in this paper, on purpose have been excluded from the table of symbols and abbreviations in section 11.

7. Putting our equations to the test

In this section we shall try and "breathe life" into the theoretical results we derived above by seeking answers to the following questions.

  1. How does the method of processing an MCT influence a student's mark?
  2. Does a student benefit by guessing?
  3. Does the effect of guessing depend on A, the number of choices per question?
  4. How does the effort in designing each question relate to A?
  5. How is the number A of choices per question related to the usable range of Raw Marks?
  6. Is the raw data line also the limiting case as A tends to infinity?

7.1 The effect of processing an MCT on a student's mark

The table in figure 7.1.1 considers a scenario of 20 questions, with 4 choices per question (or A = 4), one choice of which is correct. For convenience of presenting the numerical results in this paper, the number of correct questions has been scaled by 5 to give a percentage mark. Both columns 1 and 2 can thus be regarded effectively as the same raw data but with different scaling. Note that usually one tries to avoid negative marks, so that “K”, the number of questions whose answer a student actually knows, has not been allowed to go negative.

Fig. 7.1.1: Tabular Comparison of Raw Data and Averaging Marking Schemes

The table represents three options.

Figure 7.1.2, below, represents the results for different numbers, Q, of questions attempted. It shows that the operation area for MCTs is bounded by the triangular area which is formed by the zero penalty line (raw data), the maximum penalty line (Q = T = 20) and the x–axis. As stated above, negative results are floored to zero.

Fig. 7.1.2: Graphical Comparison of Raw Data and Averaging Marking Schemes

The graph in figure 7.1.3 is derived from the graph in figure 7.1.2 above, and shows the differences in the percentage output mark (K) (obtained by reading in the vertical direction) between the zero penalty approach and intermediate penalty approaches where the number Q of questions attempted is 4, 8, 12, and 16 respectively, with the limiting case of Q = T = 20 being equivalent to the maximum penalty approach. For Q values of 4, 8, 12, 16 and 20, the corresponding maximum differences are thus 5, 10, 15, 20 and 25% respectively. Since the maximum difference can be as much as 25%, figure 7.1.3 emphasises that the mark that a student obtains for a MCT is strongly influenced by the way in which the educator processes the raw marks generated by the MCT.

Fig. 7.1.3: Difference between Raw Data and Averaging Marking Schemes

7.2 Does a student benefit by guessing?

If there is no processing of the marks, in other words only the raw data are used as described in section 4, the answer is definitely yes for “blind” guessing. However, with the averaging approach of section 5, the answer is no. This is borne out by equation [5] and the table in figure 7.1.1. If Q, the total number of questions attempted, increases, but R, the actual score, stays the same, then the mark K actually given to the student will decrease. This is emphasised by figure 7.2.1. This shows the number of questions attempted against the resultant mark K for different raw scores (R) where R is the number of questions answered correctly, irrespective of whether this is due to the student’s actual knowledge or to guessing. (Of course, in figure 7.2.1 we could also put odd values for R, but perhaps the message is put across to you the reader in a better way with a less complicated graph.) In figure 7.2.1, the downward sloping lines for given values of R indicate diminishing returns for the student if wrong answers, denoted by (Q-R), are given for further questions attempted.

Fig. 7.2.1: Number (Q) of Questions Attempted against Final Score (K) for different numbers (R) of questions for which the correct answer is given (either from knowledge or by chance).

The effect of partial knowledge is also interesting. Suppose a student has additional partial knowledge that improves the chances of an answer being correct from 1/A to 1/P (i.e. P < A). For example, a student may be able to narrow down his or her options from 1 in 4 choices (A = 4) to 1 in 2 choices (P = 2). In this case, in equation [5], the subtracted term ((Q-R)/(A-1)) will be less in value if A is replaced by P where P < A. With partial knowledge, the negative impact on the final mark will, on average, be lessened. Naturally, this is only a very crude model of the situation since students will not normally approach each question with the same degree of partial knowledge.

7.3 Does the effect of guessing depend on A?

Does the effect of guessing depend on the number (A) of possible choices per question? Intuitively we could say that the more choices per question, the less the chance that a student hits on the right answer by blind guessing. In fact we can say that the effect of guessing is proportional to (1/A), so that as A increases, the possibility of a student picking the right answer for a given question from a set of A possible answers diminishes. If we plot the probability as a percentage (instead of as a per unit) we get the graph shown in fgure 7.3.1.

A 2345 678910
 probablity 1/A (%)   50.0  33.3   25.0  20.0   16.7  14.3   12.5  11.1   10.0 

Fig. 7.3.1: Graph and Table relating the number A of choices per question to the probability 1/A of hitting the right answer to a question by chance (expressed as a percentage).

We have thus quantified our original statement that "the more choices per question, the less is the chance that a student hits on the right answer by blind guessing". Not only that, but figure 7.3.1 also shows us quantitatively that increasing the number of choices per question "indefinitely" is likely to bring the examiner diminishing returns when it comes to insuring against blind guessing by the student.

This and the next paragraph of the present section 7.3 are not required for an understanding of the material to be developed in later sections. However, as an exercise in manipulating the knowledge we have already gained in previous sections, we can also relate quantitatively the effect of guessing to an equation which we have indeed already established. We realize that figure 7.1.2 has been created by assuming for each question, four choices (i.e. A = 4), one of which is correct. In this figure we can see that the maximum difference DMAX in marks between the zero and maximum penalty lines is given by the vertical distance between the two lines from the crossing of the x-axis by the maximum penalty line; this distance is 25%. This exercise can be repeated by replotting figure 7.1.2 for different values of A, (e.g. A = 2 to A = 10).

DMAX   =   T

We short cut this effort of replotting several times, by considering equation [8]. For this equation we assume zero knowledge on the part of the student by putting KM = 0, despite the fact that the number of "right" answers is R = DMAX. This gives us a new equation [10] just above; using it, we can then dutifully plot and tabulate the results. This also gives us the graph shown in figure 7.1.3 above, except that now the y-axis would be labelled "Maximum Mark Difference", but otherwise the resultant graph and the units for its axes are identical.

7.4 Educator's Effort and the Number of Choices per question?

What is the optimum number (A) of choices for each question? Frequently, 3 or 4 choices are used. The greater the number (A) of choices, the more time must be allocated to a student to complete a question, but the smaller will be the influence of blind guessing by the student. From the teacher’s perspective, the effort involved in providing a large number of choices per question means more time spent on assessment at the expense of contact time with the students.

Consider the following weighted cost function F(A), for which the term (αA) indicates the time spent preparing the question, where “α” is a weighting factor, and the term (β/A) indicates the effect of guessing, where “β” is a further weighting factor.

F(A)   =   α A +   β

If we assume that all variables in equation [11] are positive, then equation [11] has a minimum for:

A   =   β

If we look at the question in reverse and assume 3 or 4 choices per question, then equation [12] of course gives us an infinite number of corresponding matching ratios (β : α). The lowest values of the terms in the corresponding ratios are 9:1 and 16:1. In practical terms, this means that minimising the effect of guessing is regarded as being 9 or 16 times more important than the initial work in preparing the question. This fits in with normal practice. The emphasis on “initial work” highlights one advantage for the teacher of building up and maintaining a sizeable bank of multiple choice questions. Of course, as has already been mentioned, another advantage of such a question bank, is to ensure the effect of novelty in the questions for the student.

7.5 Choices per Question and the Usable Range of Raw Marks

If we compare raw data and averaging approaches to processing MCT marks with two different values for A, the number of choices per question, and we then plot the results as in figure 7.5.2, we notice the following. While the gradients of the set of plots for a given value of A are identical, the gradients for different values of A appear to increase with increase in A. Thus for A = 2, the gradient is arc tan (1) = π/4 radians or 45 degrees, and for A = 4, these values are steeper at arc tan (4/3) or 53.1 degrees. If A tends to infinity the angle would tend to 90 degrees.

Fig. 7.5.1: Comparing raw data and averaging approaches to processing MCT marks with two different values for A, the number of choices per question.

Fig. 7.5.2: Comparing raw data and averaging approaches to processing MCT marks with two different values for A, the number of choices per question. Plot of table in figure 7.5.1 above.

The steeper (i.e. larger) the gradient, the smaller the range of possible marks for a given value of Q. This is not desirable, but nor is the legitimate effort of producing too high a number A of choices per question. Somewhere there should be an "optimum" value of A, given these apparently conflicting aspects. To try and tackle this, we use equation [5] to establish the gradient which is then given by equation [13].

  =   1 +   1
A - 1

This enables us to calculate, for a given value of A, the inverse gradient, expressed as a percentage - where we assume a maximum possible mark of 100%. So for A = 4 we get 75% and for A = 2 we get 50%, both of which accord with the plots in figure 7.5.2. Where the maximum number of questions attempted is less than the number offered, we get a pro rata reduction in the maximum possible range of raw data marks. This also accords with figure 7.5.2. Thus, if we increase A, we increase the sensitivity of the processing of the marks to the available raw data marks. This is amplified in the table in figure 7.5.3 below. Of course, as we try and increase A indefinitely, we get diminishing returns regarding range sensitivity. In fact, as A tends to infinity, the plot in figure 7.5.2 indicates that the characteristic would tend to the characteristic (blue line) for the raw data approach, with of course, the added advantage of introducing the averaging approach.

Fig. 7.5.3: Relative ranges for raw data (R) values for different numbers A of choices per question.

To clarify the concept of ranges we can also use the diagram of figure 7.5.4 below, which highlights ranges associated (for example) with Q = 12 (i.e. 12 questions attempted), and the corresponding available (usable) ranges R4 and R2 of raw data. From this it can be seen that the ratios R4/R∞ and R2/R∞ give 0.75 and 0.50 respectively as per the above table in figure 7.5.3.

Fig. 7.5.4: Relative ranges R∞, R2 and R4 for raw data (R) values for Q = 12, and choices per question given by A = 2 and A = 4.

7.6 The Raw Data Line is the Limiting Case as A tends to Infinity

The Raw Data line (the blue diagonal line) in figures 7.5.2 and 7.5.4 can be regarded as the limiting case, as A tends to infinity, of both the raw data and the averaging methods.

8. Subtractive Strategy

8.1. Subtractive - using questions attempted

In this strategy, for every wrong answer we subtract a given number of marks from the marks total, J. Thus correct answers will cause marks to be added into a total, and incorrect answers will cause marks to be subtracted from the same total. This gives equation [14]. For example, a right answer will contribute 1 to J and a wrong answer could cause 1 to be subtracted from J. Note that now, we use J instead of K. K, as used in the previous sections above, is the number of questions actually known and at the same time, the result of the MCT. Now, however, J corresponds only to the result of the MCT, and not at the same time to the number K of questions deemed to be known. In the present subtractive case, it is not good enough for a student to know the answer to a question; there is additional pressure for a student not to give a wrong answer.

In addition, in this case, we would have G = 1. In general, G could be a factor other than one, e.g. 0.5, 1.5 or 2. With this strategy, note that "G" is separate from A, the number of choices per question.

J   =   R   −   G ( Q − R)

Let us use the following substitution for G.

G   =   1
B   −   1

This gives us equation [16], which is algebraically equivalent to equations [5] and [7], with A in those equations replaced by B, and K replaced by J.

J   =   R   −   Q − R
B − 1

Now for G = 1, 1/2, 1/3, we have equivalent values B = 2, 3, 4 respectively. This yields a mathematically convenient formulation - convenient in that it is similar to the previous formulations using A. However, it should be noted that for this marking strategy, G (and by implication B) is actually independent of, and chosen separately from, the actual number A of choices per question.

In other words, the subtractive marking method gives us two variables to manipulate, namely the number A of choices per question and the amount G by which the marker determines the penalising subtraction.

8.2. Subtractive - using questions offered

A variation of this strategy is to put Q=T, as in section 6.4. In other words, we also regard questions that have not been attempted as wrongly answered questions. In this case we replace Q by T. Equation [15] relating G to B is of course still valid, and so is the fact that for this marking strategy, G is independent of the actual number A of choices per question.

8.3. More Numerical Comparisons

Here you will see figures 8.3.1 and 8.3.2, which are essentially the same as figures 7.5.1 and 7.5.2, but with A = 2 replaced by G = 1. G = 1 is a popular value when applying the subtractive method of processing MCT marks. Essentially, for every wrong answer we subtract one mark from the final total, where, as we said, each correct answer carries one mark.

Fig. 8.3.1: Comparing raw data, averaging and subtractive approaches to processing MCT marks.

Fig. 8.3.2: Comparing raw data and averaging approaches to processing MCT marks with A = 4 and G = 1. Plot of table in figure 8.3.1 above.

In figures 8.3.1 and 8.3.2, it is assumed that we consider the number Q of questions attempted and only the total number of questions offered in the limiting case of Q = 20 (or 100% when scaled). In these figures you can thus see, that the slopes for G = 1 are steeper than those for the average MCT processing method using the popular choice of A = 4. The marking for the subtractive approach is noticeably more sensitive with change in the raw data R values than for the averaging approach. Put another way, to achieve the same output mark (K or J), a greater range of raw data marks is required in the averaging approach than in the subtractive approach - which many would see as an aspect favouring the averaging approach.

The table in figure 8.3. is constructed in the manner of the table in figure 7.5.3 and serves to amplify the way that the range of available raw data values deteriorates as G increases. As before, the maximum available range should be reduced pro rata according the maximuum number of questions actually attempted.

Fig. 8.3.3: Relative ranges for raw data (R) values for different G values.

9. Future Challenges

Multiple Choice Tests have gained great popularity with students, especially when the tests are computer-based, and with teachers who, by saving on assessment time, can devote more resources to the actual process of teaching. The following possibilities show that there are still many challenges with setting and devising MCTs.

10. Conclusions

We should remember that the computer conveniently gathers the marks for us; what we do – or ask the computer to do - with those marks is up to us as the assessors and could have a significant effect on a student’s marks in the MCT.

Whilst multiple choice tests (MCTs) readily lend themselves to automatic marking, we can see that one of the challenges of MCTs lies in the need - since we do not know the number C of questions got right by chance - to make some sort of assumption, in order to assign marks for the MCT. The averaging assumption given by equation [3] is certainly usable, especially where no other factors such as a student's partial knowledge, the way that a course has been taught or errors due some distraction are assumed.

Using the averaging assumption of equation [3], we have suggested that the raw data marking scheme is unfair to the marker or examiner, and indeed ultimately to the student, in that students can get marks by blind guessing, without knowing anything about the subject at all. The subtractive approach is unfair to the student in that it overly restricts the range of possible raw data marks that could actually produce output marks for the student.

It is suggested that the fairest approach is the averaging approach, where a selection of A = 4 for the number of choices per question provides a balance between (1) the range of raw data marks attracting processed marks, (2) a reasonable hedge against the effect of guessing by the student and (3) not too great a need to introduce fake or dummy choices for each question. Another recommendation is to consider the number of questions attempted as opposed to the total number of questions offered, in situations where the latter exceeds the former.

11. References & Links

While the amount of literature on multiple choice tests is quite extensive, there appears to be little in the way of a discussion on how to process and evaluate the marks arising from such tests, marks which are often collected automatically and at great speed. Possibly a question, by the setters of MCTs, of acting in haste and repenting at leisure! However, the first set of the following references is directly relevant to the aims of this web paper, the second set is tangentially relevant.

11.1 Main References

[1]    Eric M. Scharf and Lynne Baldwin, Assessing multiple choice question (MCQ) tests – a mathematical perspective, Active Learning in Higher Education, Sage Publications, DOI: 10.1177/1469787407074009, Vol 8(1): 33–49, 2007.
Available at:
(This paper contains a number of relevant references).
[3]    Robert B. Frary, Formula Scoring of Multiple-Choice Tests (Correction for Guessing) - Teaching Aid, Educational Measurement: Issues and Practice, National Council on Measurement in Education, 1988, 7(2),, accessed: 2012-12-12.
[4]    Australian Mathematics Trust, AMC Scoring System, Australian Mathematics Trust,, accessed: 2013-01-08.

11.2 Further References

[6]    Frank Baker, Item_response_theory, December 2001, accessed: 2012-12-12.
[7]    Michael J. Rees, Automatic assessment aids for Pascal programs,
Newsletter, ACM SIGPLAN Notices, Volume 17 Issue 10, October 1982 Pages 33 - 42, ACM New York, NY, USA, DOI: 10.1145/948086.948088,, accessed: 2012-12-12.
[8]   David Jackson, A software system for grading student computer programs, Computers & Education, Volume 27, Issues 3–4, December 1996, Pages 171–180., accessed: 2012-12-12.
[9]    Riku Saikkonen, Lauri Malmi, Ari Korhonen, Fully automatic assessment of programming exercises, ITiCSE '01 Proceedings of the 6th annual conference on Innovation and technology in computer science education, Volume 33 Issue 3, Sept. 2001, Pages 133-136, ACM New York, NY, USA ©2001, ISBN:1-58113-330-8, DOI: 10.1145/377435.377666., accessed: 2012-12-12.

12. Symbols and Abbreviations

To keep life simple, only important symbols and those that are not local to a particular section are included here. This section therefore serves an aide-memoire to try and make the life of you the reader as straight forward as possible when negotiating some of the more - sometimes inevitably intertwined - sections of this web paper.

11.1. Symbols

In the table below, symbols are grouped according to the section (§) in which they first occur. Please also remember that, in furtherance of simplicity, we have assumed, without loss of generality, one mark per question. This means, for example, that the number of raw marks (marks for right answers) translates directly and without scaling to the actual raw marks themselves.

?  number of ...   §
A  choices per question6
C  questions right by chance"
K  questions known"
Q  questions attempted"
R  raw ("right") marks"
T  questions offered"
?   description  §
F  function of A7
α  coefficient of A"
β  coefficient of 1/A"
B  "A equivalent" of G8
G  proportion of "non-right" questions"
J  result of "subtractive" MCT"

11.2. Abbreviations

MCQ  Multiple Choice Question
MCT  Multiple Choice Test

Attribution. On this web page, the picture of a multiple choice questionnaire in the header is from Microsoft's® clip-art libraries, supplied with some versions of MS Office®.