Académique Documents
Professionnel Documents
Culture Documents
0 INTRODUCTION
For this project, our group is required to build or design one complete objective
test paper which consists of 40 items or multiple choice questions. We have decided to
build an English test paper with 40 items and each item contains 4 options. Our test paper
also consists of 5 sections and it must be answered by the students within 1 hour.
mentioned below:
3) It can measure the students’ performance and differentiate between the top (high
scorers) and bottom performers (low scorers) on the test as a whole for the chosen
class.
5) It is useful for measuring whether the students can analyze material in the items
correctly or not.
6) It can be used to check whether the students have learned facts and routine
7) It encourages students to think critically before they choose the best options for
8) It enables us (teachers) to analyze each item that we have built, in which the key
aim in this item analysis is to increase the reliability and validity of the test. By
1
analyzing the items, we will be able to identify the weak items and try to come
Our group has chosen to distribute the test papers to one Form 2 class in Sekolah
Menengah Kebangsaan (SMK) Bukit Gambir, Pulau Pinang. We went to the school twice,
which is on 9th June 2008 (Monday) to ask permission from the school’s principal and
discuss with the English Language Coordinator regarding the suitable time to distribute
the test papers. On 13th June 2008 (Friday), we went to the school again to distribute the
The students begin to answer the test paper on 8.45 a.m. until 9.45 a.m. After all
the students have answered all the questions within the stipulated time, our group has
taken the opportunity to ask for the students’ feedback regarding the question paper that
we have distributed to them. Overall, the students felt that the items in the question paper
are at the average or medium level, whereby there are some difficult questions which
create confusion to the students and some easy questions which can be answered by the
students without any problem. Generally, our group felt very happy and satisfied with the
school and students’ cooperation during the distribution of the test papers.
Besides that, our group has also prepared the project’s report which contains 4
main topics such as the test specification table, the students’ performance (frequency,
mean, mode, median, variance and standard deviation), item analysis and suggestions or
2
2.0 TEST SPECIFICATION TABLE
Section A
(Graphic
Materials and
Stimuli ) –
Students would
need to study
information found
in the graphic
materials and
short texts and
answer the
questions based on
them.
3
• Item No.2 2) Students will be able to
identify the most accurate √
information of the label as
displayed in the given
picture.
4
• Item No.6 6) Students will be able to
identify the correct √
objective of the workshop
as shown in the poster
given.
5
• Item No.10 10) Students will be able to
identify the correct type √
of menu as shown in the
picture given.
Section B
(Rational
Cloze) – This
section tests
the students’
knowledge of
grammar and
vocabulary.
Students need
to learn how
they can apply
and use the
clues in the
text to get the
correct answer.
• Item No.11
11) Students will be able to
choose the most accurate √
interrogative pronoun to
construct a grammatically
correct sentence.
6
• Item No.12 12) Students will be able to
choose the most accurate √
interrogative pronoun to
construct a grammatically
correct sentence.
7
• Item No.16 16) Students will be able to
choose the most accurate √
interrogative pronoun to
construct a grammatically
correct sentence.
Section C
(Closest In
Meaning) – In
this section,
students will
learn how to
use available
clues to answer
8
the questions
on similar
expressions.
9
• Item No.23 23) Students will be able to
choose the best meaning √
for the underlined phrases
as provided in the
conversation given.
Section D
(Reading
Comprehension)
– This section
provides all
kinds of
comprehension
passage and
students should
know how to
read and
understand
different types of
comprehension
passages.
10
• Item No.25 25) Students will be able to
identify the correct type √
of bamboo sticks to make
the kite by looking at the
information in the given
descriptive passage.
11
• Item No.29 29) Students will be able to
choose the correct √
purpose of the letter by
looking at the information
in the given letter.
12
• Item No.33 33) Students will be able to
choose the best √
description about Emelda
Allyn based on the
information in the given
letter.
Section E
(Literature
Component) –
In this section,
students need
to have better
understanding
of the
literature texts.
13
• Item No.36 36) Students will be able to
choose the best answer to √
fill in the blank in the
sentence by referring to
the information in the
given poem.
14
• Item No.40 40) Students will be able to
identify the important √
moral value that the
people of Dalat had learnt
from the incident
happened in their village
correctly by looking at the
extract of the given short
story.
Percentage 40 % 5% 24 % 9% 2% - -
15
3.0 THE OVERALL REPORT OF THE STUDENTS’ PERFORMANCE
STANDARD DEVIATION
Overall, the performance of the form two students was very good. This is because none of
the student fails in this exam. Most of them managed to get a good marks range from 30 or 70%
and above. This can be shown in the score table, Table 3.2, and the frequency table, Table 3.3,
below whereby only one of the student got 22 marks out of 40 marks or 55%, 3 students got 27
marks out of 40 marks which also amount to 67.5%, 5 students got 29 marks over 40 marks that
is 72.5%, 13 students got the marks range from 30 to 34 marks over 40 marks and 8 students got
the marks range from 35 to 39 over 40 marks. In the other words, from the results shown, they
have clearly shown that the exam questions can be considered as very easy and the level of
intelligence of the students were excellent. This can be supported by the histogram shown in
Histogram 1.
16
S14 32 -0.2 0.04
S5 31 -1.2 1.44
S20 31 -1.2 1.44
S25 30 -2.2 4.84
S3 29 -3.2 10.24
S6 29 -3.2 10.24
S10 29 -3.2 10.24
S15 29 -3.2 10.24
S27 29 -3.2 10.24
S2 27 -5.2 27.04
S23 27 -5.2 27.04
S29 27 -5.2 27.04
S8 22 -10.2 104.04
Meanwhile, from the table above we can see that the mean or the average marks obtained
by 30 students of this Form Two class is equal to 32.20 which also amount to 80.5%. Whereas,
the mode which also refers to the score that occurs most frequently for this class is 34. In the
other words, most of the students had averagely scored 85% for this test. The variance or the
average of the squared differences of the students’ scores from the mean is equal to 14.16 and the
standard deviation or the average of the differences of all scores from the mean is equal to 3.763.
17
Scores (x/40) Percentage (%) Number of Students
40 100% 0
39 97.5% 1
38 95% 1
37 92.5% 0
36 90% 5
35 87.5% 1
34 85% 6
33 82.5% 2
32 80% 2
31 77.5% 2
30 75% 1
29 72.5% 5
28 70% 0
27 67.5% 3
26 65% 0
25 62.5% 0
24 60% 0
23 57.5% 0
22 55% 1
21 52.5% 0
20 50% 0
TABLE 3.2: Students Performances’ Table
18
Frequenc
Bin y
22 1
25.4 0
28.8 3
32.2 10
35.6 9
More 7
TABLE 3.3: Frequency
From the histogram shown above, it shows that the distributions of the students’ results
are negatively skewed. Meaning to say, the exam questions are quiet easy for them therefore
most of the students managed to score a very good grade for this test. This may be caused by the
level of intelligence of the students themselves, in which they are from the advance class. Thus,
their mastery of English Language is very good. Apart from that, the designed items in this exam
paper can be considered as unable to measure, to evaluate or to discriminate the real performance
of each student in the classroom as they still cannot differentiate between the excellent students
19
4.0 ITEM ANALYSIS
Item analysis is carried out to aid in evaluating the effectiveness of a test item. In conducting the
item analysis of this test, item difficulty, item discrimination and distracters quality will be
considered. The students are categorised into upper-group and lower-group. The scores of the
students were sorted descending 27% of the high-scoring students are categorised into upper-
group and 30% of the low-scoring students are categorised into lower-group students.
20
S20 31
S25 30
S3 29
S6 29
S10 29
S15 29
S27 29
S2 27 - Upper-group students
S23 27 - Lower-group students
S29 27
S8 22
The index of difficulty (p value) is the proportion of the total group who got an item right. It
ranges between 0 – 1. A p-value closer to 1 indicates the easier the item is as more students can
answer the item correctly. On the other hand, the closer the value to 0, the more difficult the item
is as fewer students got the answer right. The index of difficulty (p) of each item of the test was
where,
P = index of difficulty
Ncorrect = number of students answering correctly
Ntotal = number of students taking the test
21
7 0 0 0 *8 0 0 6 *3 9 0.667 1.000 0.333 0.667
8 0 *8 0 0 1 *7 1 0 12 0.900 1.000 0.778 0.222
9 0 0 *8 0 1 1 *7 0 12 0.900 1.000 0.778 0.222
10 0 0 0 *8 1 1 0 *7 13 0.933 1.000 0.778 0.222
11 0 0 0 *8 0 0 0 *9 13 0.967 0.875 1.000 -0.125
12 *6 2 0 0 *6 2 0 1 7 0.633 0.750 0.667 0.083
13 *7 0 0 1 *0 1 5 3 9 0.533 0.875 0.000 0.875
14 0 *8 0 0 1 *7 1 0 11 0.867 1.000 0.778 0.222
15 0 1 0 *7 2 1 3 *3 5 0.500 0.875 0.333 0.542
16 0 1 0 *7 3 0 1 *5 10 0.733 0.875 0.556 0.319
17 *7 0 1 0 *4 1 4 0 7 0.600 0.875 0.444 0.431
18 0 0 *8 0 2 0 *7 0 10 0.833 1.000 0.778 0.222
19 0 0 *8 0 2 1 *6 0 11 0.833 1.000 0.667 0.333
20 0 *8 0 0 1 *8 0 0 13 0.967 1.000 0.889 0.111
21 0 *7 0 1 0 *7 0 2 13 0.900 0.875 0.778 0.097
22 0 0 *8 0 0 0 *9 0 13 1.000 1.000 1.000 0.000
23 0 0 0 *8 0 0 0 *9 13 1.000 1.000 1.000 0.000
24 *7 0 0 1 *9 0 0 0 11 0.867 0.750 1.000 -0.250
25 *8 0 0 0 *5 1 0 3 10 0.767 1.000 0.556 0.444
26 0 *8 0 0 1 *6 2 0 13 0.900 1.000 0.667 0.333
27 0 0 *8 0 0 2 *7 0 13 0.933 1.000 0.778 0.222
28 *8 0 0 0 *8 0 0 1 11 0.900 1.000 0.889 0.111
29 *8 0 0 0 *7 0 2 0 12 0.900 1.000 0.778 0.222
30 *5 0 3 0 *5 0 3 1 6 0.533 0.625 0.556 0.069
31 2 0 0 *6 4 2 0 *3 8 0.567 0.750 0.333 0.417
32 *8 0 0 0 *4 1 3 1 11 0.767 1.000 0.444 0.556
33 *8 0 0 0 *4 2 1 2 9 0.700 1.000 0.444 0.556
34 0 *8 0 0 0 *9 0 0 9 0.867 1.000 1.000 0.000
35 0 0 *8 0 0 0 *9 0 13 1.000 1.000 1.000 0.000
36 *8 0 0 0 *8 1 0 0 13 0.967 1.000 0.889 0.111
37 0 *8 0 0 2 *7 0 1 11 0.833 1.000 0.667 0.333
38 0 0 *8 0 1 1 *7 0 13 0.967 1.000 0.889 0.111
39 0 0 *8 0 0 0 *9 0 13 1.000 1.000 1.000 0.000
40 *4 0 0 4 *4 0 1 4 10 0.633 0.500 0.556 -0.056
* Denotes Correct Answer
TABLE 4.1: Calculation of P-values And Discrimination Index for Each Item
From the table above, the p-value for each item is listed in the Pall column. The p-values
calculated range from 0.467 to 1.000. Since this test is a norm-referenced test (NRT), therefore
22
the average difficulty index should be within 0.34 to 0.66. The difficulty indices were analysed
Item with p-value lesser than 0.33 will be considered as a difficult item. Whereas item with p-
value between 0.34 to 0.66 is a moderately difficult item and item with p-value more than 0.67 is
an easy item.
Based on the table above, most items are categorised into low difficulty item which means this
test is an easy test. There seems to be an imbalance distribution among easy and difficult items
where 78% of the items have low level of difficulty whereas none of the item in the test has high
level of difficulty. However, the discrimination index should also be considered in analysing
The index of discrimination (D) is the difference between the proportion of the upper group who
answered an item correctly and the proportion of the lower group who answered the item
correctly. This index is dependent upon the difficulty of an item which means it relates with the
index of difficulty of the item. The D of each item for this test is calculated using the formula
below:
D = Pupper – Plower
23
where,
D = item discrimination for an individual item
Pupper = item difficulty for the upper group on the whole test
Plower = item difficulty for the lower group on the whole test
Ebel’s (1979) criteria and guidelines for categorising discrimination indices is a widely quoted
set of guidelines and therefore is used in this test analysis to categorise the 40 test items.
Discrimination index
Description
(D)
≤ 0.19 (Bad) The item should be eliminated or completely revised.
0.20 – 0.29 (OK) The item is marginal and needs revision.
0.30 – 0.39 (Good) Little or no revision is required.
≥ 0.40 (Very Good) The item is functioning quite satisfactorily.
TABLE 4.4: Ebel’s Guidelines
Based on the Ebel’s guidelines in the above table, the 40 test items can be categorised as follows:
Discrimination Index
Items Count
(D)
≤ 0.19 (Bad) 2,3,4,11,12,20,21,22,23,24,28,30,34,35,36,38,39,40 18
0.20 – 0.29 (OK) 6,8,9,10,14,18,27,29 8
0.30 – 0.39 (Good) 5,16,19,26,37 5
≥ 0.40 (Very Good) 1,7,13,15,17,25,31,32,33 9
TABLE 4.5: Item Categorisation Based on Discrimination Index
The results indicate that about 65% of the test items are week in discriminating between the
‘good’ and ‘weak’ students and these items need to be looked closely as they may need
Item Difficulty
Item Medium (0.34 –
Discrimination High (≤ 0.33) Low (≥ 0.67)
0.66)
≤ 0.19 (Bad) - 12,30,40 2,3,4,11,20,21,22,23,24,28,
24
34,35,36,38,39
0.20 – 0.29 (OK) - - 6,8,9,10,14,18,27,29
0.30 – 0.39 (Good) - 5 16,19,26,37
≥ 0.40 (Very Good) - 1,13,15,17,31 7,25,32,33
TABLE 4.6: Item Categorisation Based On The Relationship Item Discrimination and Item
Difficulty
The table above shows the relationship between the item discrimination and item difficulty for
each item. Items which are listed in the black coloured cell need to be revised or eliminated if the
Distracter analysis examines the proportion of students who selected each of the response
options. According to Tucker (2007), “on a well-designed multiple choice item, high scoring
students should select the correct option even from highly plausible distracters. Those who are
ill-prepared should select randomly from available distracters. In this scenario, the item would be
a good discriminator of knowledge and should be considered for future assessments. In other
scenarios, a distracter analysis may reveal an item that was mis-keyed, contained a proofreading
error, or contains a distracter that appears plausible even by those that scored well on an
assessment”. The proportion for each option is calculated by using the following formula:
where,
Prop = Proportion student choosing the distracter
n = number of students choosing the distracter
N = total number of students taking the test
The table below shows the proportion of students in the upper and lower group who selected the
correct answers as well as the proportion of students choosing each alternative for each item.
25
Upper Group (8 Students) Lower Group (9 Students) Discrim
Item Pall -ination
Index
A B C D A B C D
*0.62 0.46
1 0.125 0.125 0.125 0.111 0.556 *0.111 0.222 0.514
5 7
26
*1.00 *0.88 0.96
2 0.000 0.000 0.000 0.000 0.000 0.111 0.111
0 9 7
*1.00 *0.88 0.93
3 0.000 0.000 0.000 0.000 0.000 0.111 0.111
0 9 3
*0.75 *0.77 0.76
4 0.125 0.000 0.125 0.222 0.000 0.000 -0.028
0 8 7
*0.87 *0.55 0.53
5 0.125 0.000 0.000 0.444 0.000 0.000 0.319
5 6 3
*0.62 *0.33 0.66
6 0.000 0.125 0.250 0.222 0.333 0.111 0.292
5 3 7
*1.00 *0.33 0.66
7 0.000 0.000 0.000 0.000 0.000 0.667 0.667
0 3 7
*1.00 *0.77 0.90
8 0.000 0.000 0.000 0.111 0.111 0.000 0.222
0 8 0
*1.00 *0.77 0.90
9 0.000 0.000 0.000 0.111 0.111 0.000 0.222
0 8 0
*1.00 *0.77 0.93
10 0.000 0.000 0.000 0.111 0.111 0.000 0.222
0 8 3
*1.00 *1.00 0.96
11 0.000 0.000 0.000 0.000 0.000 0.000 -0.125
0 0 7
*0.75 *0.66 0.63
12 0.250 0.000 0.000 0.222 0.000 0.111 0.083
0 7 3
*0.87 *0.00 0.53
13 0.000 0.000 0.125 0.111 0.556 0.333 0.875
5 0 3
*1.00 *0.77 0.86
14 0.000 0.000 0.000 0.111 0.111 0.000 0.222
0 8 7
*0.87 *0.33 0.50
15 0.000 0.125 0.000 0.222 0.111 0.333 0.542
5 3 0
*0.87 *0.55 0.73
16 0.000 0.125 0.000 0.333 0.000 0.111 0.319
5 6 3
*0.87 *0.44 0.60
17 0.000 0.125 0.000 0.111 0.444 0.000 0.431
5 4 0
*1.00 *0.77 0.83
18 0.000 0.000 0.000 0.222 0.000 0.000 0.222
0 8 3
*1.00 *0.66 0.83
19 0.000 0.000 0.000 0.222 0.111 0.000 0.333
0 7 3
*1.00 *0.88 0.96
20 0.000 0.000 0.000 0.111 0.000 0.000 0.111
0 9 7
*0.87 *0.77 0.90
21 0.000 0.000 0.125 0.000 0.000 0.222 0.097
5 8 0
*1.00 *1.00 1.00
22 0.000 0.000 0.000 0.000 0.000 0.000 0.000
0 0 0
*1.00 *1.00 1.00
23 0.000 0.000 0.000 0.000 0.000 0.000 0.000
0 0 0
*0.87 *1.00 0.86
24 0.000 0.000 0.125 0.000 0.000 0.000 -0.250
5 0 7
27
*1.00 *0.55 0.76
25 0.000 0.000 0.000 0.111 0.000 0.333 0.444
0 6 7
*1.00 *0.66 0.90
26 0.000 0.000 0.000 0.111 0.222 0.000 0.333
0 7 0
*1.00 *0.77 0.93
27 0.000 0.000 0.000 0.000 0.222 0.000 0.222
0 8 3
*1.00 *0.88 0.90
28 0.000 0.000 0.000 0.000 0.000 0.111 0.111
0 9 0
*1.00 *0.77 0.90
29 0.000 0.000 0.000 0.000 0.222 0.000 0.222
0 8 0
*0.62 *0.55 0.53
30 0.000 0.375 0.000 0.000 0.333 0.111 0.069
5 6 3
*0.75 *0.33 0.56
31 0.250 0.000 0.000 0.444 0.222 0.000 0.417
0 3 7
*1.00 *0.44 0.76
32 0.000 0.000 0.000 0.111 0.333 0.111 0.556
0 4 7
*1.00 *0.44 0.70
33 0.000 0.000 0.000 0.222 0.111 0.222 0.556
0 4 0
*1.00 *1.00 0.86
34 0.000 0.000 0.000 0.000 0.000 0.000 0.000
0 0 7
*1.00 *1.00 1.00
35 0.000 0.000 0.000 0.000 0.000 0.000 0.000
0 0 0
*1.00 *0.88 0.96
36 0.000 0.000 0.000 0.111 0.000 0.000 0.111
0 9 7
*1.00 *0.77 0.83
37 0.000 0.000 0.000 0.222 0.000 0.111 0.333
0 8 3
*1.00 *0.77 0.96
38 0.000 0.000 0.000 0.111 0.111 0.000 0.111
0 8 7
*1.00 *1.00 1.00
39 0.000 0.000 0.000 0.000 0.000 0.000 0.000
0 0 0
*0.50 *0.44 0.63
40 0.000 0.000 0.500 0.000 0.111 0.444 -0.056
0 4 3
* Denotes Correct Answer
TABLE 4.7: The Proportion of Students in the Upper and Lower Group Who Selected
Each Option
This statistic measures test reliability of inter-item consistency. A KR-20 value ranges from 0 – 1.
A higher value indicates a strong relationship between items on the test. While a lower value
28
indicates a weak relationship between test items. Therefore, a test has a better reliability when
where,
KR-20 - Kuder Richardson 20
k - Number of items in the test
p - Item difficulty
2
S - Variance of the raw scores or standard deviation squared
The value of KR-20 calculated for this test is 0.63. It seems that this test is moderately reliable.
In this chapter, we will be analyzing the item in order to discover items that consist of ambiguous
item, miskeyed, too easy or too difficult item and non-discriminating item.
The reasons for this analyzing are to enhance the technical quality of an examination by pointing
out options that are nonfunctional and that should be improved or eliminated.
29
Item that virtually everyone gets right are useless for discriminating among students and should
be replaced by more difficult items or eliminated. This could be seen from the proportion of
students answering an item correctly. This point may be summarized by saying that items
discriminate.
When all items are extremely easy, most test scores will be extremely high. In either case, test
scores will show very little variability. Thus, extreme p values directly restrict the variability of
test scores. When everyone taking the test chooses the correct response as is seen in Table 1. An
item with a p value of .0 or a p value of 1.0 does not contribute to measuring individual
Example
Question 22:-
30
Table 1
Options A B C D
Upper 0 0 8 0
group
Lower 0 0 9 0
group
Note. * denotes correct response
Item difficulty: (8+9)/17 = 1.00p
Discrimination Index: (8-9)/17 = 0.05
A. it is still new
B. it is interesting
C. it is not dirty
D. it is in a good condition
Table 2
Options A B C D
Upper 0 0 0 8
group
31
Lower 0 0 0 9
group
Note. * denotes correct response
Item difficulty: (8+9)/17 = 1.00p
Discrimination Index: (8-9)/17 = 0.05
Based on table one (1) and table two (2,) we could suggest that the items must be improved by
making the distracter more attractive and these items should be replace or eliminate in order to
One measure of item ambiguity is the extent to which students in the upper group select an
incorrect option with about the same frequency as they select the correct one. Ambiguity defined
in this way is the inability of the highest-scoring students on the test to discriminate between a
“correct” alternative and one judged by the teacher to be “wrong”. An ambiguous item could also
be defined as one that allows for more than one “correct” alternative as judged by a group of
experts, although a question that is clear to experts may be ambiguous to students who lack
40. What was the important moral value that the people of Dalat had learned from the incident
A. to value peace
B. never listen to brothers
C. customs are waste of time they should obey their siblings
D. they should obey their siblings
32
Table 3
Item Difficulty
Options A B C D
Upper 4 0 0 4
group
Lower 4 0 1 4
group
In the example, the item appears to be ambiguous because the two options can be justified.
When students in the upper portion of the class select a “correct” option and an “incorrect”
option with about equal frequency, the item is ambiguous either because students lack knowledge
or because the options or the item was defective. Which of these reasons is applicable to a given
item is determined by examining the highly selected but “incorrect” options to see if more than
one answer can be justified. Our suggestion to this item is to look at their favorite alternative
again, and see if we can find any reason they could be choosing it.
5.1.3. Miskeying
33
When we first correcting students’ test paper, we have identified a problem where students have
giving the correct answer but miskeyed by us. Miskeying is another common error that can be
corrected before students’ papers are returned. One way of detecting potentially miskeyed items
is to examine the responses of the students in the upper portion of the class. An “incorrect”
option selected by a large number of these students suggests a keying error, as in the following
example:
28. A string must be tied to each __________ of the stick to make frame.
Because the majority of the most capable students selected ‘end’ as the correct alternative and so
few agreed with the “keyed” answer, the teacher should check the key for a possible error.
Analyzing the distracters is useful in determining the relative usefulness of the decoys in each
item.
Example: Item 5.
A. vary
B. come
C. fly
D. watch
34
Table 4
Options A B C D
Upper group 7 1 0 0
Lower group 5 4 0 0
Based on the example, we could clearly identified that no one fell for options C and D. Option B
is a good distracter however options C and D is not a plausible alternatives. Our suggestion is, if
a distracter elicits very few or no responses, then it may not be functioning as a distracter and
Example: Item 6
A. visit parks
B. hold a talk
C. promoting membership
D. enlighten students
Table 5
Effective distracters
35
Options A B C D
Upper 0 1 2 5
group
Lower 2 3 1 3
group
Note. * denotes correct response
Item difficulty = 0.667
Discrimination Index = 0.292
Distracters should be carefully examined when items show large positive p-values. Our comment
is, items should be modified if students consistently fail to select certain multiple choice
alternatives. The alternatives are probably totally implausible and therefore of little use as decoys
in multiple choice items. Also some distracters may be too appealing, causing the item to be too
difficulty, discriminateable, or variability can be redeemed by the revision of one or two of the
response options.
Based on the item analysis, we could clearly identify that the students have answer the question
without guessing or respond to the questions randomly. This is mainly because of; the test was
6.0 CONCLUSIONS
Overall, we could say that this test is moderately reliable. However, most of the items are too
easy to be given to advance level students. As a result, most of the items were not functioning
very well and could not perfectly differ between upper and lower students. Nevertheless, the item
36
analysis statistics of item discrimination (IF) and item difficulty (ID) help us to decide which
items to keep and which to discard in creating a new revised version of the test. Moreover, the
distracter efficiency analysis is also useful for spotting items that are miskeyed and for tuning up
those items that have options that are not working as would be expected.
7.0 REFERENCES
Ebel, R.L. (1979). Essentials of educational measurement (3rd ed.). Englewood Cliffs, NJ:
Prentice Hall.
Henning, G. (1987) A Guide To Language Testing Development, Evaluation, Research. London:
Newbury House Publisher.
Tucker, S. (2007, September 21). Retrieved June 26, 2008, from University of Maryland:
http://www.umaryland.edu/cits/testscoring/pdf/sop_deconstructingtestscoring.pdf
37
8.0 APPENDIX
38