Vous êtes sur la page 1sur 18

5/4/12

VALIDITY & RELIABILIT Click to edit Master subtitle style Y

5/4/12

Validity

5/4/12

(Dictionary): The state of quality of being valid: to question the validity of the argument Extent to which a test measures what it claims to measure. It is vital for a test to be valid in order for the result to be accurately applied and interpreted. In terms of assessment:
validity

refers to the extent to which a test's content is representative of the actual skills learned and whether the test can allow accurate conclusions concerning achievement.

5/4/12

Category of validity
i. ii. iii.

Content/face validity Validity related to the criteria Validity of constructs

5/4/12

It It

does not assess memory, when it is supposed to be assessing problem-solving (andvice versa). does not grade someone on the quality of their writing, when writing skills are not relevant to the topic being assessed, but it does when they are.

5/4/12

Content validity

Does the content of the test measure stated objectives? Important validity for teachers to prepare exam questions and the contents in the exams incorporated with the contents, syllabus ,curriculum specification and the objectives The contents depends on the students achievement, difficulties, and suitable tested items. Example: final year exams that only focus on certain topics but ignore other topics thus it invalid.

5/4/12

Validity related to the criteria


Relate 2

the validity to the stated criteria.

types: predict the future performance

Predictive validity

Concurrent validity the new exams are identical to the past exams. Accuracy to identify the current based on status regarding

5/4/12

Validity of constructs
Constructs,

such as creativity, honesty, and intelligence, are concepts used by social scientists to help explain empirical data about a phenomenon or to construct a theory validity measures the extent to which an examination actually measures what the specification says it is intended to measure.

Construct

5/4/12

languages

examinations are intended to measure candidates' skills in understanding, speaking, reading and writing the foreign language. a specification requires candidates to acquire practical skills, the assessment should include a practical assessment and should not just require candidates to, for example, write about how they would carry out a practical activity.

If

5/4/12

5/4/12

Reliability

5/4/12

What is reliability ?
used

in experiment, test or any measuring procedure. that will show the same result when we repeat. Without any condition of the independent observers that able to repeat the research procedure. research, the term reliability means "repeatability" or "consistency". A measure is considered reliable if it would give us the same result over and over

In

5/4/12 How to determine the reliability of instruments Test-retest method through on the same subjects at a future date. Test is repeated

Results are compared with initial test to determine stability.

Equivalent form

Determine by relating two sets of test scores to one another to highlight the degree of relationship.

Inter-rater Reliability

Dependent upon the ability of two or more individuals to be consistent. For example, when two or more teachers use a rating scale with which they are rating the students oral responses in an interview (1 being most negative, 5 being most positive). If a teacher gives a 1 to a student, while

5/4/12

Internal

Consistency

Assess the same characteristics, skill or quality. teachers interpret data and predict the value of scores and the limits of relationship among students.

Helps

Split
The

Half Method

Spearman-Brown prophecy formula is applied to the correlation to determine the reliability. that helps teachers to determine

Methods

5/4/12

Factor affecting reliability

Test length. the longer a test is, the more reliable it is.

Speed.
When a test is a speed test, reliability can be problematic. It is inappropriate to estimate reliability using internal consistency, test-retest, or alternate form methods. This is because not every student is able to complete all of the items in a speed test. In contrast, a power test is a test in which every student is able to complete all the items.

Item difficulty
When there is little variability among test scores, the reliability will be low. Thus, reliability will be low if a test is so easy that every student gets most or all of the items correct or so difficult that every student gets most or all of the items wrong.

5/4/12

Objectivity Objectively scored tests, rather than subjectively scored tests, show a higher reliability

Variation with the testing situation Difficulties in the testing situation. For example, students misunderstanding or misreading test instructions, noise level, distractions, and sickness can cause test scores to vary.

5/4/12

Relationship between validity & reliability


Validity A

and reliability are closely related.

test cannot be considered valid unless the measurements resulting from it are reliable

5/4/12

Likewise, results from a test can be reliable and not necessarily valid.

Vous aimerez peut-être aussi