Académique Documents
Professionnel Documents
Culture Documents
net/publication/242742418
CITATIONS READS
3 6,882
2 authors:
Some of the authors of this publication are also working on these related projects:
Educational outcomes among survivors of childhood cancer in British Columbia, Canada: Report of the Childhood/Adolescent/Young Adult Cancer Survivors (CAYACS)
Program View project
All content following this page was uploaded by Alain Desrochers on 21 April 2015.
Introduction
Reading assessment is typically carried out to guide changes (Gaudreau, 2001) either
in individual interventions, instructional programs, or curricula. The result of an
individualized reading assessment may indicate that some children are ‘at risk’ of
developing reading skills that are significantly below the level of their peers, and that
some form of intervention is warranted: remediation, individualized instruction, or
placement into a special program. When the target of change is the instructional
program or curriculum, the assessment compares children as a group to the reading
goals that were initially set out in the curriculum; reading assessment results may then
indicate that the reading program needs to be upgraded.
Diagnostic assessment. Children may face difficulty in learning to read for a variety of
reasons. This diversity is due to the complexity of reading acquisition, which is based on
a large set of elementary skills (for detailed analyses, see Coltheart, 2005; Seymour,
1986; Sprenger-Charolles, Colé, & Serniclaes, 2006). For instance, efficient reading
requires a normal ability to discriminate and recognize visual patterns, process speech
sounds, convert graphemes into speech sounds, recognize and read out whole words,
and access meanings from printed words. Diagnostic test batteries are designed to
assess the strengths and weaknesses of readers on these elementary skills and identify
the components of reading that an intervention should target (for examples, see Reid,
Hresko, & Hammill, 2001; Wagner, Torgesen, & Rashotte, 1999).
Program evaluation. The focus of reading outcome assessment need not be on the
child; it can be on the reading instruction program itself. The assessment is then
intended to evaluate the merits and weaknesses of a curriculum, an instructional
program, the consequences of an educational reform, or the success of a program
implementation. Outcome results can serve to modify the orientation or improve specific
components of the program or its implementation. This type of assessment is typically
based on the reading curriculum that pertains to a particular population of school-age
students.
In order to interpret test scores, three elements are required: a) a normative framework
to determine if individual scores are at, above or below what is expected from an
average child, b) a framework that can help link specific performance scores with
specific cognitive skills (e.g., phonological decoding), and c) a link to provincial
curriculum and learning outcomes.
5. How will an educator know if the scores from a reading test are accurate?
No measure of human ability is absolutely perfect; all measures entail a margin of error.
To maximize accuracy most measures of reading ability involve multiple items (e.g.,
words, sentences, passages of text). Several indices may be computed and examined
to assess the level of accuracy of a reading test: internal consistency, temporal stability,
and measurement error (for a detailed discussion, see Bertrand & Blais, 2004; Kline,
2005; Laveault & Grégoire, 2002; Sax, 1997).
Internal consistency refers to the inter-relationships among the items that comprise the
test. Responses to these items are expected to be determined by a common set of
abilities and the extent to which they are influenced by the same factors can be
measured. The most common index of internal consistency is the Cronbach Alpha
coefficient. If a particular reading skill (e.g., oral reading of words) is measured over two
consecutive days on the same children, we would expect these two measures to be
identical if no learning has taken place and if there is no measurement error. An
estimate of temporal stability (also called test-retest reliability) can be obtained by
calculating the correlation coefficient between the scores observed over two occasions
separated in time, on the same test, and from the same individuals. Because internal
consistency and temporal stability are never perfect, all test scores are expected to
‘wobble’ around their true value. The estimate of this ‘wobble’ is called the standard
error of measurement and it can be used to define an interval of confidence of a reading
score. This corresponds to the score interval within which the true score has a 95%
chance of falling.
Content. In many cases, content-related validity can be easily established. If the test is
intended to measure children’s ability to convert letters into speech sounds (e.g., as in
Association among reading-related skills. Many skills that are relevant to reading are
strongly correlated with one another. For instance, phonological awareness is strongly
associated with reading ability (for a review, see Kirby, Desrochers, Roth, & Lai, 2008).
We would then expect a good measure of phonological awareness to be significantly
correlated with specific aspects of reading such as oral decoding. This relationship is
observed in the correlations among concurrent measures. It is also observed when
measures of phonological awareness are used to predict the level of reading
performance achieved several weeks or months later. This form of criterion-related
evidence is particularly useful in the development of tests for screening children at risk
for reading failure.
Basis for decision making. Reading assessment is typically intended to guide one’s
decisions or actions. A common decision consists of determining if a child needs
remedial intervention and, if so, which components of reading should be targeted for
remedial instruction. These decisions depend, in part, on the test’s capacity to gauge
the severity of the child’s reading difficulties and to guide the ensuing intervention.
Further evidence of validity can be gained by assessing the goodness of the match
between the child’s reading profile, as revealed by the assessment tool, and the
recommended reading intervention based on the assessment.
Future Directions
We now discuss three research directions that may be considered in the future: a)
providing educators with a complete toolbox for the assessment of reading skills, b)
developing norms that are relevant to all Canadian children, and c) linking reading
assessment to reading intervention.
A complete toolbox for reading assessment. Presently, no single assessment tool can
measure all reading-related skills that are required to chart a profile of children’s
strengths and weaknesses. This goal can only be achieved by gathering information
from different assessment tools. Educators could benefit from some guidance on how to
select a complete set of assessment tools for their purposes: screening children at risk
of reading failure, reading progress monitoring, diagnostic assessment, or reading
program evaluation. A rigorous matching procedure between current needs and current
resources would permit us to determine what is presently lacking in Canada to build a
complete toolbox and how our test-development efforts should be invested.
Reading performance norms that are relevant to Canadian children. Most reading
assessment tools currently in use in Canada were developed in the United States or
United Kingdom (for English) or Europe (for French). This means that the performance
norms that are currently available for these tests were developed with populations of
children in countries other than Canada. Since the level of performance on reading tests
depends largely on reading curricula and programs, which are decided by provincial
ministries of education in Canada, some discrepancy may be present between the
average Canadian reader and the average reader represented in the norms established
in other countries. There may also be differences among average readers in different
Canadian provinces or regions. Should a common normative frame of reference be
developed for the whole of Canada or for each Canadian province or region? How
should Canadian linguistic diversity be addressed in developing these norms? At the
present time, examiners consider themselves fortunate to have any norms available to
base their decisions on. Further research on the development of reading performance
norms that are relevant to Canadian children would provide a more reliable basis for
interpreting test scores and guiding educational decisions and actions.
Conclusions
Desrochers, A., & Glickman, V. (2009). Criteria for the evaluation of reading
assessment tools. Encyclopedia of Language and Literacy Development (pp. 1-
9). London, ON: Canadian Language and Literacy Research Network.
Retrieved from http://literacyencyclopedia.ca/pdfs/topic.php?topId=280