Académique Documents
Professionnel Documents
Culture Documents
That people differ from each other is obvious. How and why they differ is less clear and is the
subject of the study of Individual differences (IDs). Although to study individual differences
seems to be to study variance, how are people different, it is also to study central tendency, how
well can a person be described in terms of an overall within-person average. Indeed, perhaps the
most important question of individual differences is whether people are more similar to
themselves over time and across situations than they are to others, and whether the variation
within a single person across time and situation is less than the variation between people. A
related question is that of similarity, for people differ in their similarities to each other. Questions
of whether particular groups (e.g., groupings by sex, culture, age, or ethnicity) are more similar
within than between groups are also questions of individual differences.
O
Taxonomic work has focused on categorizing the infinite ways in which individuals differ in
terms of a limited number of latent or unobservable constructs. This is a multi-step, cyclical
process of intuition, observation, deduction, induction, and verification that has gradually
converged on a consensual descriptive organization of broad classes of variables as well as on
methods for analyzing them. Most of the measurement and taxonomic techniques used
throughout the field have been developed in response to the demand for selection for schooling,
training, and business applications.
OO
Consider the case of differences in vocabulary in a particular language (e.g., English). Although
it is logically possible to organize people in terms of the specific words they know in English, the
more than 2^(500,000) possible response patterns that could be found by quizzing people on
each of the more than 500,000 words in English introduces more complexity rather than less.
Classical Test Theory (CTT) ignores individual response patterns and estimates an individual's
total vocabulary size by measuring performance on small samples of words. Words are seen as
random replicates of each other and thus individual differences in total vocabulary size are
estimated from observed differences on these smaller samples. The Pearson Product Moment
Correlation Coefficient (r) compares the degree of covariance between these samples with the
variance within samples. As the number of words sampled increases, the correlation of the
individual differences within each sample and with those in the total domain increases
accordingly.
Estimates of ability based upon Item Response Theory (IRT) take into account
parameters of the words themselves (i.e., the difficulty and discriminability of each
word) and estimate a single ability parameter for each individual. Although CTT and
IRT estimates are highly correlated, CTT statistics are based on decomposing the
sources of variance within and between individuals while IRT statistics focus on the
precision of an individual estimate without requiring differences between individuals.
CTT estimates of reliability of ability measures are assessed across similar items
(internal consistency), across alternate forms, and across different forms of assessment
as well as over time (stability). Tests are reliable to the extent that differences within
individuals are small compared to those between individuals when generalizing across
items, forms, or occasions. CTT reliability thus requires between subject variability.
IRT estimates, on the other hand, are concerned with the precision of measurement for
a particular person in terms of a metric defined by item difficulty.
The test theory developed to account for sampling differences within domains can be
generalized to account for differences between domains. Just as different samples of
words will yield somewhat different estimates of vocabulary, different cognitive tasks
(e.g., vocabulary and arithmetic performance) will yield different estimates of
performance. Using multivariate procedures such as Principal Components Analysis
or Factor Analysis, it is possible to decompose the total variation into between domain
covariance, within domain covariance, and within domain variance. One of the most
replicable observations in the study of individual differences is that almost all tests
thought to assess cognitive ability have a general factor (g) that is shared with other
tests of ability. That is, although each test has specific variance associated with
content (e.g., linguistic, spatial), form of administration (e.g., auditory, visual), or
operations involved (e.g., perceptual speed, memory storage, memory retrieval,
abstract reasoning), there is general variance that is common to all tests of cognitive
ability.
ð
Although to some the term personality refers to all aspects of a person's individuality, typical
usage divides the field into studies of ability and personality. Tests of ability are viewed as
maximal performance measures. Ability is construed as the best one can do on a particular
measure in a limited time (speed test) or with unlimited time (power test). Personality measures
are estimates of average performance and typically include reports of preferences and estimates
of what one normally does and how one perceives oneself and is perceived by others.
The same procedures used to clarify the structure of cognitive abilities have been
applied to the question of identifying the domains of personality. Many of the early
and current personality inventories use self-descriptive questions (e.g., do you like to
go to lively parties; are you sometimes nervous) that are rationally or theoretically
relevant to some domain of interest for a particular investigator. Although there is
substantial consistency across inventories developed this way, some of this agreement
could be due to conceptually overlapping item pools. Other researchers have
advocated a lexical approach to the taxonomic problem, following the basic
assumption that words in the natural language describe all important individual
differences. This shifts the taxonomic question from how are individuals similar and
different from each other to how are the words used to describe individuals (e.g.,
lively, talkative, nervous, anxious) similar and different from each other.
Measures of ability and personality reflect observations aggregated across time and
occasion and require inferences about stable latent traits thought to account for the
variety of observed behaviors. However there are other individual differences that are
readily apparent to outside observers and require little or no inference about latent
traits. The most obvious of such variables include sex, age, height, and weight.
Differences that require some knowledge and inference are differences in ethnicity
and social economic status. These obvious group differences are sometimes analyzed
in terms of the more subtle measures of personality and ability or of real life outcomes
(e.g, sex differences in neuroticism, mathematics ability, or income).
ð
Individual differences are important only to the extent that they make a difference.
Does knowing that people differ on a trait X help in predicting the likelihood of their
doing behavior Y? For many important outcome variables the answer is a resounding
yes. In their review of 85 years of selection in personnel psychology, Frank Schmidt
and John Hunter (Psychological Bulletin, 1998, 124, 262-274) show how differences
in cognitive ability predict differences in job performance with correlations averaging
about .50 for mid complexity jobs. These correlations are moderated by job
complexity and are much higher for professional-managerial positions than they are
for completely unskilled jobs. In terms of applications to personnel psychology, a
superior manager (one standard deviation above the mean ability for managers)
produces almost 50% more than an average manager. These relationships diminish as
a function of years of experience and degree of training. General mental ability (g)
also has substantial predictive powers in predicting non-job related outcomes, such as
likelihood of completing college, risk for divorce and even risk for criminality.
The non-cognitive measures of individual differences also predict important real life
criteria. Extraversion is highly correlated with total sales in dollars among
salespeople. Similarly, impulsivity can be used to predict traffic violations.
Conscientiousness, when added to g substantially increases the predictability of job
performance. Although the size of the correlation is much lower, conscientiousness
measured in adolescence predicts premature mortality over the next fifty years.
The taxonomic and predictive studies of individual differences are descriptive organizations of
thoughts, feelings, and behaviors that go together and how they relate to other outcomes. But this
categorization is descriptive rather than causal and is analogous to grouping rocks in terms of
density and hardness rather than atomic or molecular structure. Causal theories of individual
differences are being developed but are in a much earlier stage than are the descriptive
taxonomies.
Descriptive taxonomies are used to organize the results of studies that examine
genetic bases of individual differences. By applying structural modeling techniques to
the variances and covariances associated with various family constellations it is
possible to decompose phenotypic trait variance into separate sources of genetic and
environmental variance. The most common family configurations that are used are
comparisons of identical (monozygotic) with fraternal (dizygotic) twins. Additional
designs include twins reared together or apart, and biological versus adoptive parents,
children and siblings. Conclusions from behavioral genetics for most personality traits
tend to be similar: Across different designs, with different samples from different
countries, roughly 40-60% of the phenotypic variance seems to be under genetic
control with only a very small part of the remaining environmental variance
associated with shared family environmental effects. Additional results suggest that
genetic sources of individual differences remain important across the lifespan.
However, this should not be taken to mean that people do not change as they mature
but rather that the paths one takes through life are similar to those taken by genetically
similar individuals.
Genes do not code for thoughts, feelings or behavior but rather code for proteins that
regulate and modulate biological systems. Although promising work has been done
searching for the biological bases of individual differences it is possible to sketch out
these bases only in the broadest of terms. Specific neurotransmitters and brain
structures can be associated with a broad class of approach behaviors and positive
affects while other neurotransmitters and structures can be associated with a similarly
broad class of avoidance behaviors and negative affects. Reports relating specific
alleles to specific personality traits emphasize that the broad personality traits are
most likely under polygenic influence and are moderated by environmental
experience.
O
O
O
O
!
O
O
O
c
h
!
"
#
$
%&
' ! (# !
%&
!
' !
)
# !
"
!
%&
#
# # #
%&
)
*
a å
' !+
,-
(# !
)
a )
)
.
' / 0 ")
1
2
%&
%
+%&-
%&
!
+
34
-
+
33
-
h
1 %&
!
"
"
!
'
1 *5
6h
'
71
A test in which questions are written according to specific predetermined criteria. A
student knows what your standards are for passing and only competes against him or herself
while completing the test.
Ä
" #
Ä O$
%
j
Ä
"
&
Ä O
O
O
school
!
v ' education
Ä O
(
%
'
) "
*
Ä O
+
O +
%
O
) "
*
h
ã O
"
"
O Ä O
O Ä O
+
O
," ,
O
%
%
-
Ä O
'
"
,,
"
*
.)
O
/
0 Ä
"
"
-
1 O
!
O
'
)2
)2
3
O
1 O
O Ä O
O
) O
#
$
+
-
!
O
!
/
O *
&
-
-
"
%
4
,*
-
5O 3 /
!
-
6
/-
'
- *
4,
5
7
7ã8 /
&
9Schooling6
!
"
¢
There are two ways that reliability is usually estimated: test/retest and internal
consistency.
O¢
Test/retest is the more conservative method to estimate reliability. Simply put, the
idea behind test/retest is that you should get the same score on test 1 as you do on test
2. The three main components to this method are as follows:
1.) implement your measurement instrument at two separate times for each
subject;
2). compute the correlation between the two separate measurements; and
3) assume there is no change in the underlying condition (or trait you are trying
to measure) between test 1 and test 2.
2
Internal consistency estimates reliability by grouping questions in a questionnaire that
measure the same concept. For example, you could write two sets of three questions
that measure the same concept (say class participation) and after collecting the
responses, run a correlation between those two groups of three questions to determine
if your instrument is reliably measuring that concept.
One common way of computing correlation values among the questions on your
instruments is by using Cronbach's Alpha. In short, Cronbach's alpha splits all the
questions on your instrument every possible way and computes correlation values for
them all (we use a computer program for this part). In the end, your computer output
generates one number for Cronbach's alpha - and just like a correlation coefficient, the
closer it is to one, the higher the reliability estimate of your instrument. Cronbach's
alpha is a less conservative estimate of reliability than test/retest.
O
There are four types of validity commonly examined in social research.
1. Conclusion validity asks is there a relationship between the program and the
observed outcome? Or, in our example, is there a connection between the
attendance policy and the increased participation we saw?
2. Internal alidity asks if there is a relationship between the program and the
outcome we saw, is it a causal relationship? For example, did the attendance
policy cause class participation to increase?
4. External validity refers to our ability to generalize the results of our study to
other settings. In our example, could we generalize our results to other
classrooms?
O O2
There are three main types of threats to internal validity - single group, multiple group
and social interaction threats.
O apply when you are studying a single group receiving a
program or treatment. Thus, all of these threats can be greatly reduced by adding a
control group that is comparable to your program group to your study.
A j
occurs when an historical event affects your program group such
that it causes the outcome you observe (rather than your treatment being the cause). In
our earlier example, this would mean that the stricter attendance policy did not cause
an increase in class participation, but rather, the expulsion of several students due to
low participation from school impacted your program group such that they increased
their participation as a result.
A
to internal validity occurs when standard events over the course
of time cause your outcome. For example, if by chance, the students who participated
in your study on class participation all "grew up" naturally and realized that class
participation increased their learning (how likely is that?) - that could be the cause of
your increased participation, not the stricter attendance policy.
A
to internal validity is simply when the act of taking a pre-test affects
how that group does on the post-test. For example, if in your study of class
participation, you measured class participation prior to implementing your new
attendance policy, and students became forewarned that there was about to be an
emphasis on participation, they may increase it simply as a result of involvement in
the pretest measure - and thus, your outcome could be a result of a testing threat - not
your treatment.
An
to internal validity could occur if the effect of increased
participation could be due to the way in which that pretest was implemented.
A
to internal validity occurs when subjects drop out of your study,
and this leads to an inflated measure of your effect. For example, if as a result of a
stricter attendance policy, most students drop out of a class, leaving only those more
serious students in the class (those who would participate at a high level naturally) -
this could mean your effect is overestimated and suffering from a mortality threat.
In sum, these single group threats must be addressed in your research for it to remain
credible. One primary way to accomplish this is to include a control group comparable
to your program group. This however, does not solve all our problems, as I'll now
highlight the multiple group threats to internal validity.
þ
O to internal validity involve the comparability of the two
groups in your study, and whether or not any other factor other than your treatment
causes the outcome. They also (conveniently) mirror the single group threats to
internal validity.
A
j threat occurs when an event occurring between the pre and post
test affects the two groups differently.
A
threat occurs when there are different rates of growth
between the two groups between the pre and post test.
threat is the result of the different effect from taking tests between
the two groups.
A
threat occurs when the test implementation affects the
groups differently between the pre and post test.
A
Threat occurs when there are different rates of dropout between
the groups which leads to you detecting an effect that may not actually occur.
Finally, a
threat occurs when the two groups regress towards the
mean at different rates.
Okay, so know that you have dragged yourself through these extensive lists of threats
to validity - you're wondering how to make sense of it all. How do we minimize these
threats without going insane in the process? The best advice I've been given is to use
two groups when possible, and if you do, make sure they are as comparable as is
humanly possible. Whether you conduct a randomized experiment or a non-random
study --> YOUR GROUPS MUST BE AS EQUIALENT AS POSSIBLE! This is
the best way to strengthen the internal validity of your research.The last type of threat
to discuss involves the social pressures in the research context that can impact your
results. These are known as
to internal validity.
occurs when the comparison group learns
about the program group and imitates them, which will lead to an equalization of
outcomes between the groups (you will not see an effect as easily).
means that the comparison group develops a competitive
attitude towards the program group, and this also makes it harder to detect an effect
due to your treatment rather than the comparison groups reaction to the program
group.
is a threat to internal validity that exaggerates the posttest
differences between the two groups. This is because the comparison group (upon
learning of the program group) gets discouraged and no longer tries to achieve on
their own.
is the only threat that is a result of the
actions of the research staff - it occurs when the staff begins to compensate the
comparison group to be "fair" in their opinion, and this leads to an equalization
between the groups and makes it harder to detect an effect due to your program.
I know, I know - you're thinking - no I just can't go on. Let's take a deep breath and I'll
remind you what construct validity is, and then we'll look at the threats to it one at a
time. OK? OK.
Constuct validity is the degree to which inferences we have made from our study can
be generalized to the concepts underlying our program in the first place. For example,
if we are measuring self-esteem as an outcome, can our definition (operationalization)
of that term in our study be generalized to the rest of the world's concept of self-
esteem?
Ok, let's address the threats to construct validity slowly - don't be intimidated by their
lengthy academic names - I'll provide an English translation.
simply means we did not
define our concepts very well before we measured them or implemented our
treatment. The solution? Define your concepts well before proceeding to the
measurement phase of your study.
simply means we only used one version of our independent
variable (our program or treatment) in our study, and hence, limit the breadth of our
study's results. The solution? Try to implement multiple versions of your program to
increase your study's utility.
simply put, means that you only used one measure or observation
of an important concept, which in the end, reduces the evidence that your measure is a
valid one. The solution? Implement multiple measures of key concepts and do pilot
studies to try to demonstrate that your measures are valid.
occurs when the testing in combination with
the treatment produces an effect. Thus you have inadequately defined your
"treatment," as testing becomes part of it due to its influence on the outcome. The
solution? Label your treatment accurately.
means that it was a combination of our treatment
and other things that brought about the effect. For example, if you were studying the
ability of Tylenol to reduce headaches and in actuality it was a combination of
Tylenol and Advil or Tylenol and exercise that reduced headaches -- you would have
an interaction of different treatments threatening your construct validity.
!
"
simply put, means that there were
some unanticipated effects from your program, that may make it difficult to say your
program was effective.
occurs when you are unable to detect an effect from your
program because you may have mislabeled your constructs or because the level of
your treatment wasn't enough to cause an effect.
As with internal validity, there are a few social threats to construct validity also. These
include:
See, that wasn't so bad. We broke things down and attacked them one at a time. You
may be wondering why I haven't given you along list of threats to conclusion and
external validity - the simple answer is it seems as if the more critical threats involve
internal and construct validity. And, the means by which we improve conclusion and
external validity will be highlighted in the section on Strengthening Your Analysis.
The real difference between reliability and validity is mostly a matter of definition.
Reliability estimates the consistency of your measurement, or more simply the degree
to which an instrument measures the same way each time it is used in under the same
conditions with the same subjects. alidity, on the other hand, involves the degree to
which your are measuring what you are supposed to, more simply, the accuracy of
your measurement. It is my belief that validity is more important than reliability
because if an instrument does not accurately measure what it is supposed to, there is
no reason to use it even if it measures consistently (reliably).
c &
' !
Description
O
a ¢
a ¢
a þ
2ndividual exercises
2
O
V °
°
2
V
V
!
°
V
!
°
"#
"
rne-to-one exercises
2
$
$
°
°
O
°
°
%
$
2
°
%
O
$
$
°
&
oroup exercises
o
°
'
O¢
(
O
°
°
°
)
°
$
°
'
°
O
°
¢
*
'°+,-.
$
/0$ $ 10
elf-assessment exercises
°
O
2
a 2
a
a 2
O
$
Development
3
°
°
%
%
2dentify criteria
2
3
%
4
$$
$$
O
%
Develop exercises
þ
O
3
2
5
°
2
$ °
°
°
O
°
elect assessors
*
%o
%°
$
þ
)
O
Run the assessment center
2
°
O
a 3
°
°
a 2
°
a
a
a *
a
a ( &
&
a O
a h
6
hollow-up
°
Discussion
2+,17
18
'
$+,/9°
098°
+,,9
:-8
V
O
°
$°
)
rrigins
O
O;O°
+ '
0 (
7 2$
. O
$
- <
%
: <
1 =
!
/
,
alidity
¢
°
+,::
O
Criticisms
O
%
%>
%%
°
=
!
O
.$-
þ
?
)
O
'
3:
2
&
&
-
#&&-$
9
& O9
3/-
&
6
; <*
1
=
-
ã
ã )
O
)
O 9
6
O
6
> &
%
O
&
&
9 ?@ '
, )+ -
O
'
6
/
'
O
5
a
a
-
%
%
#
9A-"
%
$
a
'
a
'
a
/ "
a
'
7%
,) . .**'.* *
'
*
+*
* *
' /**'0
&&-
;
O
&&-
O
O
!
O
:
*
;12"
3
23
3
3
3+ - 3 +4 3
5
)* *
' 2
'
+
O
7
)
6
a
2726B
;
"
a
**
6B
/
a
+6B
;
A cr itical component of validity is rei iity. alidity embodies not only what positive
outcomes a selection approach may predict, but also how consistenty(i.e., reliably) it does
so. In this chapter we will (1) r eview ways of improving the consistency or reliability of
the selection process; (2) discuss two methods for measuring validity; and (3) pr esent two
cases that illustrate these methods. First, however, let¶s consider a legal issue that is
closely connected to validity: employment discrimination.
In psychology, a projective test is a type of personality test in which the individual offers
responses to ambiguous scenes, words or images. This type of test emerged from the
psychoanalytic school of thought, which suggested that people have unconscious thoughts or
urges. These projective tests were intended to uncover such unconscious desires that are hidden
from conscious awareness.
jðO
In many projective tests, the participant is shown an ambiguous image and then asked to give the
first response that comes to mind. The key to projective tests is the ambiguity of the stimuli.
According to the theory behind such tests, clearly defined questions result in answers that are
carefully crafted by the conscious mind. By providing the participant with a question or stimulus
that is not clear, the underlying and unconscious motivations or attitudes are revealed.
There are a number of different types of projective tests. The following are just a few examples
of some of the best-known projective tests.
a OO O!OO"
In the Thematic Apperception Test, an individual is asked to look at a series of ambiguous
scenes. The participant is then asked to tell a story describing the scene, including what is
happening, how the characters are feeling and how the story will end. The examiner then
scores the test based on the needs, motivations and anxieties of the main character as well as
how the story eventually turns out.
ðO
Projective tests are most frequently used in therapeutic settings. In many cases, therapists use
these tests to learn qualitative information about a client. Some therapists may use projective
tests as a sort of icebreaker to encourage the client to discuss issues or examine thoughts and
emotions.
While projective tests have some benefits, they also have a number of weaknesses and
limitations. For example, the respondent's answers can be heavily influenced by the examiner's
attitudes or the test setting. Scoring projective tests is also highly subjective, so interpretations of
answers can vary dramatically from one examiner to the next.
Additionally, projective tests lack both validity and reliability. alidity refers to whether or not a
test is measuring what it purports to measure, while reliability refers to the consistency of the test
results.
The
of a chart is the ratio of a given distance on the chart to the actual distance which it
represents on the earth. It may be expressed in various ways. The most common are:
1. A simple ratio or fraction, known as the . For example, 1:80,000
or 1/80,000 means that one unit (such as a meter) on the chart represents 80,000 of the
same unit on the surface of the earth. This scale is sometimes called
the
or
scale.
2. A that a given distance on the earth equals a given measure on the chart, or
vice versa. For example, ³30 miles to the inch´ means that 1 inch on the chart represents
30 miles of the earth¶s surface. Similarly, ³2 inches to a mile´ indicates that 2 inches on
the chart represent 1 mile on the earth. This is sometimes called the
.
3. A line or bar called a
may be drawn at a convenient place on the chart and
subdivided into nautical miles, meters, etc.
All charts vary somewhat in scale from point to point, and in some projections the scale is not
the same in all directions about a single point. A single subdivided line or bar for use over an
entire chart is shown only when the chart is of such scale and projection that the scale varies a
negligible amount over the chart, usually one of about 1:75,000 or larger. Since 1 minute of
latitude is very nearly equal to 1 nautical mile, the latitude scale serves as an approximate
graphic scale.
On most nautical charts the east and west borders are subdivided to facilitate distance
measurements. On a Mercator chart the scale varies with the latitude. This is noticeable on a
chart covering a relatively large distance in a north-south direction. On such a chart the border
scale near the latitude in question should be used for measuring distances.
Of the various methods of indicating scale, the graphical method is normally available in some
form on the chart. In addition, the scale is customarily stated on charts on which the scale does
not change appreciably over the chart. The ways of expressing the scale of a chart are readily
interchangeable. For instance, in a nautical mile there are about 72,913.39 inches. If the natural
scale of a chart is 1:80,000, one inch of the chart represents 80,000 inches of the earth, or a little
more than a mile. To find the exact amount, divide the scale by the number of inches in a mile,
or 80,000/72,913.39 = 1.097. Thus, a scale of 1:80,000 is the same as a scale of 1.097 (or
approximately 1.1) miles to an inch.
Stated another way, there are: 72,913.39/80,000 = 0.911 (approximately 0.9) inch to a mile.
Similarly, if the scale is 60 nautical miles to an inch, the representative fraction is 1:(60 x
72,913.39) = 1:4,374,803.
A chart covering a relatively large area is called a
#
and one covering a
relatively small area is called a
#
. Since the terms are relative, there is no sharp
division between the two. Thus, a chart of scale 1:100,000 is large scale when compared with a
chart of 1:1,000,000 but small scale when compared with one of 1:25,000.
As scale decreases, the amount of detail which can be shown decreases also. Cartographers
selectively decrease the detail in a process called
$ when producing small scale
charts using large scale charts as sources. The amount of detail shown depends on several
factors, among them the coverage of the area at larger scales and the intended use of the chart.
) +
Ä
+
+
Ä
+
#O
"
/ $/
/
+
"
/
O
/
5
a 9
+
a E
&
+
-+
/
F
a D
+
E
,'
,
a '
O
+
)
4
-
!%"
A
*
%
B
%
%
O
5
a Ä
+
a E
a E
a
'
/
G
+
B
+
>
"
O
*
5
a
a B
+
'
%
)7 +
O
'
,
,
O
"+
O
O
"
5
a Ä
+
B
"%
a
*
'
%
a -"
+
-
)
-
E
!
*
>
>
O
%>
"
'
-
D
"
5
a
'
(
a &
!
'
a *
a E
'
)1
6
'
!
B
%
+
"
"
>
>+
!
O
!
5
a -
+
+
%
a
- +
a
+
>
%+
a &
:
)'
-
O
%
'
+
+
O
!
5
)
'
O
"
O
5
+ O
+
O
5
a 4
'
a O
a -
a :
))+")
*
6E
56E
/
B
Ä
O
1
'
'
""
/
!
5
)7
'
%
% 6
"
6
&
%+
-
Ä
/
5
a O
B
4
'
'
!
B
'
a '
'
a O
'
'
-
5,'
,,/
,
a Ä
-
+
a
*
a &
a O
)<"
Ä
/
%
!
/
"
4
O
>
/
>
-
/
5
a D
-
a &
a
a D
5
+
c 6
0=
D
°
%
)
%
2
°
%
a
%
O %
%
a %
%
a D
%
a D
%
°
2
a O %
a D
%
#
%
%
a 2
%
%
%
a O
a D
%
%
2
%
° %
°
O
°
*
%
%
3
"2
%
° °
"*
D
%°
D
*
O
D
@D
$
@þ
¢
$'*2
70-.0
'*2
$70-0,?A
%
B
2
D
%
°
%
°
O
%
O %
Explanation:
D
%
þ
%
°
°
°
%
2
!
2
°
2llustration:
O O
*
3
%
D
o
þ
2
%
°
°
O
-
-
H
%&
%
+
-
%O '1!#
$O
5(
D
v8G*
IÄ
2
O ?
O ? ?
O
%J?
K6
0* +