Vous êtes sur la page 1sur 22

SAMPLING, DATA COLLECTION METHODS, STATISTICAL TECHNIQUES

AND HYPOTHESIS TESTING

Sampling Design

Surveys can gather information through different methods of observation. However, most
surveys employ a questionnaire to measure specific characteristics of the population. There
are two main ways to gather this information.

A census survey collects complete information from all participants in the population.
General criteria of a census survey include:

 Establish and maintain a complete list of the primary sampling unit (PSU)
components.
 All members of the PSU in must be included
 Validation (used to correct for missing and misreported data)
 Enforceable and enforced

A sample survey uses a representative group of a given population to determine


characteristics of the entire population. General requirements of a sample survey include:

 The sample must be representative of the entire population.


 The data collected from the representative sample is expanded to produce an estimate
of the total population.
 The estimate has two parts: Point estimate and margin of error.
 Assumptions should be tested to identify and measure potential bias if possible.

Importance of sampling survey

Sample surveys are used when it is not possible or practical to conduct a census to count each
individual of an entire population. Everyday examples of sample surveys include political
polling, health statistics, and television ratings. Sample surveys are a proven, effective
method for gathering accurate information if they are properly designed and the sample
design is accounted for in the estimation methods.

Sampling method refers to the rules and procedures by which some elements of the
population are included in the sample.

Sampling Estimator. The estimation process for calculating sample statistics is called the
estimator. Different sampling methods may use different estimators.

Objectives of sampling:

(i) To obtain the optimum results, i.e., the maximum information about the characteristics of
the population with the available sources at our disposal in terms of time, money and
manpower by studying the sample values only.

(ii) To obtain the best possible estimates of the population parameters.


Characteristics of a good sample

1. Representativeness

2. Accuracy

3. Precision

4. Size

Process of Sampling Design

Sampling frame: In statistics, a sampling frame is the source material or device from which a
sample is drawn. It is a list of all those within a population who can be sampled, and may include
individuals, households or institutions.

Sampling unit: A unit is the thing being studied. Usually in social research this is people. There may
also be additional selection criteria used to choose the units to study, such as 'people who have been
police officers for at least five years.'

Sample size: The larger your sample size, the more sure you can be that their answers truly reflect the
population. This indicates that for a given confidence level, the larger your sample size, the smaller
your confidence interval.
Determining sample size is a very important issue because samples that are too large may
waste time, resources and money, while samples that are too small may lead to inaccurate
results.

Sampling Instrument: A questionnaire is a research instrument consisting of a series of questions and


other prompts for the purpose of gathering information from respondents.

Questionnaire and Schedule

Basis for
Questionnaire Schedule
Comparison
Schedule is a formalized set of questions,
Questionnaire refers to a technique
statements and spaces for answers,
of data collection which consist of
Meaning provided to the enumerators who ask
a series of written questions along
questions to the respondents and note
with alternative answers.
down the answers.
Filled by Respondents Enumerators
Response Rate Low High
Coverage Large Comparatively small
Cost Economical Expensive
Respondent's
Not known Known
identity
Success relies Honesty and competence of the
Quality of the questionnaire
on enumerator.
Only when the people are literate
Usage Used on both literate and illiterate people.
and cooperative.

Other methods of collecting data through primary data collection are observation,
experimentation, simulation, panel methods and interview methods.

Sampling Techniques:
Probability or Random Sampling Techniques:

A probability sampling method is any method of sampling that utilizes some form of random
selection. In order to have a random selection method, you must set up some process or procedure that
assures that the different units in your population have equal probabilities of being chosen.

1. Simple Random Sampling: The simplest form of random sampling is called simple random
sampling. Simple random sampling is not the most statistically efficient method of sampling and you
may, just because of the luck of the draw, not get good representation of subgroups in a population.

2. Stratified Random Sampling, also sometimes called proportional or quota random sampling,
involves dividing your population into homogeneous subgroups and then taking a simple random
sample in each subgroup.

a. Proportionate stratification is a type of stratified sampling . With proportionate stratification, the


sample size of each stratum is proportionate to the population size of the stratum. This means that
each stratum has the same sampling fraction .

b. Disproportionate stratification is a type of stratified sampling . With disproportionate stratification,


the sample size of each stratum does not have to be proportionate to the population size of the stratum.
This means that two or more strata will have different sampling fractions .

3. Systematic sampling is a type of probability sampling method in which sample members


from a larger population are selected according to a random starting point and a fixed
periodic interval. This interval, called the sampling interval, is calculated by dividing the
population size by the desired sample size. Despite the sample population being selected in
advance, systematic sampling is still thought of as being random if the periodic interval is
determined beforehand and the starting point is random.

4. Cluster Sampling: Cluster sampling (also known as one-stage cluster sampling) is a


technique in which clusters of participants that represent the population are identified and
included in the sample. Cluster involves cluster of participants that represent the population
are identified and included in the sample. This is popular in conducting marketing
researches. The main aim of cluster sampling can be specified as cost reduction and
increasing the levels of efficiency of sampling. This specific technique can also be applied in
integration with multi-stage sampling. A major difference between cluster and stratified
sampling relates to the fact that in cluster sampling a cluster is perceived as a sampling unit,
whereas in stratified only specific elements of strata are accepted as sampling unit.
Accordingly, in cluster sampling a complete list of clusters represent the sampling frame.
Then, a few clusters are chosen randomly as the source of primary data.
Non Probability or Non-Random Sampling Techniques:

1. Purposive Sampling: In purposive sampling, we sample with a purpose in mind. We


usually would have one or more specific predefined groups we are seeking. For instance,
have you ever run into people in a mall or on the street who are carrying a clipboard and who
are stopping various people and asking if they could interview them? Most likely they are
conducting a purposive sample (and most likely they are engaged in market research.)

2. Snowball Sampling: In snowball sampling, the researcher begins by identifying someone


who meets the criteria for inclusion in your study. You then ask them to recommend others
who they may know who also meet the criteria. Although this method would hardly lead to
representative samples, there are times when it may be the best method available. Snowball
sampling is especially useful when you are trying to reach populations that are inaccessible or
hard to find. For instance, if you are studying the homeless, you are not likely to be able to
find good lists of homeless people within a specific geographical area. However, if you go to
that area and identify one or two, you may find that they know very well who the other
homeless people in their vicinity are and how you can find them.

3. Convenience sampling is sometimes referred to as haphazard or accidental sampling. It is


not normally representative of the target population because sample units are only selected if
they can be accessed easily and conveniently.

4. Quota sampling: This is one of the most common forms of non-probability sampling.
Sampling is done until a specific number of units (quotas) for various sub-populations have
been selected. Since there are no rules as to how these quotas are to be filled, quota sampling
is really a means for satisfying sample size objectives for certain sub-populations.

Sampling errors and Non sampling errors

Sampling error: Sampling error is the deviation of the selected sample from the true characteristics,
traits, behaviors, qualities or figures of the entire population. The errors which arise because of
studying only a part of the total population.

Non Sampling error: These are errors which arise from sources other than sampling. The
errors of observation, errors of measurement and errors of responses are some non-sampling
errors.

Methods of Data collection

Data Collection is an important aspect of any type of research study. Inaccurate data
collection can impact the results of a study and ultimately lead to invalid results. Data
collection methods for impact evaluation vary along a continuum. At the one end of this
continuum are quantitative methods and at the other end of the continuum are Qualitative
methods for data collection.

Primary data is an information collected from sources such as personal interviews,


questionnaires or surveys with a specific intention and on a specific subject, and observation
and discussion by the researcher him or herself, which information is then assessed by that
person. It is a direct approach and, as it is tailored to a company's particular needs, reveals
apparently, much-needed information to that company which started the research; that is, the
results are used for the purpose for which they were originally intended. It can be a lengthy
process but does provide first-hand information.

Secondary data is information that is already available somewhere, whether it be in journals,


on the internet, in a company's records or, on a larger scale, in corporate or governmental
archives. Secondary data allows for comparison of, say, several years worth of statistical
information relating to, for example, a sector of the economy, where the information may be
used to measure the effects of change or whatever it is that is being researched.
Advantages of Primary Data collection

1) Targeted Issues are addressed. The organization asking for the research has the complete
control on the process and the research is streamlines as far as its objectives and scope is
concerned. Researching company can be asked to concentrate their efforts to find data
regarding specific market rather than concentration on mass market.

2) Data interpretation is better. The collected data can be examined and interpreted by the
marketers depending on their needs rather than relying on the interpretation made by
collectors of secondary data.

3) Recency of Data. Usually secondary data is not so recent and it may not be specific to the
place or situation marketer is targeting. The researcher can use the irrelevant seeming
information for knowing trends or may be able to find some relation with the current
scenario. Thus primary data becomes a more accurate tool since we can use data which is
useful for us.

4) Proprietary Issues. Collector of primary data is the owner of that information and he need
not share it with other companies and competitors. This gives an edge over competitors
replying on secondary data.

Disadvantages of Primary Data collection

1) High Cost. Collecting data using primary research is a costly proposition as marketer has
to be involved throughout and has to design everything.

2) Time Consuming. Because of exhaustive nature of the exercise, the time required to do
research accurately is very long as compared to secondary data, which can be collected in
much lesser time duration.

3) Inaccurate Feed-backs. In case the research involves taking feedbacks from the targeted
audience, there are high chances that feedback given is not correct. Feedbacks by their basic
nature are usually biased or given just for the sake of it.

4) More number of resources are required. Leaving aside cost and time, other resources
like human resources and materials too are needed in larger quantity to do surveys and data
collection.

Advantages of Secondary Data collection

1. The first advantage of using secondary data (SD) has always been the saving of time. New
technology has revolutionized this world. The process has been simplified. Precise
information may be obtained via search engines. All worth library has digitized its collection
so that students and researchers may perform more advance searches.

2. Strongly connected to the previous advantages is the saving of money.


3. Generating new insights from previous analyses.

Disadvantages of Secondary Data collection

1. Inappropriateness of the data. Data collected by oneself (primary data) is collected with
a concrete idea in mind. Usually to answer a research question or just meet certain
objectives. In this sense, secondary data sources may provide you with vast amount of
information, but quantity is not synonymous of appropriateness. This is simply because it has
been collected to answer a different research question or objectives.

2. Lack of control over data quality: Government and other official institutions are often a
guarantee of quality data, but it is not always the case.

Primary Data collection Methods:

1. Observation – Participant observation, non-participant observation, direct observation,


indirect observation, controlled and uncontrolled observation

2. Experimentation – Laboratory and Field experiment

3. Simulation – Man simulation, computer simulation and man-computer simulation

4. Interview – structured or directive interview, unstructured or non-directive interview,


focused interview, clinical interview, depth interview, telephonic interview and group
interview

5. Mail survey

6. Panel method

7. Projective techniques – Rorschach test, Thematic Apperception Test (TAT), Picture


frustration test, Holtzman Inkblot test and Word Association test.

Tools for Data collection

1. Questionnaire

2. Check list

3. Rating scale

4. Inventories

5. Interview schedule

6. Interview guide

7. Observation schedule
Measurement and Scaling Techniques

1. Nominal Scale: Nominal scales are used for labelling variables, without any quantitative
value. “Nominal” scales could simply be called “labels.” Here are some examples,
below. Notice that all of these scales are mutually exclusive (no overlap) and none of them
have any numerical significance. A good way to remember all of this is that “nominal”
sounds a lot like “name” and nominal scales are kind of like “names” or labels.

2. Ordinal Scale: With ordinal scales, it is the order of the values is what‟s important and
significant, but the differences between each one is not really known.

3. Interval scales are numeric scales in which we know not only the order, but also the exact
differences between the values.

The classic example of an interval scale is Celsius temperature because the difference
between each value is the same. For example, the difference between 60 and 50 degrees is a
measurable 10 degrees, as is the difference between 80 and 70 degrees. Time is another good
example of an interval scale in which the increments are known, consistent, and measurable.

Interval scales are nice because the realm of statistical analysis on these data sets opens
up. For example, central tendency can be measured by mode, median, or mean; standard
deviation can also be calculated.

4. Ratio scales are the ultimate nirvana when it comes to measurement scales because they
tell us about the order, they tell us the exact value between units, AND they also have an
absolute zero–which allows for a wide range of both descriptive and inferential statistics to be
applied.

Arbitrary scale, Rensis Likert Scale, Semantic Differential Scale (SD), Factor scale, Attitude
scales are some of the common scales used to collect data.

Scales and its Measurement

Order Distance Origin


Nominal Scale X X X
Ordinal Scale YES X X
Interval Scale YES YES X
Ration Scale YES YES YES
HYPOTHESIS & TESTING OF HYPOTHESIS

It is a supposition or proposed explanation made on the basis of limited evidence as a starting


point for further investigation.

Types of hypothesis

1. Simple hypothesis – it is that one in which there exists relationship between two variables
(causation effect – cause and effect) Ex: Smoking leads to cancer

2. Complex hypothesis – It is that one in which as relationship among variables exist.

Ex: Smoking and other drugs lead to cancer and chest infection etc.

3. Empirical / Working hypothesis – It is that one which is applied to a field.

4. Null hypothesis (Ho)– It is contrary to the positive working hypothesis. There is no


relationship between variables.

5. Alternate hypothesis (H1) – The workable and most efficient hypothesis.

6. Logical hypothesis – Is is that one which is verified logically. We may have agreement,
disagreement, difference and residue as the outcomes.

7. Statistical hypothesis – which should be tested statistically.

Type I and Type II Errors (Alpha and Beta error)

When you do a hypothesis test, two types of errors are possible: type I and type II. The risks
of these two errors are inversely related and determined by the level of significance and the
power for the test. Therefore, you should determine which error has more severe
consequences for your situation before you define their risks.

Type I error - When the null hypothesis is true and you reject it, you make a type I error.
The probability of making a type I error is α, which is the level of significance you set for
your hypothesis test. An α of 0.05 indicates that you are willing to accept a 5% chance that
you are wrong when you reject the null hypothesis. To lower this risk, you must use a lower
value for α. However, using a lower value for alpha means that you will be less likely to
detect a true difference if one really exists.

Type II error - When the null hypothesis is false and you fail to reject it, you make a type II
error. The probability of making a type II error is β, which depends on the power of the test.
You can decrease your risk of committing a type II error by ensuring your test has enough
power. You can do this by ensuring your sample size is large enough to detect a practical
difference when one truly exists.
Type I and Type II Errors
PROCEDURE FOR HYPOTHESIS TESTING

There are five steps in hypothesis testing:

1. Making assumptions
2. Stating the research and null hypotheses and selecting (setting) alpha
3. Selecting the sampling distribution and specifying the test statistic
4. Computing the test statistic
5. Making a decision and interpreting the results

10 Key concepts in Hypothesis Testing:

1. Null hypothesis - The null hypothesis is a clear statement about the relationship between
two (or more) statistical objects. These objects may be measurements, distributions, or
categories. Typically, the null hypothesis, as the name implies, states that there is no
relationship.In the case of two population means, the null hypothesis might state that the
means of the two populations are equal.

2. Alternative hypothesis - Once the null hypothesis has been stated, it is easy to construct
the alternative hypothesis. It is essentially the statement that the null hypothesis is false. In
our example, the alternative hypothesis would be that the means of the two populations are
not equal.

3. Significance - The significance level is a measure of the statistical strength of the


hypothesis test. It is often characterized as the probability of incorrectly concluding that the
null hypothesis is false.

The significance level is something that you should specify up front. In applications, the
significance level is typically one of three values: 10%, 5%, or 1%. A 1% significance level
represents the strongest test of the three. For this reason, 1% is a higher significance level
than 10%

4. Power - Related to significance, the power of a test measures the probability of correctly
concluding that the null hypothesis is true. Power is not something that you can choose. It is
determined by several factors, including the significance level you select and the size of the
difference between the things you are trying to compare.

Unfortunately, significance and power are inversely related. Increasing significance decreases
power. This makes it difficult to design experiments that have both very high significance
and power.

5. Test statistic - The test statistic is a single measure that captures the statistical nature of
the relationship between observations you are dealing with. The test statistic depends
fundamentally on the number of observations that are being evaluated. It differs from
situation to situation.

6. Distribution of the test statistic - The whole notion of hypothesis rests on the ability to
specify (exactly or approximately) the distribution that the test statistic follows. In the case of
this example, the difference between the means will be approximately normally distributed
(assuming there are a relatively large number of observations).

7. One-tailed vs. two-tailed tests - Depending on the situation, you may want (or need) to
employ a one- or two-tailed test. These tails refer to the right and left tails of the distribution
of the test statistic. A two-tailed test allows for the possibility that the test statistic is either
very large or very small (negative is small). A one-tailed test allows for only one of these
possibilities.

8. Critical value - The critical value in a hypothesis test is based on two things: the
distribution of the test statistic and the significance level. The critical value(s) refer to the
point in the test statistic distribution that give the tails of the distribution an area (meaning
probability) exactly equal to the significance level that was chosen.

9. Decision - Your decision to reject or accept the null hypothesis is based on comparing the
test statistic to the critical value. If the test statistic exceeds the critical value, you should
reject the null hypothesis. In this case, you would say that the difference between the two
population means is significant. Otherwise, you accept the null hypothesis.

10. P-value - The p-value of a hypothesis test gives you another way to evaluate the null
hypothesis. The p-value represents the highest significance level at which your particular test
statistic would justify rejecting the null hypothesis. For example, if you have chosen a
significance level of 5%, and the p-value turns out to be .03 (or 3%), you would be justified
in rejecting the null hypothesis.

STATISTICAL TECHNIQUES

Correlation – It is a statistical measure that indicates the extent to which two or more
variables fluctuate together. A positive correlation indicates the extent to which those
variables increase or decrease in parallel; a negative correlation indicates the extent to which
one variable increases as the other decreases.

Types of correlation

1. Positive correlation occurs when an increase in one variable increases the value in
another. The line corresponding to the scatter plot is an increasing line.

2. Negative correlation occurs when an increase in one variable decreases the value of
another. The line corresponding to the scatter plot is a decreasing line.

3. Perfect correlation occurs when there is a functional dependency between the variables.
In this case all the points are in a straight line.

Methods of finding correlation coefficient:

1. Karl Pearson‟s correlation coefficient

2. Spearman‟s Rank Correlation


Correlation values and its interpretation: The value of correlation always lies between -1 to +1.

Correlation value Interpretation


0 No Correlation
0 to 0.4 Low Positive Correlation
0 to -0.4 Low Negative Correlation
0.4 to 0.7 Moderate Positive Correlation
-0.4 to -0.7 Moderate Negative Correlation
0.7 to 0.9 High Positive Correlation
-0.7 to -0.9 High Negative Correlation
More than 0.9 Very high Positive Correlation
More than -0.9 Very high Negative Correlation
+1 Perfect Positive Correlation
-1 Perfect Negative Correlation

Regression

Simple linear regression is a statistical method that allows us to summarize and study
relationships between two continuous (quantitative) variables:

 One variable, denoted x, is regarded as the predictor, explanatory, or independent


variable.
 The other variable, denoted y, is regarded as the response, outcome, or dependent
variable.

Because the other terms are used less frequently today, we'll use the "predictor" and
"response" terms to refer to the variables encountered in this course. The other terms are
mentioned only to make you aware of them should you encounter them in other arenas.
Simple linear regression gets its adjective "simple," because it concerns the study of only one
predictor variable. In contrast, multiple linear regression, which we study later in this course,
gets its adjective "multiple," because it concerns the study of two or more predictor variables.

Regression coefficient : The constant „b‟ in the regression equation (Ye = a + bX) is called as
the Regression Coefficient. It determines the slope of the line, i.e. the change in the value of
Y corresponding to the unit change in X and therefore, it is also called as a “Slope
Coefficient.

Properties of Regression coefficients:

1. The correlation coefficient is the geometric mean of two regression coefficients.

Symbolically, it can be expressed as:

2. The value of the coefficient of correlation cannot exceed unity i.e. 1. Therefore, if one of
the regression coefficients is greater than unity, the other must be less than unity.
3. The sign of both the regression coefficients will be same, i.e. they will be either positive
or negative. Thus, it is not possible that one regression coefficient is negative while the other
is positive.

4. The coefficient of correlation will have the same sign as that of the regression
coefficients, such as if the regression coefficients have a positive sign, then “r” will be
positive and vice-versa.

5. The average value of the two regression coefficients will be greater than the value of
the correlation. Symbolically, it can be represented as

6. The regression coefficients are independent of the change of origin, but not of the scale

Difference between correlation and Regression


Chi Square test:

The Chi-Square test of Independence is used to determine if there is a significant relationship


between two nominal (categorical) variables. The frequency of one nominal variable is
compared with different values of the second nominal variable. The data can be displayed in
an R*C contingency table, where R is the row and C is the column. For example, a
researcher wants to examine the relationship between gender (male vs. female) and empathy
(high vs. low). The chi-square test of independence can be used to examine this
relationship. If the null hypothesis is accepted there would be no relationship between gender
and empathy. If the null hypotheses is rejected the implication would be that there is a
relationship between gender and empathy (e.g. females tend to score higher on empathy and
males tend to score lower on empathy).

PARAMETRIC AND NON-PARAMETRIC TESTS

Parametric Tests: A parametric statistical test is one that makes assumptions about the
parameters (defining properties) of the population distribution(s) from which one's data are
drawn.

Assumptions are the underlying source population(s) to be normally distributed; they


generally also assume that one's measures derive from an equal-interval scale.

Parametric tests are having some advantages and disadvantages. They are,

Advantages :

1. Don‟t require data


2. Quite easy to calculate them
3. They give you all necessary information

Disadvantages :

1. They are not valid


2. The size of the sample is always very big

Non-parametric Tests: A non-parametric test is one that makes no such assumptions. In


this strict sense, "non-parametric" is essentially a null category, since virtually all statistical
tests assume one thing or another about the properties of the source population.
Difference between Parametric and Non-parametric tests
Parametric Non-parametric
Assumed distribution Normal Any
Assumed variance Homogeneous Any
Typical data Ratio or Interval Ordinal or Nominal
Data set relationships Independent Any
Usual central measure Mean Median
Can draw more Simplicity; Less
Benefits
conclusions affected by outliers
Tests
Choosing parametric Choosing a non-
Choosing
test parametric test
Correlation test Pearson Spearman
Independent measures, Independent-measures
Mann-Whitney test
2 groups t-test
Independent measures, One-way, independent-
Kruskal-Wallis test
>2 groups measures ANOVA
Repeated measures, 2
Matched-pair t-test Wilcoxon test
conditions
Repeated measures, >2 One-way, repeated
Friedman's test
conditions measures ANOVA

Parametric tests (means) Nonparametric tests (medians)


1-sample t test 1-sample Sign, 1-sample Wilcoxon
2-sample t test Mann-Whitney test
One-Way ANOVA Kruskal-Wallis, Mood‟s median test
Factorial DOE with one factor and one
Friedman test
blocking variable

Student t tests:

Introduced by William Gosset in 1908. A t-test is any statistical hypothesis test in which the
test statistic follows a Student's t-distribution under the null hypothesis. It can be used to
determine if two sets of data are significantly different from each other.

One sample t test: In testing the null hypothesis that the population mean is equal to a
specified value μ0, one uses the statistic.
The independent samples t-test is used when two separate sets of independent and
identically distributed samples are obtained, one from each of the two populations being
compared.

Paired t test: A test of the null hypothesis that the difference between two responses
measured on the same statistical unit has a mean value of zero.

The assumptions underlying a t-test are that

 X follows a normal distribution with mean μ and variance σ2


 s2 follows a χ2 distribution with p degrees of freedom under the null hypothesis, where
p is a positive constant
 Z and s are independent.

ANALYSIS OF VARIANCE (ANOVA)

The one-way analysis of variance (ANOVA) is used to determine whether there are any
statistically significant differences between the means of three or more independent
(unrelated) groups.

Assumptions

There are four basic assumptions used in ANOVA.

 the expected values of the errors are zero


 the variances of all errors are equal to each other
 the errors are independent
 they are normally distributed

The ANOVA technique applies when there are two or more than two independent groups.
The ANOVA procedure is used to compare the means of the comparison groups and is
conducted using the same five step approach used in the scenarios discussed in previous
sections. Because there are more than two groups, however, the computation of the test
statistic is more involved. The test statistic must take into account the sample sizes, sample
means and sample standard deviations in each of the comparison groups.

If one is examining the means observed among, say three groups, it might be tempting to
perform three separate group to group comparisons, but this approach is incorrect because
each of these comparisons fails to take into account the total data, and it increases the
likelihood of incorrectly concluding that there are statistically significant differences, since
each comparison adds to the probability of a type I error. Analysis of variance avoids these
problemss by asking a more global question, i.e., whether there are significant differences
among the groups, without addressing differences between any two groups in particular
(although there are additional tests that can do this if the analysis of variance indicates that
there are differences among the groups).

The fundamental strategy of ANOVA is to systematically examine variability within groups


being compared and also examine variability among the groups being compared.

EXAMPLE OF ANOVA

A study is designed to test whether there is a difference in mean daily calcium intake in
adults with normal bone density, adults with osteopenia (a low bone density which may lead
to osteoporosis) and adults with osteoporosis. Adults 60 years of age with normal bone
density, osteopenia and osteoporosis are selected at random from hospital records and invited
to participate in the study. Each participant's daily calcium intake is measured based on
reported food intake and supplements. The data are shown below.

Normal Bone Osteopenia Osteoporosis


Density
1200 1000 890
1000 1100 650
980 700 1100
900 800 900
750 500 400
800 700 350

Is there a statistically significant difference in mean calcium intake in patients with normal
bone density as compared to patients with osteopenia and osteoporosis?

Solving ANOVA using the five-step approach.

 Step 1. Set up hypotheses and determine level of significance

H0: μ1 = μ2 = μ3 H1: Means are not all equal α=0.05

 Step 2. Select the appropriate test statistic.

The test statistic is the F statistic for ANOVA, F=MSB/MSE.

 Step 3. Set up decision rule.

In order to determine the critical value of F we need degrees of freedom, df1=k-1 and df2=N-
k. In this example, df1=k-1=3-1=2 and df2=N-k=18-3=15. The critical value is 3.68 and the
decision rule is as follows: Reject H0 if F > 3.68.

 Step 4. Compute the test statistic.


To organize our computations we will complete the ANOVA table. In order to compute the
sums of squares we must first compute the sample means for each group and the overall
mean.

Normal Bone Osteopenia Osteoporosis


Density
n1=6 n2=6 n3=6

If we pool all N=18 observations, the overall mean is 817.7.


We can now compute:

Substituting:

Finally,

Next,

SSE requires computing the squared differences between each observation and its group
mean. We will compute SSE in parts. For the participants with normal bone density:

Normal Bone (X - (X -
Density 938.3) 938.3)2
1200 261.7 68,486.9
1000 61.7 3,806.9
980 41.7 1,738.9
900 -38.3 1,466.9
750 -188.3 35,456.9
800 -138.3 19,126.9
Total 0 130,083.4
Thus,

For participants with osteopenia:

Osteopenia (X - 715.0) (X - 715.0)2


1000 285.0 81,225.0
1100 385.0 148,225.0
700 -15 225.0
800 85.0 7,225.0
500 -215.0 46,225.0
700 -15 225.0
Total 0 283,350.0

Thus,

For participants with osteoporosis:

Osteoporosis (X - 715.0) (X - 715.0)2


890 90 8,100.0
650 -150 22,500.0
1100 300 90,000.0
900 100 10,000.0
400 -400 160,000.0
350 -450 202,500.0
Total 0 493,100.0

Thus,

We can now construct the ANOVA table.


Source of Sums of Squares Degrees of Mean Squares F
Variation (SS) freedom (df) (MS)
Between 152,429.6 2 76,214.8 1.26
Treatments
Error or Residual 906,533.4 15 60,435.6
Total 1,058,963.0 17

 Step 5. Conclusion.

We do not reject H0 because 1.26 < 3.68. We do not have statistically significant evidence at
a =0.05 to show that there is a difference in mean calcium intake in patients with normal bone
density as compared to osteopenia and osterporosis.

ANALYSIS OF COVARIANCE (ANACOVA)

Analysis of covariance (ANCOVA) is a general linear model which blends ANOVA and regression.
ANCOVA evaluates whether population means of a dependent variable (DV) are equal
across levels of a categorical independent variable (IV) often called a treatment, while
statistically controlling for the effects of other continuous variables that are not of primary
interest, known as covariates (CV) or nuisance variables.

Mathematically, ANCOVA decomposes the variance in the DV into variance explained by


the CV(s), variance explained by the categorical IV, and residual variance. Intuitively,
ANCOVA can be thought of as 'adjusting' the DV by the group means of the CV(s).

CONCLUSION

All the statistical techniques/tests can be classified under five heads.

Classification Statistical Techniques/Tests


Measures of Central Mean, Median, Mode, Harmonic Mean and Geometric Mean
Tendency
Measures of Range, Quartile Deviation, Standard Deviation, Mean Deviation,
Dispersion Coefficient of variation
Measures of Correlation, regression, chi square test, factor analysis, cluster
associations/relations analysis, discriminant analysis and canonical analysis
Analysis of variance One-way ANOVA, two-way ANOVA, MANOVA, ANACOVA and
MANACOVA
Time series analysis Seasonal, cyclical, trend and erratic variations

Compiled by

Dr. A. G. Vijayanarayanan
M.Com., MBA, PGDSBSA, Ph.D.,

Vous aimerez peut-être aussi