Académique Documents
Professionnel Documents
Culture Documents
discussions, stats, and author profiles for this publication at: http://www.researchgate.net/publication/5378506
CITATIONS
READS
34
114
2 AUTHORS, INCLUDING:
Ulrich Guller
Kantonsspital St. Gallen
139 PUBLICATIONS 2,841 CITATIONS
SEE PROFILE
EDUCATION
Background
Let us assume that we are evaluating patients after laparoscopic cholecystectomy. The primary end point (also
known as the outcome or dependent variable) of the
investigation is the length of hospital stay. For the sake of
simplicity, let us say that our sample includes five patients. The lengths of hospital stay of the patients who all
had a postoperative course without complications were 1
day, 1 day, 2 days, 3 days, and 3 days. In this example, the
mean and median are identical (2 days). Now hypothesize that the fifth patient suffered from a postoperative
infection that led to a generalized sepsis and respiratory
failure requiring prolonged intubation, intravenous antibiotics, and transfer to the intensive care unit and that
this patient was finally dismissed after 56 days of hospitalization. The mean length of hospital stay is now 12.6
days, but the median remains unchanged.
Outlierspatients who behave very differently from
the majority of patientsare frequently present in medical literature and might render interpretation of the
study findings difficult.9 Use of the median helps prevent potential distortion of study findings caused by
extreme values and should be preferably used if outliers
are present.
441
ISSN 1072-7515/04/$30.00
doi:10.1016/j.jamcollsurg.2003.09.017
442
CI
confidence interval
COPD chronic obstructive pulmonary disease
IBD inflammatory bowel disease
SD
standard deviation
SEM standard error of the mean
VMA vanillymandelic acid
But both mean and median sometimes fail to appropriately reflect the nature of the data. Consider a different sample of seven patients undergoing laparoscopic
cholecystectomy. Let us assume that the lengths of hospital stays were 1 day, 1 day, 1 day, 2 days, 9 days, 11
days, and 14 days. The median in this example, 2 days,
poorly summarizes the middle of the data. The mean,
5.6 days, doesnt give a clear picture of the nature of the
data either. It is critical that a measure of the variability
of the data and a mean or median be used to summarize
a data distribution.
How to interpret standard deviations and standard
errors of the mean
Although the standard deviation (SD) and standard error of the mean (SEM) are distinctly different statistics,
they are often used interchangeably in medical literature.
Confusion about their correct interpretation is
considerable.2,9
The SD is a measure of the variability (scatter) of a
data distribution. It is a measure of the degree to which
the individual values deviate from the population mean.
Larger deviations from the mean indicate more extensive
scatter and result in larger standard deviations. The sample SD is an estimate of the population SD and is computed using deviations from the sample mean. There is a
common misconception in the medical community
about interpretation of the standard deviation: it is often
believed that the standard deviation decreases with increasing sample size. Regardless of the sample size, the
standard deviation will be large if the data are highly
scattered.
The standard error of the mean (SEM) reflects the
variability in the distribution of sample means from the
population mean. If several different samples were available (which is not generally the case), their means would
vary and would have a standard deviation, which is the
SEM. Contrary to the SD, the SEM is an inferential
statistic strongly dependent on the sample size.9 The
J Am Coll Surg
larger the sample size, the smaller the SEM, and the
more precisely the sample mean estimates the overall
population mean.
Because both sample SD and SEM are statistics frequently used in medical literature, it is important to
know how to convert one to the other. The SEM can be
obtained by dividing the SD by the square root of the
number of patients in the sample (SEM SD/n).
Multiplication of the SEM by the square root of the
number of patients will result in the SD (SD SEM
n). Again, the SD is the appropriate statistic to describe the scatter of the data; the SEM estimates the
variability of an estimate of the sample mean. Some investigators display error bars using the sample-based estimate of the SEM instead of the SD for graphic presentation of the data scatter, leading readers to believe that
there is little data dispersion.2,9 The graphic presentation
of SEM instead of SD to indicate data scatter is
misleading.
A standard error can be computed not only for a mean
but for any kind of sample statistic, eg, proportions,
differences between means and proportions, regression
parameters (as described below), and so forth. We will
discuss the standard errors in context with confidence
intervals.
How to interpret risk ratios, absolute risk
reduction, odds ratios, and number needed to treat
Although the display and analysis of continuous outcomes (such as tumor size, length of hospital stay, total
bilirubin concentration in the blood, and so forth) are
easy and intuitive, the interpretation of percentages for
dichotomous (binary, yes or no) end points is often
more challenging.10 Examples of dichotomous outcomes are death, tumor relapse, liver failure, gastrointestinal bleeding, and so forth. Results of dichotomous end
points are frequently presented in medical literature using risk ratios, odds ratios, absolute risk reduction, and
the number needed to treat. Nonetheless, confusion exists in the medical community about the interpretation
of these different analytical tools. To facilitate the understanding, let us consider the following hypothetical
example.
There is extensive evidence in the medical literature
that consumption of a low-fiber diet is a causative factor
in the development of colon diverticulosis.11-13 In the
US, approximately one-third of people over 45 years of
age and two-thirds of people over 85 years of age suffer
Yes
No
Present
A
C
Absent
90
40
10
60
B
D
100
100
The number needed to treat has been recently introduced into medical literature as a measure of treatment
or prevention efficacy.15 The number needed to treat (or
in our example, the number needed to prevent) represents the number of patients that must be treated, or
from whom a certain risk factor must be removed, to
prevent the occurrence of one case. The number needed
to treat is the inverse of the absolute risk reduction (in
443
444
Yes
No
Present
A
C
70
30
Absent
15
85
B
D
85
115
J Am Coll Surg
samples of the same size, from the same overall population, and calculated the CI in the same manner, the true
parameter in the overall population would be included
95% of the time. There remains a 5% chance that the
true population parameter is outside the 95% CI. If an
investigator wishes to be more sure that the confidence
interval based on the patient sample includes the true
population value, a 99% CI can be chosen, but the 99%
CI is wider than a 95% CI because there is a smaller
degree of uncertainty. The higher the level of confidence,
the wider the confidence limits. A 95% CI can be computed for means, proportions, differences of means and
proportions, risk ratios, odds ratios, sensitivity, specificity, and so forth. Again, computing confidence intervals
only makes sense if the sample is representative of a
larger population for which inferences can be drawn.
The width of the confidence interval indicates the precision of an estimate and is dependent on the variation in
the data and the number of subjects in the sample. The
width of the 95% CI and the standard error (SE) are
closely related: for samples of sufficiently large size
(n 60), the 95% CI is usually calculated as the
mean 2 SEM. (The factor, with which the SEM is
multiplied, varies with the samples size. For n 60 it is
exactly 2, for n 10 it is 2.3, for an infinitely large
sample size, the factor is 1.96. But the choice of a factor
2 is a good approximation in the vast majority of applications.) The greater the dispersion of data and the
smaller the sample size, the larger the SE and the wider
the confidence interval. Conversely, the less scattered the
data and the larger the patient sample, the narrower the
confidence interval. A wide confidence interval indicates
that the sample data are insufficient for precisely estimating the effect in the overall population and must be
interpreted cautiously, regardless of whether or not the
results are statistically significant.17,18
Example
Twenty-four hour urine measurement of vanillylmandelic acid (VMA) represents a sensitive and specific test
in the diagnosis of pheochromocytoma patients.19,20 Let
us consider a hypothetical sample of 10 patients with
pheochromocytoma. The urine measurements in these
patients yielded VMA values of 50, 60, 70, 80, 90, 100,
110, 120, 130, and 140 mg/24 h (normal: 7 mg/24 h).
The 95% CI for the mean VMA value ranges from 73.3
to 116.7 mg/24 h (sometimes displayed as: 95% CI
(73.3, 116.7 mg/24 h). The first value is called the lower
The findings of a study comparing two groups of patients can be wrong in two ways21:
1. The results might lead to the erroneous conclusion that
there is a difference between the study groups when, in
reality, there is none.
2. The results might lead to the erroneous conclusion that
there is no difference between the study groups when, in
reality, a difference exists.
445
With larger sample sizes, larger true differences between the populations from which the patient samples
have been drawn, or higher acceptance of false-positive
results, the power of the study increases.
Before initiating the study, power and sample size
must be determined. For sample size computations, investigators start by defining a clinically meaningful difference between treatments A and B, which is believed to
be true for the overall patient population. This difference is usually based on preliminary data of small phase
II studies or retrospective reviews, but is sometimes specified according to clinical intuition. If the investigator is
satisfied with an 80% probability of obtaining a statistically significant difference between the study groups, if
such a difference truly exists, a smaller sample size is
required than if 90% power were chosen. In other
words, a larger sample size corresponds to a higher level
of power. Ideally, both alpha and beta would be set at 0 to
avoid false-positive and false-negative findings. This
would require a prohibitively large sample size, rendering any trial unfeasible. For a patient sample of given
size, there is a tradeoff between alpha and beta: the more
stringent alpha (the lower the false-positive rate), the
higher beta (increased rate of false-negative results, lower
power), and vice versa.21,23 In general, one should choose
a small alpha level if avoiding false-positive results is
particularly important (eg, testing the efficacy of a new
chemotherapy regimen with serious adverse effects).
Similarly, a small beta level should be chosen if obtaining
a false-negative result would be deleterious; for example,
446
J Am Coll Surg
Table 3. Sample Size Computations* for a Study Showing the 5-Year Overall Survival Difference Between Colorectal Cancer
Patients With and Without Disseminated Tumor Cells
Expected 5-y overall survival
of patients with
disseminated tumor cells (%)
45
45
45
50
50
50
55
55
55
55
55
55
Alpha
Beta
75
75
75
75
75
75
75
75
75
70
70
70
0.05
0.05
0.05
0.05
0.05
0.05
0.05
0.05
0.05
0.05
0.05
0.05
0.20
0.15
0.10
0.20
0.15
0.10
0.20
0.15
0.10
0.20
0.15
0.10
79
89
105
108
124
145
161
185
215
287
328
384
*All sample size computations are based on a 25% positivity rate of the samples, a type I error probability of 0.05 (two-sided), a projected accrual period of 3 y,
and a followup interval of 5 y. The program used for sample size computations was based on references 61 through 65.
3.
Example
4.
5.
6.
447
448
J Am Coll Surg
Example
Researchers are more likely to report p values than confidence intervals.26,30 But confidence intervals provide
much more information to the astute reader than do p
values alone3,17,26 and are now requested by many journals in the reporting of study findings.1,39 Although p
values and confidence intervals might seem different at
first glance, closer scrutiny reveals that they are complementary. Both are computed using the same underlying
assumptions. If the 95% confidence interval includes
the value of the null hypothesis (eg, 0 for the difference
between means, or 1 for a risk ratio and or odds ratio),
the p value will be greater than 0.05. On the other hand,
449
if the 95% CI does not include the value of no difference, the p value will be less than 0.05. Confidence
intervals are also very helpful in interpreting nonsignificant p values.3,17 If the range of the 95% CI of differences or risk ratios includes values that are clinically
trivial, one can assume the results to be irrelevant with
higher confidence. If the 95% CI includes values that
you find clinically important, the study should be considered inconclusive because a small sample size might
be a reason for not reaching statistical significance.
Example
450
J Am Coll Surg
viduals in Western countries. The etiology and pathogenesis of IBD await additional clarification. It is generally accepted that IBD occurs predominantly in
genetically susceptible people51,52 yet many genetic loci
involved in the pathogenesis of this disease need to be
determined.53 Satsangi and colleagues54 performed a
genome-wide search for susceptibility genes in patients
with IBD. The authors investigated 260 different genetic loci for potential linkage to IBD. Because 260 null
hypotheses were tested, there was a definite risk of obtaining false-positive results if the standard alpha level of
0.05 was used. So the investigators correctly chose a level
of statistical significance of p 0.001, which is 50 times
smaller than the commonly used cutoff (p 0.05).
Also, the investigators viewed their study as hypothesis
generating rather than hypothesis testing.
How to interpret survival curves
451
Figure 2. Hypothetical survival curve comparing two groups of colorectal cancer patients. The solid line survival curve plots time to
death for node negative, the dotted curve for node-positive patients.
In the group without lymph node metastases, only one patient dies
at 15 mo of followup. Nine node-negative patients die, two (step is
twice as big) at 12 mo of followup. Median survival for the nodepositive group is approximately 10 mo; the median survival for the
negative group cannot be computed because fewer than half of the
subjects are deceased at the end of the study. Note that 16 patients
were censored, and that the survival curve remains horizontal during
censoring.
452
J Am Coll Surg
In medical literatureespecially in nonrandomized observational studiesconfounding can be a major problem. For instance, in an observational study comparing
two treatments, patients undergoing treatment A might
substantially differ from patients having treatment B
with respect to factors such as age, gender, race, socioeconomic status, comorbid diseases, and tumor staging
and grading. All of these are probably also related to the
outcomes of interest, so are potential confounding variables. The investigator seeks to assess the relationship
between the primary predictor variable (type of therapy,
A versus B) and the outcomes under investigation after
the potential distortion through covariates has been
eliminated. The use of stratification when multiple potential confounders are present is cumbersome. The preferred method of adjusting for many confounding variables simultaneously is multivariable analyses.
Depending on the type of outcome variable, one of
three different multivariable analyses is generally appropriate: For continuous outcomes (eg, postoperative
length of hospital stay), multiple linear regression analysis is appropriate; for categorical or dichotomous outcomes (eg, presence of metastases), multiple logistic regression analysis is useful; and for time to event
453
Table 5. Hypothetical Multiple Linear Regression Model: Comparison of Length of Hospital Stay Between Open and
Laparoscopic Appendectomy
Variable
Laparoscopic surgery
Age (y)
COPD
Beta coefficient
Standard error
p Value
0.67
0.02
0.37
0.04
0.001
0.04
0.0001
0.0001
0.0001
[0.74, 0.61]
[0.019, 0.024]
[0.29, 0.45]
In all three types of multivariable analysis, the variables beta coefficient (not to be confused with the previously mentioned rate of false-negative results) indicates how the dependent variable responds to changes of
the independent variable, after adjusting for all other
covariates in the model. As the type of model changes for
different types of end points, the interpretation of the
beta coefficient changes too (Table 4).
In multiple linear regression analysis, the outcome is
continuous. A positive beta coefficient signifies that the
independent variable and the mean value of the dependent variable vary in the same direction (either both
increasing or both decreasing). Conversely, a negative
coefficient indicates that while the independent variable
increases the mean outcome decreases, and vice versa.
In multiple logistic regression analysis, the appropriate statistical tool to assess the risk-adjusted impact of
covariates on categoric end points, the logit is modeled.
The logit is the natural logarithm of the odds of experiencing the outcomes. The beta coefficient in logistic
regression models expresses how the logit changes with
changes in the independent variable. Although the logit
is difficult to interpret on its own, it can be easily transformed into the odds ratio by taking the antilogarithm
(exponentiating) of the beta coefficient.
As mentioned previously, proportional hazards regression is frequently used if a time to event end point is
being evaluated. Time to event outcomes incorporate
more information than dichotomous end points if the
followup period is sufficiently long (eg, we not only
know whether or not death occurred but also how long
after cancer resection).32,36,60 In proportional hazards regression analysis, the beta coefficient represents the
change in the natural logarithm of the relative hazard for
a one-unit change in the independent variable. The relative hazard represents the ratio of instantaneous risk of
experiencing the event for patients having a certain risk
factor to the instantaneous risk of experiencing the event
for patients where this risk factor is absent. Similar to the
coefficients in logistic regression analyses, the relative
454
J Am Coll Surg
Table 6. Hypothetical Multiple Logistic Regression Model: Comparison of Postoperative Infections Between Open and
Laparoscopic Appendectomy
Variable
Laparoscopic surgery
Age (y)
COPD
Beta coefficient
Standard error
p Value
0.535
0.007
0.10
0.05
0.001
0.06
0.0001
0.0001
0.1
pendectomy were coded as 1 and laparoscopic appendectomy as 0, the beta coefficient would have been
0.67, meaning that open appendectomy patients have
a length of hospital stay 0.67 days longer than laparoscopic appendectomy patients.
Age. The beta coefficient for age is 0.02. So, for each
year of age, the average length of hospital stay increases
by 0.02 days. In other words, 60-year-old patients are
expected to have a 0.8-day (40 0.02 days/year 0.8
days) longer length of hospital stay than 20-year-old
patients.
Chronic obstructive pulmonary disease (COPD).
COPD was coded as present (1) and absent (0). The
presence of COPD is associated with hospital stays 0.37
days longer than those for patients who do not have
COPD. Again, the interpretation of coding is critical. If
absence of COPD were coded as 1, the beta coefficient
would have been 0.37.
For all three covariates in this model, the beta coefficients are more than twice as large as the standard errors,
so the p values are statistically significant. Also, the 95%
confidence intervals do not include the value of the null
hypothesis (for continuous outcomes 0), indicating
that the p values are below the level of statistical
significance.
Example (table 6)
smaller than 1. Conversely, if the beta coefficient is positive, the corresponding odds ratio will exceed 1.
Laparoscopic surgery. The beta coefficient is negative, resulting in an odds ratio smaller than 1. The odds
ratio (laparoscopic versus open surgery) of having a postoperative infection is 0.59 (antilogarithm of 0.535).
Again, interpretation of the coding is critical. In this
example, laparoscopic appendectomy was coded as 1
(and open as 0). So, the odds of having a postoperative
infection after laparoscopic surgery is only 0.59 times
the odds of experiencing an infection after open surgery.
If open appendectomy were coded as 1, the beta coefficient would be 0.535, and the corresponding odds
ratio 1.69 (1/.59), meaning that the odds of having a
postoperative infection after open surgery are 1.69 times
those of laparoscopic surgery. The change in coding
would not result in a change of the p value.
COPD. The beta coefficient is positive, so the corresponding odds ratio is greater than 1. Patients with
COPD are 1.11 times more likely to experience postoperative infections than patients without COPD. Again,
if the coding were reversed (absence of COPD 1,
presence of COPD 0), the beta coefficient would
become negative (without a change in the absolute
value: 0.10), and the odds ratio would be 0.9 (1/
1.11) and would be interpreted as the odds for nonCOPD patients relative to patients with COPD. The
95% confidence interval crosses the value of the null
hypothesis (for ratios 1), resulting in a nonsignificant
p value. Also, the beta coefficient is smaller than twice
the standard error, which is indicative that the results are
not significant.
How to assess the performance of
statistical models
455
As discussed previously, the most frequently encountered end points in medical literature are continuous,
categorical, or time to event. Although continuous outcomes cover a range (eg, tumor shrinkage measured in
millimeters after neoadjuvant chemotherapy), categorical outcomes have only certain possible values (eg, dichotomous outcomes, such as response to chemotherapy, presence of metastases, and so forth). The time to
event outcome is frequently used in surgical oncology,
where studies evaluate the time after tumor resection to
relapse or death.
Paired versus unpaired data
456
J Am Coll Surg
Parametric test
Corresponding
nonparametric
test
Continuous outcome
Two sample
(unpaired) t test
Mann-Whitney-U
test
Wilcoxon matched
pair test
Kruskal-Wallis test
Friedman test
(analogue of
repeated measures
ANOVA)
One sample
(paired) t test
Regression on
ranks, other less
frequently used
tests
Dichotomous/categoric outcome
Chi-square test or Fishers exact test
McNemars test
Example
For small sample sizes ( 20 patients), the chi-square test does not provide accurate results and Fishers exact test must be used. For large samples, both Fishers
exact test and chi-square test yield similar results.
REFERENCES
1. Altman DG. Statistics in medical journals: developments in the
1980s. Stat Med 1991;10:18971913.
2. Hayden GF. Biostatistical trends in pediatrics: implications for
the future. Pediatrics 1983;72:8487.
3. Rothman KJ. A show of confidence. N Engl J Med 1978;299:
13621363.
4. How to read clinical journals: I. why to read them and how to
start reading them critically. Can Med Assoc J 1981;124:555
558.
5. Evidence-Based Medicine Working Group. Evidence-based
medicine. A new approach to teaching the practice of medicine.
JAMA 1992;268:24202425.
6. Guyatt GH, Rennie D. Users guides to the medical literature.
JAMA 1993;270:20962097.
7. Berwick DM, Fineberg HV, Weinstein MC. When doctors
meet numbers. Am J Med 1981;71:991998.
8. Friedman SB, Phillips S. Whats the difference? Pediatric residents and their inaccurate concepts regarding statistics. Pediatrics 1981;68:644646.
457
9. Brown GW. Standard deviation, standard error. Which standard should we use? Am J Dis Child 1982;136:937941.
10. Jaeschke R, Guyatt G, Shannon H, et al. Basic statistics for
clinicians: 3. Assessing the effects of treatment: measures of association. CMAJ 1995;152:351357.
11. Munakata A, Nakaji S, Takami H, et al. Epidemiological evaluation of colonic diverticulosis and dietary fiber in Japan. Tohoku
J Exp Med 1993;171:145151.
12. Brodribb AJ, Humphreys DM. Diverticular disease: three studies. Part IRelation to other disorders and fibre intake. Br
Med J 1976;1:424425.
13. Aldoori WH, Giovannucci EL, Rockett HR, et al. A prospective
study of dietary fiber types and symptomatic diverticular disease
in men. J Nutr 1998;128:714719.
14. Kohler L, Sauerland S, Neugebauer E. Diagnosis and treatment
of diverticular disease: results of a consensus development conference. The Scientific Committee of the European Association
for Endoscopic Surgery. Surg Endosc 1999;13:430436.
15. Laupacis A, Sackett DL, Roberts RS. An assessment of clinically
useful measures of the consequences of treatment. N Engl J Med
1988;318:17281733.
16. Bland JM, Altman DG. Statistics notes. The odds ratio. BMJ
2000;320:1468.
17. Altman DG, Gore SM, Gardner MJ, Pocock SJ. Statistical
guidelines for contributors to medical journals. Br Med J (Clin
Res Ed) 1983;286:14891493.
18. Katz MH. Interpreting the analysis. Multivariable analysis. A
practical guide for clinicians. Cambridge: Cambridge University
Press; 1999:133.
19. Hanson MW, Feldman JM, Beam CA, et al. Iodine 131-labeled
metaiodobenzylguanidine scintigraphy and biochemical analyses in suspected pheochromocytoma. Arch Intern Med 1991;
151:13971402.
20. Lenders JW, Pacak K, Walther MM, et al. Biochemical diagnosis
of pheochromocytoma: which test is best? JAMA 2002;287:
14271434.
21. Berwick DM. Experimental power: the other side of the coin.
Pediatrics 1980;65:10431045.
22. Freiman JA, Chalmers TC, Smith H Jr, Kuebler RR. The importance of beta, the type II error and sample size in the design
and interpretation of the randomized control trial. Survey of 71
negative trials. N Engl J Med 1978;299:690694.
23. Perneger TV. Whats wrong with Bonferroni adjustments. BMJ
1998;316:12361238.
24. Simon R, Altman DG. Statistical aspects of prognostic factor
studies in oncology. Br J Cancer 1994;69:979985.
25. Altman DG, Schulz KF, Moher D, et al. The revised CONSORT statement for reporting randomized trials: explanation
and elaboration. Ann Intern Med 2001;134:663694.
26. Pocock SJ, Hughes MD, Lee RJ. Statistical problems in the
reporting of clinical trials. A survey of three medical journals.
N Engl J Med 1987;317:426432.
27. Vogel I, Francksen H, Soeth E, et al. The carcinoembryonic
antigen and its prognostic impact on immunocytologically detected intraperitoneal colorectal cancer cells. Am J Surg 2001;
181:188193.
28. Schott A, Vogel I, Krueger U, et al. Isolated tumor cells are
frequently detectable in the peritoneal cavity of gastric and colorectal cancer patients and serve as a new prognostic marker. Ann
Surg 1998;227:372379.
29. Lindemann F, Schlimok G, Dirschedl P, et al. Prognostic sig-
458
30.
31.
32.
33.
34.
35.
36.
37.
38.
39.
40.
41.
42.
43.
44.
45.
46.
47.
J Am Coll Surg
48. Lee KL, McNeer JF, Starmer CF, et al. Clinical judgment and
statistics. Lessons from a simulated randomized trial in coronary
artery disease. Circulation 1980;61:508515.
49. Rothman KJ. No adjustments are needed for multiple comparisons. Epidemiology 1990;1:4346.
50. Stewart LA, Parmar MK. Bias in the analysis and reporting of
randomized controlled trials. Int J Technol Assess Health Care
1996;12:264275.
51. Duerr RH. The genetics of inflammatory bowel disease. Gastroenterol Clin North Am 2002;31:6376.
52. Karban A, Eliakim R, Brant SR. Genetics of inflammatory
bowel disease. Isr Med Assoc J 2002;4:798802.
53. Yang H, Rotter JI. The genetic background of inflammatory
bowel disease. Hepatogastroenterology 2000;47:514.
54. Satsangi J, Parkes M, Louis E, et al. Two stage genome-wide
search in inflammatory bowel disease provides evidence for susceptibility loci on chromosomes 3, 7 and 12. Nat Genet 1996;
14:199202.
55. Pocock SJ, Clayton TC, Altman DG. Survival plots of time-toevent outcomes in clinical trials: good practice and pitfalls. Lancet 2002;359:16861689.
56. Altman DG, Bland JM. Time to event (survival) data. BMJ
1998;317:468469.
57. Mullner M, Matthews H, Altman DG. Reporting on statistical
methods to adjust for confounding: a cross-sectional survey.
Ann Intern Med 2002;136:122126.
58. Ramstedt M. Per capita alcohol consumption and liver cirrhosis
mortality in 14 European countries. Addiction 2001;96(Suppl
1):S1933.
59. Singh GK, Hoyert DL. Social epidemiology of chronic liver
disease and cirrhosis mortality in the United States, 19351997:
trends and differentials by ethnicity, socioeconomic status, and
alcohol consumption. Hum Biol 2000;72:801820.
60. Green MS, Symons MJ. A comparison of the logistic risk function and the proportional hazards model in prospective epidemiologic studies. J Chronic Dis 1983;36:715723.
61. Taulbee JD, Symons MJ. Sample size and duration for cohort
studies of survival time with covariables. Biometrics 1983;39:
351360.
62. Schoenfeld DA, Richter JR. Nomograms for calculating the
number of patients needed for a clinical trial with survival as an
endpoint. Biometrics 1982;38:163170.
63. Schoenfeld DA. Sample-size formula for the proportionalhazards regression model. Biometrics 1983;39:499503.
64. Makuch RW, Simon RM. Sample size requirements for comparing time-to-failure among k treatment groups. J Chronic Dis
1982;35:861867.
65. Lachin JM. Introduction to sample size determination and
power analysis for clinical trials. Control Clin Trials 1981;2:93
113.