Académique Documents
Professionnel Documents
Culture Documents
Paper
presented at the 22nd annual meeting of the Society for Industrial and Organizational Psychology, New York,
NY.
This study explores reactions and responses to open-ended questions on Web-based surveys. An
experimental design is used to examine the effects of requiring participants to answer open-ended
questions. Findings show that requiring these responses increases perceptions of item importance and the
number of usable comments without increasing negative affective reactions.
It is difficult to think of any aspect of the Trends in Survey Technology and Open-Ended
working world that is untouched by technology. Questions
Organizational surveying is one area where Survey technologies continue to advance,
technology has caused a major change from providing new and improved options for both
traditional practices (Couper, 2000). Not only researchers and practitioners. One development
has technology affected the survey experience that has had a significant impact on survey
for respondents, but it has also resulted in new practice is the transition from paper-based to
approaches to constructing surveys and web-based surveys. An environmentally friendly
gathering data (DeMay, Chandonnet, & alternative to paper questionnaires, web-based
Fenlason, 2005). From a data collection surveys can be efficiently distributed to many
standpoint, survey-related technologies have individuals simultaneously. Web-based
reduced many of the barriers to gathering technology has not only facilitated the
responses in general and open-ended comments dissemination of surveys, it has also reduced
in particular. This has prompted a growing many of the traditional costs associated with
reliance on open-ended survey questions and an using open-ended questions for both respondents
increased need to expand our understanding of and survey designers (Borg, 2005; Kulesa &
how people view and react to these types of Bishop, 2006; Thompson, Surface, Martin, &
questions. The present study addresses this Sanders 2003). Whereas paper surveys collect
research need. The purpose of this research is to written responses that need to be entered or
determine whether people view open-ended typed, web-based surveys automatically capture
questions as less important than closed-ended comment data in a digital format, thereby
questions. In addition, we examine whether eliminating the time-consuming step of typing or
requiring and prompting survey comments otherwise entering open-ended responses.
affects the perceived importance of open-ended Meanwhile, advances in computer-aided text
questions as well as reactions to the survey in analysis programs have eased the process of
general. Finally, this study determines whether organizing and summarizing survey comments
requiring and prompting open-ended responses obtained from both paper- and web-based
increases the number of usable comments surveys.
gathered from a survey. Accompanying these technological advances
is a renewed interest in including open-ended
questions on organizational questionnaires.
Open-ended comments 2
Open-ended survey questions are valued as a unimportant. This perception could stem from
data collection tool because of the in-depth the larger proportion of closed-ended questions
information they can provide (Kraut, 1996; compared to open-ended questions that are
Wagner & Spencer, 1996). Edwards. Thomas, typically found on surveys (Church &
Rosenfeld, and Booth-Kewley (1997) argue that Waclawski, 1998). The greater number of
“open-ended questions can add flesh to the closed-ended questions may be interpreted as an
sometimes sterile numbers and deductions indication of the importance of these questions.
drawn from responses to closed-ended items” (p. Another possibility has to do with the placement
26). While they have been considered useful for of open-ended questions, which commonly
quite some time, open-ended survey questions appear at the end of a survey (Edwards et al.,
seem to be growing in popularity due to the ease 1997), perhaps causing them to seem like a
with which they can now be collected and survey design “afterthought.” The fact that
analyzed. responses to closed- but not open-ended
The increasing inclusion of open-ended questions on web-based surveys are often
questions on surveys has sparked a movement in explicitly required may also contribute to the
the research literature to explore many perception that comment data are
previously unanswered questions about the data inconsequential. Finally, the wording of the
collection methodology commonly known as the open-ended question may affect responses.
“survey comment block” (Borg, 2005; Kulesa & Mossholder, Settoon, Harris, and Armenakis
Bishop, 2006; Poncheri, Lindberg, Thompson, (1995) argue that most open-ended questions are
& Surface, in press). The present study extends worded in a way that implies that they are
this line of research. optional, and therefore unimportant.
Although there are reasons to believe that
Why Don’t Participants Respond to Open-Ended open-ended comments are viewed as less critical
Questions? than closed-ended ratings, this possibility has
Nonresponse error is a ubiquitous problem not been tested empirically. The present study
in survey research and occurs when respondents addresses this research need.
differ from the population of interest in ways Hypothesis 1: Open-ended questions
that can have a meaningful impact on survey will be viewed as less important
results (Dillman, 2000; Sills & Song, 2002). than closed-ended questions in a
Two major types of nonresponse have been standard Web-based survey.
identified in survey research, unit and item
nonresponse (Bosnjak & Tuten, 2001; Bosnjak, Requiring Open-Ended Comments
Tuten, & Wittman, 2005). Unit nonresponse If support is found for the hypothesis that
occurs when a respondent does not answer any open-ended questions are viewed as relatively
questions on a survey, resulting in the loss of a unimportant, then it becomes essential to
survey unit (i.e., respondent). Item nonresponse examine how survey researchers and
occurs when a respondent answers some of the practitioners interested in obtaining comment
questions on a survey, but skips others. data can change this perception and persuade
Although the majority of this research has more respondents to provide comments. It is
focused on closed-ended questions, nonresponse possible that open-ended questions are perceived
to open-ended questions can be viewed as a as relatively unimportant because respondents
form of item nonresponse as well. Several are usually allowed to submit surveys even if
studies have shown that most respondents do not they have left their “comment blocks” blank.
provide open-ended comments, resulting in a Therefore, one way to change perceptions and
high rate of item nonresponse to these types of encourage a higher level of response may be to
questions (Borg, 2005; Poncheri et al., in press; require survey comments.
Siem, 2005). Many web-based survey applications have a
One potential reason that survey respondents feature which allows survey programmers to
may choose not to answer open-ended questions require responses to open- as well as closed-
is that these questions might be perceived as ended questions (Dillman, 2000). That is, web-
Open-ended comments 3
based surveys can be designed so that Notably, DeRouvray and Couper’s (2002)
respondents are not allowed to advance to the prompting strategy could be applied to open-
next item or screen if a previous item has been ended questions. This would entail encouraging,
left blank. Dillman (2000) notes that this feature but not mandating, responses to open-ended
“is often promoted as a distinct advantage of questions through the use of a prompt. Such an
Web surveys” (p. 394). Although many survey approach may help communicate the importance
designers require responses to closed-ended of open-ended items without forcing otherwise
questions, Dillman (2000) cautions survey unwilling people to comment.
practitioners about the implications of doing so. In sum, ongoing debates in the literature
He argues that there are situations in which it is suggest that requiring responses to open-ended
legitimate for people to have no response to an questions will likely communicate their
item and that forcing a response can result in importance but could also adversely affect
feelings of frustration and possibly attrition from reactions to the survey. For this reason,
the survey. employing a prompting strategy rather than an
Based on the potentially negative outcomes absolute response requirement may be a better
of response requirements, Dillman (2000) way to increase perceptions that open-ended
recommends requiring responses, but also comments are important and valuable. The
providing a “prefer not to answer” or “don’t assumption that response requirements and
know” option. Whereas this recommendation is prompting strategies influence respondents’
technologically feasible, there are questions affective reactions to surveys as well as their
about the effects this practice may have on the perceptions of the importance of open-ended
data collected. Research has shown that when comments awaits empirical investigation. To
“don’t know” options are offered, a larger address this need, the present study tests
proportion of respondents choose this option as Dillman’s (2000) assertion regarding the
opposed to choosing from other response frustration-inducing effects of response
options provided (e.g., Schuman & Scott, 1989). requirements by examining the affective
This is problematic because it is unclear whether reactions of respondents who are and are not
respondents truly have no opinion or if they are required to provide survey comments.
satisficing by choosing this response Hypothesis 2: Respondents who are
(DeRouvray & Couper, 2002). Although the required to complete open-ended
preceding debate has occurred in the context of questions will have more negative
closed-ended questions, it also pertains to open- reactions to the survey than those
ended items. Thus, there is a concern that simply who are not required to comment.
allowing respondents to select an option In addition, this study examines whether
indicating they have “no comment” will do little prompting as well as requiring comments can
to increase the number of open-ended responses communicate their importance.
received. Hypothesis 3: Respondents who are
DeRouvray and Couper (2002) recently required and respondents who are
explored four different strategies for reducing prompted to complete open-ended
the number of “no opinion” responses to questions will view open-ended
fourteen closed-ended questions on a web-based questions as more important than
survey. The strategy which resulted in the least respondents who are not required
amount of missing data was one in which an and not prompted to complete open-
explicit “decline to answer” response was not ended questions.
offered, but the respondent was prompted each
time a closed-ended question was not answered. One final consideration is related to the
The prompt allowed respondents to choose quality of responses that are provided when
whether they would respond to the item or skip individuals are either required or prompted to
to the next item. This strategy was superior to comment. It is unclear what type of response
three other strategies in which an explicit would result from requiring a comment,
“decline to answer” response was offered. especially if the respondent does not have a
Open-ended comments 4
legitimate answer to an item (Dillman, 2000). To an online survey (containing both closed-ended
determine which survey conditions will result in and open-ended questions) as well as an online
the greatest number of usable comments, the follow-up evaluation questionnaire. In terms of
following research question is proposed: response to the online survey, there were 103
Research Question: Does the survey participants in the control group, 99 participants
condition (i.e., required, prompted, in the required comment group, and 98
or non-prompted) affect the participants in the prompted comment group.
tendency to provide a usable open- Those in the control group were not required to
ended comment? complete the open-ended questions in the
survey. This condition was created to represent a
Method standard organizational survey. If the individuals
assigned to this condition did not provide
Participants comments, they did not receive an error message
The participants in this study were 300 and were able to proceed and submit the survey.
students at a large southeastern university who Individuals assigned to the required comment
volunteered to fulfill a course research group were required to answer the open-ended
requirement. Most participants (85.7%) were 20 questions. If the participants assigned to this
or younger. Half of the sample (50.2%) was condition did not answer the questions, they
male and the majority (79.8%) was Caucasian. received the following error message: “The
Overall, the majority of participants provided a following required questions were left
favorable technological evaluation of their unanswered: [Item]” and were not allowed to
experiences with the web-based survey used in continue to the next set of questions or submit
this study. Table 1 shows how participants rated the survey unless a response was provided.
the ease of using the technology needed to Finally, participants assigned to the prompted
complete the survey administered in the present comment group received a follow-up message if
study. Descriptive baseline information on they did not provide an answer to the open-
participants’ past experiences with surveys is ended questions. The prompt used in this study:
also provided in Table 1. “We would very much like to have your answer
to this question. If you would like to provide an
Design and Procedure answer to this question, please do so in the space
This study included one independent provided. If you would prefer not to answer this
variable, which was our non-response follow-up question, please type ‘Skip’ in the comment box.
manipulation. It had three levels: control, Then select the 'Next’ button to continue the
required comment, and prompted comment. survey. Thank you!” was adapted from the one
Four dependent variables were examined: used in an earlier study by DeRouvray and
perceived importance of closed-ended questions, Couper (2002).
perceived importance of open-ended questions,
affective reactions toward the survey, and the Survey Instrument
presence/absence of a usable survey comment. Closed-ended items. The survey
Individuals were randomly assigned to one administered in this study was the Student
of the three conditions based on the order in Assessment of the College Environment (SACE;
which they signed up to participate in the study. Baker, 1997). The SACE was designed to
A list of random numbers in which each number evaluate the organizational climate of
was associated with one of the three conditions college/university life from the student
was generated prior to data collection and perspective. This measure was chosen because it
participants were assigned to conditions based gathers opinions about organizational issues of
on this list. Due to the fact that the numbers for importance to the student participants yet also
each condition were randomly generated and resembles climate surveys typically
compiled in the same list, a slightly different administered to personnel operating in an
numbers of participants were assigned to each of employment setting. The SACE is composed of
the three conditions. All three groups completed 58 closed-ended items which comprise five
Open-ended comments 5
categories: Instructional Services, Student question which used the same wording to ask
Services, Administrative and Physical Services, about open-ended item importance. Example
Student Focus, and Social and Cultural Services. items include “It was okay to skip the open-
Participants were asked to indicate their level of ended questions" (reverse scored) and
satisfaction with various elements of the college “Answering the closed-ended questions was
environment using a five-point Likert-type scale important." Participants were asked to indicate
ranging from 1 (very dissatisfied) to 5 (very their level of agreement with these items using a
satisfied). An example item is “The quality of five-point Likert-type scale ranging from 1
instruction in my field of interest.” The survey (strongly disagree) to 5 (strongly agree).
was programmed so that none of the respondents
were explicitly required to complete these Usable Comment Coding
questions. Finally, each open-ended comment was
Open-ended items. One comment block coded as usable (coded as 1) or non-usable
followed each of the five categories on the (coded as 2). The following are examples of
SACE (Baker, 1997), resulting in a total of five non-usable responses: “no opinion,” “no
open-ended questions. For example, the open- additional comment,” “skip.”
ended item related to Instructional Services read:
“Please provide any additional comments about Results
the instructional services at your institution.”
Table 2 provides descriptive statistics and
Follow-Up Evaluation Questionnaire correlations among the dependent measures
After respondents completed the SACE, assessed via the follow-up evaluation
they were advanced to a web page where they questionnaire. Hypothesis 1 predicted that open-
were asked if they wanted to complete a follow- ended questions would be viewed as less
up questionnaire containing 22 items assessing important than closed-ended questions in a
the dependent measures examined in this study. standard Web-based survey. To examine this
Almost all participants (96%) who completed prediction, a t-test for two dependent samples
the SACE also opted to complete the follow-up was conducted using responses from the control
evaluation. Table 2 provides descriptive group (N=100) only. The two treatment groups
statistics and coefficient alpha values for the were omitted from this analysis because we were
constructs assessed on the follow-up evaluation. interested in the manner in which individuals
Reactions. Affective reactions to the survey responding to “typical” web-based surveys (i.e.,
were measured with four items which asked surveys where open-ended responses are neither
participants to indicate their level of agreement prompted nor required) view the importance of
using a five-point Likert scale ranging from 1 the two types of items. Results demonstrated
(strongly disagree) to 5 (strongly agree). An that open-ended questions were viewed as less
example item is “I enjoyed completing this important (M=3.39, SD=.69) than closed-ended
survey.” Responses to the four items were questions (M=3.89, SD=.59), t(99)= -6.67,
averaged to create an affective reactions scale p<.001, indicating support for Hypothesis 1. An
score for each respondent. alpha level of .05 was used for all statistical
View of item importance. To assess the tests.
degree to which participants viewed the survey Hypothesis 2 predicted that requiring
items as important, two variables were respondents to complete open-ended questions
measured: view of open-ended item importance would adversely impact affective reactions to
(five items, which were averaged) and view of the survey. Meanwhile, Hypothesis 3 predicted
closed-ended item importance (five items, which that both requiring and prompting respondents to
were also averaged). The items measuring answer open-ended questions would increase
perceptions of open- and closed ended item perceptions of open-ended item importance. See
importance mirrored each other. That is, for each Table 3 for a summary of the affective reaction
question assessing perceptions of closed-ended and perceived importance scores per condition.
item importance there was a corresponding
Open-ended comments 6
researchers can design surveys so that comments hold value for organizations restricted to paper
are either required or simply prompted. surveys. Future research should examine
This study also shows that when given methods for encouraging responses to open-
appropriate cues, more participants provide ended items on paper surveys.
useful qualitative information than when these
cues are absent. This suggests that participants
are not skipping open-ended questions because
they do not have a legitimate answer as
suggested by Dillman (2000), but rather
skipping may be better understood as a form of
satisficing. Survey consumers should be
cautious when interpreting open-ended
responses that are not explicitly required or
encouraged since it is likely that the
nonrespondents to these questions have opinions
they are not providing.
Bosnjak, M., & Tuten, T. L. (2001). Classifying Kraut, A. I. (Ed.). (1996). Organizational
response behaviors in web-based surveys: Tools for assessment and
surveys. Journal of Computer-Mediated change. San Francisco: Jossey-Bass.
Communication, 6(3). Available from
http://jcmc.indiana.edu/vol6/issue3/bozn Kulesa, P., & Bishop, R.J. (2006). What did they
jak.html really mean? New and Emerging
methods for analyzing themes in open-
Bosnjak, M., Tuten, T. L., Wittman, W. W. ended comments. In A. I. Kraut (Ed.),
(2005). Unit (non)response in web- Getting action from organizational
based access panel surveys: An surveys: New concepts, technologies,
extended planned-behavior approach. and applications (pp. 238-263). San
Psychology and Marketing, 22(6), 489- Francisco: Jossey-Bass.
505.
Mossholder, K. W., Settoon, R. P., Harris, S. G.,
Church, A. H., & Waclawski, J. (1998). & Armenakis, A. A. (1995). Measuring
Designing and using organizational emotion in open-ended survey
surveys. Vermont, NH: Gower. responses: An application of textual data
analysis. Journal of Management, 21,
Couper, M. P. (2000). Web surveys: A review of 335-355.
issues and approaches. Public Opinion
Quarterly, 64, 464-494. Poncheri, R. M., Lindberg, J. T., Thompson, L.
F., & Surface, E. A. (in press). A
DeMay, C. C., Chandonnet, A., & Fenlason, K. comment on employee surveys:
J. (2005, April). Realizing the full Negativity bias in open-ended
potential of open-ended comments: responses. Organizational Research
Leveraging online technology. In K. J. Methods.
Fenlason (Chair), Comments: Where
have we been? Where are we going? Schuman, S., & Scott, J. (1989). Response
Symposium conducted at the 20th effects over time: Two experiments.
annual conference of the Society for Sociological Methods and Research, 17,
Industrial and Organizational 398-408.
Psychology, Los Angeles, California.
Open-ended comments 9
Table 1
N M SD
Past Experiences with Surveys
Technological Evaluation
Note. Past experiences with surveys response scale: 1 = never, 2 = rarely, 3 = sometimes, 4 = often, 5 = very often.
Technological evaluation response scale: 1 = strongly disagree, 2 = disagree, 3 = neither disagree nor agree, 4 =
Table 2
M SD 1 2 3
1. Affective Reactions 3.93 .63 (.85)
Note. N = 293. This N represents the number of individuals who provided responses to all items on these three
scales; ** Correlation is significant at the .01 level. Response scale: 1 = strongly disagree, 2 = disagree, 3 = neither
disagree nor agree, 4 = agree, 5 = strongly agree. Coefficient alpha values for each scale are provided on the
diagonal.
Open-ended comments 12
Table 3
Closed-Ended Item
102 3.89 .59 99 3.75 .53 98 4.11 .60
Importance
Open-Ended Item
101 3.39 .69 98 4.06 .57 98 3.69 .61
Importance
a
Sample size varied across conditions because respondents who failed to rate all of the items comprising a given
Note. Response scale: 1 = strongly disagree, 2 = disagree, 3 = neither disagree nor agree, 4 = agree, 5 = strongly
agree.
Open-ended comments 13
100 Usable
Non-Usable
80
Number of Comments
60
40
20
0
Control Required Comment Prompted Comment
Note. An omnibus test [Χ2(2, N = 312) = 33.65, p<.001] indicated that this study’s non-response
Services. Three 2 x 2 follow-up chi-square tests of independence showed that participants in the
required group and participants in the prompted group were more likely to provide usable
comments than those in the control group. No significant differences were found between the
100 Usable
Non-Usable
80
Number of Comments
60
40
20
0
Control Required Comment Prompted Comment
Note. An omnibus test [Χ2(2, N = 312) = 54.27, p<.001] indicated that this study’s non-response
follow-up manipulation affected the number of usable comments pertaining to Student Services.
Three 2 x 2 follow-up chi-square tests of independence showed that participants in the required
group and participants in the prompted group were more likely to provide usable comments than
those in the control group. No significant differences were found between the required and
prompted groups.
Open-ended comments 15
100 Usable
Non-Usable
80
Number of Comments
60
40
20
0
Control Required Comment Prompted Comment
Note. An omnibus test [Χ2(2, N = 312) = 52.49, p<.001] indicated that this study’s non-response
and Physical Services. Three 2 x 2 follow-up chi-square tests of independence showed that
participants in the required group and participants in the prompted group were more likely to
provide usable comments than those in the control group. A significant difference was found
between the required and prompted groups for this dimension, such that respondents in the
required comment group were morel likely to provide usable comments than respondents in the
100 Usable
Non-Usable
80
Number of Comments
60
40
20
0
Control Required Comment Prompted Comment
Note. An omnibus test [Χ2(2, N = 312) = 54.37, p<.001] indicated that this study’s non-response
follow-up manipulation affected the number of usable comments pertaining to Student Focus.
Three 2 x 2 follow-up chi-square tests of independence showed that participants in the required
group and participants in the prompted group were more likely to provide usable comments than
those in the control group. No significant differences were found between the required and
prompted groups.
Open-ended comments 17
100 Usable
Non-Usable
80
Number of Comments
60
40
20
0
Control Required Comment Prompted Comment
Note. An omnibus test [Χ2(2, N = 312) = 53.94, p<.001] indicated that this study’s non-response
follow-up manipulation affected the number of usable comments pertaining to Social and
participants in the required group and participants in the prompted group were more likely to
provide usable comments than those in the control group. No significant differences were found