Loading [Contrib]/a11y/accessibility-menu.js
Skip to main content
Survey Practice
  • Menu
  • Articles
    • Articles
    • Editor Notes
    • In-Brief Notes
    • Interview the Expert
    • Recent Books, Papers, and Presentations
    • All
  • For Authors
  • Editorial Board
  • About
  • Issues
  • Blog
  • Subscribe
  • search

RSS Feed

Enter the URL below into your favorite RSS reader.

http://localhost:22663/feed
Articles
Vol. 7, Issue 2, 2014March 31, 2014 EDT

Measuring Propensity to Join the Military: Survey Data are Consistent Regardless of Response Option Order

Jennifer Romano Bergstrom, Jennifer L Gibson, Joy Hackenbracht,
militarypropensitysurvey responsesresponse option order
https://doi.org/10.29115/SP-2014-0009
Survey Practice
Romano Bergstrom, Jennifer, Jennifer L Gibson, and Joy Hackenbracht. 2014. “Measuring Propensity to Join the Military: Survey Data Are Consistent Regardless of Response Option Order.” Survey Practice 7 (2). https:/​/​doi.org/​10.29115/​SP-2014-0009.
Save article as...▾
Download all (1)
  • Figure 1  No difference in self-reported intention to join, by response option order.
    Download

Sorry, something went wrong. Please try again.

If this problem reoccurs, please contact Scholastica Support

Error message:

undefined

View more stats

Abstract

Measuring Propensity to Join the Military: Survey Data are Consistent Regardless of Response Option Order

Introduction

Subtle differences in response option order can affect survey data (Krosnick and Alwin 1987; Schuman and Presser 1981; Schwarz et al. 1985, 1991; Tourangeau, Rips, and Rasinski 2000; see Schwarz and Hippler 1991, for a review). Survey respondents expect response options to follow a hierarchical format, such as decreasing in importance (i.e., left and top mean first), and they take longer to respond when the format is inconsistent with this expectation (Tourangeau, Couper, and Conrad 2004). These expectations persist regardless of culture, primary language, and whether people are accustomed to increasing or decreasing scales (Chan 1991; Toepoel et al. 2009). Respondents have consistently demonstrated a bias to items presented first – for example, respondents rate the reputation of their college higher when Strongly Agree is on the left (Friedman, Herksovitz, and Pollack 1994), respondents rate items as “describ[ing] me very well” more often when this option appears first (Chan 1991), and participants look at and select the first few options in a list of response options more often than items at the end of the list (Galesic et al. 2008). Thus, design decisions can have large implications for survey data validity.

The intention to join the military is often measured using a single ordinal item, “What is the likelihood that you will serve in the Military in the next few years?”(e.g., Ford et al. 2013) For the sake of convention and comparability across studies, the response options are typically presented in the following order: Definitely, Probably, Probably Not, and Definitely Not. However, this convention may affect data when the other items in the survey are displayed in the opposite order: Definitely Not, Probably Not, Probably, and Definitely. From a user’s experience perspective, the inconsistent direction of the response options can pose a problem, as people may not recognize that the order has changed, and thus, they may select the incorrect response option. The present study examines this concern by manipulating response option order on the intention to join the military item on a web-based survey.

The intention item was embedded within a survey comprised entirely of items with negative response options listed first. Two versions were fielded: one in which Definitely was presented first and one in which Definitely Not was first. We hypothesized that the “negative first” convention would hold for successive items, and consequently, respondents would select the left-most option more often, and the resulting data would show that intention to join is higher when Definitely is presented first compared to when Definitely Not is presented first.

Method

Participants

Respondents (ages 16 to 24) first completed a self-administered paper survey about media habits and opinions. They were selected using address-based probability sampling, and the entire sample was weighted to be representative of the U.S. population in terms of gender, race/ethnicity, Census division, and education. Respondents refusing more than 75 percent of the survey items were excluded, and removal of these cases occurred before weights were computed.

All respondents who reported using social media were then invited to participate in a follow-up web-based survey, in which the experiment was embedded. There were 1686 young adults (Mage=20.27, SE=0.03) who participated. The sample was mostly white (63 percent) and evenly split across gender (57 percent female).

Procedure

Among items about lifestyle habits and preferences, all participants were asked: In the next few years, how likely is it that you will be serving in the Military? Respondents were randomly assigned to one of two experimental conditions: one in which Definitely was presented as the first response option and one in which Definitely Not was presented first. In order to ensure an adequate sample for the primary study, 70 percent received the control condition, and 30 percent received the experimental condition; thus, there were more participants in the control condition, in which Definitely appeared first (see Table 1 for participant demographics). The power to detect a small effect under a 0.05 alpha level was 0.96, which exceeds the generally accepted cutoff representing sufficient power, 0.80.

Table 1  Participant demographics.
Definitely first Definitely not first
n 1184 502
Age M=19.7, SD=2.5 M=19.6, SD=2.3
Race/Ethnicity 67% White, Non-Hispanic 68% White, Non-Hispanic
9% Black, Non-Hispanic 8% Black, Non-Hispanic
12% Hispanic 15% Hispanic
6% Asian 4% Asian
5% Other, Non-Hispanic 5% Other, Non-Hispanic
1% Refused 1% Refused
Gender 47% Male 49% Male

All analyses were conducted using Stata version 12.1 (StataCorp, 2011), and we accounted for the stratified survey design using the svy prefix. We compared the means and distributions of responses (including refusals) by conducting a series of between-group comparisons.

Results

We conducted an ordinary least squares (OLS) logistic regression and a two-way chi-square of experimental condition and propensity response. As can be clearly seen in Figure 1, regression results indicated that average propensity did not differ as a function of response option order. That is, intention to join the military was the same regardless of whether Definitely (M=1.40, SE=0.02) or Definitely Not (M=1.45, SE=0.03) was presented as the first response option, B=–0.04, t=1.07, p=0.28. Response order did not affect the distribution of responses or nonresponses, Ïx2 (4)=2.80, p=0.73, and most respondents selected Definitely Not, followed by Probably Not, regardless of where they appeared in the response option order.

Figure 1  No difference in self-reported intention to join, by response option order.

Conclusions

In this research note, we demonstrate that contrary to previous findings on the effect of response scale characteristics on survey data, response option order did not alter a single-item ordinal measure of intention to join the military. This finding has implications for surveys that use the same question but that may have different response scales. Our findings demonstrate no effects in this case – the order of response option presentation did not affect respondents’ selection for this item.

There are several limitations to our study. First, we did not have the opposite order experiment – one in which all the other items on the survey were positive first. It is not clear if swapping the presentation order of all the other items would yield a similar effect. Second, intentions to join the military may be less susceptible to response order effects than other psychological constructs. More malleable concepts, such as attitudes toward products and services, may be more susceptible to order effects. Third, we did not manipulate the axes (i.e., vertical vs. horizontal), and it is not clear if the “top means first” expectation would have been revealed with such a manipulation.

Future research should examine other constructs and whether the response order effect is insignificant with fixed constructs. Research may also examine the effect of other factors, such as the visual design of the response scale (Christian and Dillman 2004; Toepoel, Das, and Soest 2006), and the mode of data collection (Ye, Fulton, and Tourangeau 2011) on intentions to join the military and other constructs. Such factors may or may not influence fixed constructs.

Nevertheless, our study demonstrates that changing the order of response options for one item on a survey such that the order is inconsistent with the rest of the survey does not affect data quality. Despite the limited data we have provided here, our study is a small step in the direction of identifying visual modifications that may or may not affect survey data quality.

Author Note

This report is released to inform interested parties of research and to encourage discussion.

Any views expressed on methodological issues are those of the authors and not necessarily those of Fors Marsh Group. This research was based on data from the U.S. Department of Defense. The views, opinions, and findings contained in this paper are solely those of the authors and should not be construed as an official Department of Defense position, policy, or decision, unless so designated by other documentation.

References

Chan, J.C. 1991. “Response Order Effects in Likert Type Scales.” Educational and Psychological Measurement 51 (3): 532–40.
Google Scholar
Christian, L.M., and D.A. Dillman. 2004. “The Influence of Graphical and Symbolic Language Manipulations on Responses to Self-Administered Questions.” Public Opinion Quarterly 68 (1): 58–81.
Google Scholar
Ford, M.T., J.L. Gibson, A.L. DeCasare, S.M. Marsh, and B.K. Griepentrog. 2013. “Pre-Entry Expectations, Attitudes, and Intentions to Join Predict Military Tenure.” Military Psychology 25 (1): 36–45.
Google Scholar
Friedman, H.H., P.J. Herksovitz, and S. Pollack. 1994. “Biasing Effects of Scale-Checking Styles on Responses to a Likert Scale.” In Proceedings of the American Statistical Association Annual Conference: Survey Research Methods, 792–95.
Google Scholar
Galesic, M., R. Tourangeau, M.P. Couper, and F.D. Conrad. 2008. “Eye-Tracking Data. New Insights on Response Order Effects and Other Cognitive Shortcuts in Survey Responding.” Public Opinion Quarterly 72 (5): 892–913.
Google Scholar
Krosnick, J., and D.F. Alwin. 1987. “An Evaluation of a Cognitive Theory of Response-Order Effects in Survey Measurement.” Public Opinion Quarterly 51 (2): 201–19.
Google Scholar
Schuman, H., and S. Presser. 1981. Questions and Answers in Attitude Surveys: Experiments on Question Form, Wording, and Context. Orlando, FL: Academic Press, Inc.
Google Scholar
Schwarz, N., and H. Hippler. 1991. “Response Alternatives: The Impact of Their Choice and Presentation Order.” In Measurement Errors in Surveys, edited by P.P. Biemer, R.M. Groves, L.E. Lyberg, N.A. Mathiowetz, and S. Sudman. New York: Wiley.
Google Scholar
Schwarz, N., H.J. Hippler, B. Deutsch, and F. Strack. 1985. “Response Categories: Effects on Behavioral Reports and Comparative Judgments.” Public Opinion Quarterly 49 (3): 388–95.
Google Scholar
Schwarz, N., B. Knäuper, H.J. Hippler, E. Noelle-Neumann, and L. Clark. 1991. “Rating Scales: Numeric Values May Change the Meaning of Scale Labels.” Public Opinion Quarterly 55 (4): 570–82.
Google Scholar
Toepoel, V., M. Das, and A. Soest. 2006. Design of Web Questionnaires: The Effect of Layout in Rating Scales. Tilburg University: CentERdata Discussion Paper No. 2005-114.
Google Scholar
Toepoel, V., C.M. Vis, J.W.M. Das, and A.H.O. van Soest. 2009. “Design of Web Questionnaires: An Information-Processing Perspective for the Effect of Response Categories.” Sociological Methods and Research 37 (3): 371–92.
Google Scholar
Tourangeau, R., M.P. Couper, and F. Conrad. 2004. “Spacing, Position, and Order: Interpretive Heuristics for Visual Features of Survey Questions.” Public Opinion Quarterly 68:368–93.
Google Scholar
Tourangeau, R., L. Rips, and K. Rasinski. 2000. The Psychology of Survey Response. Cambridge, England: Cambridge University Press.
Google Scholar
Ye, C., J. Fulton, and R. Tourangeau. 2011. “More Positive or More Extreme? A Meta-Analysis of Mode Differences in Response Choice.” Public Opinion Quarterly 75 (2): 349–65.
Google Scholar

This website uses cookies

We use cookies to enhance your experience and support COUNTER Metrics for transparent reporting of readership statistics. Cookie data is not sold to third parties or used for marketing purposes.

Powered by Scholastica, the modern academic journal management system