Processing math: 100%
Skip to main content
Survey Practice
  • Menu
  • Articles
    • Articles
    • Editor Notes
    • In-Brief Notes
    • Interview the Expert
    • Recent Books, Papers, and Presentations
    • All
  • For Authors
  • Editorial Board
  • About
  • Issues
  • Blog
  • Subscribe
  • search

RSS Feed

Enter the URL below into your favorite RSS reader.

http://localhost:44846/feed
Articles
Vol. 13, Issue 1, 2020August 06, 2020 EDT

Interviewer Face Coverings and Response to Personal Visit Surveys: A Case Study of the 2020 U.S. Census

Nancy Bates, Laura Kail, Amanda Price,
public attitudesus censusinterviewer effects
https://doi.org/10.29115/SP-2020-0011
Photo by Gayatri Malhotra on Unsplash
Survey Practice
Bates, Nancy, Laura Kail, and Amanda Price. 2020. “Interviewer Face Coverings and Response to Personal Visit Surveys: A Case Study of the 2020 U.S. Census.” Survey Practice 13 (1). https:/​/​doi.org/​10.29115/​SP-2020-0011.
Save article as...▾

View more stats

Abstract

The survey research literature has demonstrated how interviewer characteristics can affect cooperation during interviewer-administered surveys. These studies suggest that characteristics such as gender, race/ethnicity, and age may be predictive of a sample respondent’s decision to participate in surveys. At least one other study suggests that what the interviewer is wearing can also impact how respondents behave. Given the controversy of wearing masks in the time of COVID-19, we explore the notion of face coverings worn by enumerators and whether this may influence a household’s decision to participate in the 2020 U.S. Census nonresponse follow-up operation.

1. Introduction

The survey research literature has demonstrated how interviewer characteristics can affect cooperation during interviewer-administered surveys (Davis et al. 2010; Lavrakas 2008; West and Blom 2016). These studies suggest that characteristics such as gender, Race and Hispanic origin, and age may be predictive of a sample respondent’s decision to participate in surveys. At least one other study suggests that what the interviewer is wearing can also impact how respondents behave (Benstead 2014). In this article, we explore the notion of face coverings worn by enumerators and whether this may influence a household’s decision to participate in the 2020 U.S. Census nonresponse follow-up operation.

Given the COVID-19 pandemic and adjusted timelines, in July 2020, the 2020 U.S. Census was winding down the self-response phase and preparing for personal-visit nonresponse follow-up, slated to begin in August. During this phase, census enumerators go door-to-door to interview households that have not self-responded. Census management sought information on public attitudes regarding census enumerators’ use of face coverings when contacting households. As the pandemic continued, the use of face coverings in public became a polarizing topic, with mandates to do so that varied from state to state and place to place (Flood 2020; Kim, Andrew, and Froio 2020; Stanley-Becker 2020; Wilson 2020).

2. Methodology

To understand the ramifications of enumerators wearing masks, a question on the topic was added to the 2020 Census Attitudes Daily Tracking Survey. This survey collected data on awareness, attitudes, and intent to participate in the U.S. 2020 Census and consisted of a probability random-digit-dial phone survey and a nonprobability online survey. The analysis in this article draws from the nonprobability, online survey responses collected over four weeks (June 12 – July 8). Studies suggest that whether a question is posed positively (e.g., 'Gay marriage should be allowed) or negatively ('Gay marriage should be banned) can systematically affect answers (Bregje et al. 2016; Hippler and Schwarz 1986). Consequently, we included a questionnaire design experiment.

The online survey leveraged respondents from two independent web panels, Dynata and ThinkNow, a Hispanic-focused panel. Dynata recruited respondents via direct email in partnership with globally recognized brands, while ThinkNow recruited panel participants in Spanish from targeted Google Ad campaigns and other Spanish-language websites. Quotas were set for the number of respondents originating from each panel. Hispanic English and Spanish speaking respondents, which make up 15% of the quota, originated from ThinkNow and all other respondents originated from the Dynata panel. Responses from both panels were combined, processed, and weighted together. Sample quotas ensured a balanced sample across the following key demographic characteristics: Age × Gender, Region, Race and Hispanic Origin, Age × Non-Hispanic White/Not Non-Hispanic White, and Education. Since the online survey is a nonprobability opt-in survey, a response rate cannot be calculated. However, the American Association for Public Opinion Research (AAPOR) participation rate was 34.4% (American Association for Public Opinion Research 2016).

The results discussed in this article are limited to findings from univariate and bivariate estimates. Bivariate tabulations compared each of the experimental questions across key independent variables of interest, including: age,[1] Race and Hispanic origin,[2] and region.[3] The tables also included the results of statistical hypothesis tests, which were performed pairwise on all combinations of pairs within a given independent variable. Hypothesis tests show differences using a two-proportion t-test at an α=0.05 significance level, adjusting for multiple comparisons within a single independent variable using the Bonferroni method.[4] Variance estimations use the Taylor Linearization estimator.

Persons indicating they had not yet self-responded to the 2020 Census at the time of the tracking survey were asked the face covering questions (approximately 17–24% indicated they had not yet self-responded, depending upon date of interview). Experimental question wording was administered using a randomized split panel, whereby half of the sample got the “wearing a face mask” positive version and half got the “NOT wearing a face mask” negative wording (see below).

Version 1 - Positive: "If the census taker came to your door and was wearing a face mask, would you be more likely to respond, less likely to respond, or would it not make a difference?

Version 2 – Negative: “If the census taker came to your door and was NOT wearing a face mask, would you be more likely to respond, less likely to respond, or would it not make a difference?”

3.1 Results of the split panel

Table 1 presents results from the split panels that received either the positive “presence of a mask” or negative “absence of a mask” version. The percentage of respondents who reacted negatively to enumerators wearing masks (less likely to respond to Version 1 and more likely to respond to Version 2) was significantly different between the two versions. When presented with the scenario of enumerators wearing a face covering, approximately 22% indicated they were less likely to respond compared with just under 12% who indicated they were more likely to respond if the enumerator was not wearing a face covering. However, once combined, those who reacted favorably to face coverings or said it would not make a difference were large majorities, regardless of the version asked (around 77% for Version 1 and 88 % for Version 2).

Table 1:Results of questionnaire wording experiment (standard errors in parentheses)
Version 1 – Wearing a Mask Version 2 – Not Wearing a Mask
More likely to respond 34.76 (1.93) Less likely to respond 57.10 (2.00)
Less likely to respond 21.66 (1.64)* More likely to respond 11.49 (1.26)*
Would not make a difference 42.63 (2.00) Would not make a difference 30.91 (1.90)
Refusal 0.95 (0.34) Refusal 0.50 (0.26)
N 789 N 782

* Indicates pairwise statistically significant difference

3.2 Results by demographics and region

For simplicity, we present data from the “presence of a mask” version only when examining age, Race and Hispanic origin, and region. Likelihood to respond across these characteristics was similar for both question versions. For all age groups, the majority reacted favorably to face-coverings or said it would not make a difference (see Table 2). Compared to 45–64-year-olds, significantly more 18–34-year-olds said they were less likely to respond if a census taker was wearing a mask. For the “not wearing a mask” question, +65-year-olds were significantly more likely to say they would be less likely to respond if the enumerator was not wearing a mask compared to 35–44-year-olds (74% vs. 49%, data not shown).

Table 2:Likelihood of responding to enumerators wearing a face covering by age (standard errors in parentheses)
  18–34 35–44 45–64 65+
Subgroup* 1 2 3 4
More likely to respond 35.52 (2.59) 31.05 (4.46) 38.54 (4.40) 27.42 (6.39)
Less likely to respond 27.10 (2.46)3 18.75 (3.47) 14.32 (2.91)1 13.84 (5.16)
Would not make a difference 36.08 (2.60)4 49.50 (4.80) 47.14 (4.45) 56.58 (7.29)1
Refusal 1.30 (0.54) 0.70 (0.70) 0.00 (0.00) 2.16 (2.13)
N 426 146 157 60

* Each subgroup has a value of 1 through 4. Pairwise statistical significance is denoted by a superscript with a value of 1 through 4. The superscript indicates the value of the subgroup from which the estimate significantly differs.

Table 3:Likelihood of responding to enumerators wearing a face covering by Race and Hispanic origin (standard errors in parentheses)
  Hispanic, any Race Non-Hispanic White Non-Hispanic Black Non-Hispanic Asian
Subgroup 1 2 3 4
More likely to respond 36.22 (4.57) 35.20 (2.65) 36.16 (5.07) 37.18 (6.47)
Less likely to respond 21.94 (3.66) 19.71 (2.22) 33.62 (5.00) 21.50 (5.35)
Would not make a difference 40.90 (4.70) 44.07 (2.74)3 28.54 (4.70)2 41.32 (6.75)
Refusal 0.94 (0.94) 1.02 (0.47) 1.68 (1.19) 0.00 (0.00)
N 144 394 118 82

Over one-third of each Race and Hispanic origin group reacted favorably to enumerators wearing masks (see Table 3). Non-Hispanic Black respondents were significantly less likely to say it does not make a difference (compared to Non-Hispanic Whites). When asked as “not wearing a mask,” significantly more Non-Hispanic Asians said they were less likely to respond if the enumerator was not wearing a mask (compared to Non-Hispanic Whites; 72 % vs. 52%, data not shown).

Table 4:Likelihood of responding to enumerators wearing a face covering by region (standard errors in parentheses)
  Northeast Midwest South West
Subgroup 1 2 3 4
More likely to respond 33.42 (4.29) 37.48 (4.46) 36.01 (3.05) 31.55 (4.10)
Less likely to respond 22.37 (3.82) 25.16 (3.84) 20.76 (2.50) 19.73 (3.52)
Would not make a difference 41.41 (4.61) 36.75 (4.29) 43.04 (3.15) 47.66 (4.39)
Refusal 2.80 (1.41) 0.61 (0.61) 0.19 (0.19) 1.07 (0.77)
N 154 148 325 162

Finally, Table 4 indicated a large majority of respondents who were from all four regions of the country were more likely to respond or say it would not make a difference if the enumerator had a face covering. Depending upon the region, around 20–25% indicated they would be less likely to respond if the census enumerator wore a mask, with no significant differences between them.

4. Conclusions

Survey organizations that conduct personal-visit interviews are experiencing major disruptions to “business as usual” during the COVID-19 pandemic. This article reports findings from an online tracking survey to inform about public opinion regarding enumerators’ use of face coverings. We embedded a split-panel questionnaire wording experiment as a randomized controlled trial, but several limitations are noted. First, the online responses were generated from a nonprobability opt-in sample. As a result, while the web sample enacts quotas and weights to reflect the general population, unknown biases may undermine the true “representativeness” of the results. Second, although respondents were randomly assigned to one of the two questions, there were a few underlying differences in the sample composition for each question. For example, significantly more males aged 25–34 saw the positively worded question. Otherwise, there were no significant differences between the samples (for the demographics we examine here). Finally, the data were collected over a four-week period as mandates to wear masks were changing rapidly across the U.S. with some areas adding the requirement over the course of data collection.

Results of the split-panel questionnaire wording experiment suggest that likelihood to respond during 2020 Census nonresponse follow-up would be higher (or not influenced) if census enumerators make their door-to-door contacts wearing face coverings. This was true across age groups, Race and Hispanic origin, and region of the country. A few within-subgroup differences were found, for example, compared to Non-Hispanic Whites, Non-Hispanic Blacks expressed less ambivalence toward face coverings. We also found significantly fewer respondents reacted favorably to the absence of face coverings when the question was posed in the negative direction (enumerators not wearing masks). This is in line with negativity bias literature suggesting that respondents have more difficulty endorsing topics when presented in a negative direction. Regardless, a sizeable minority overall (around 20–25%) still indicated lower likelihood of cooperating with enumerators wearing face coverings. Survey practitioners should consider these findings as organizations look for ways to move forward with personal-visit interviewing in the COVID-19 world.

Contact information

nancy.a.bates@census.gov
301.763.5248
U.S. Census Bureau, Associate Directorate for Research and Methodology Room 5K140, Washington DC 20233


  1. Based on their response to the age question, each respondent was assigned to one of the following four age range categories: 18–34, 35–44, 45–64, or 65+.

  2. Based on their response to race and Hispanic origin questions, each respondent was assigned to one of five race and Hispanic origin categories: Hispanic of any race or non-race, Non-Hispanic (NH) White only, NH Black/African American only, NH Asian only, or NH Other (races such as NH Native Hawaiian and Pacific Islander, NH American Indian and Alaska Native only, other races, and multiple races). NH Other was excluded from this analysis due to small sample sizes.

  3. Based on their response to the zip code question, each respondent was assigned to one of the following four geographic regions: Northeast, Midwest, South, or West.

  4. Because multiple comparisons induce an increased likelihood of committing Type I errors, a Bonferroni correction was applied to each alpha value to ensure an accurate assessment of significance. To operationalize the Bonferroni adjustment, a p-value, pi, for a given hypothesis test is adjusted by taking the product m∗pi, where m is all combinations of pairs within a given independent variable for a given table. The null hypothesis is rejected using a Bonferroni adjustment if p-value ≤ α/m. It was operationalized by using: m*p-value ≤ 0.05.

References

American Association for Public Opinion Research. 2016. “Standard Definitions: Final Dispositions of Case Codes and Outcome Rates for Surveys.” https:/​/​www.aapor.org/​AAPOR_Main/​media/​publications/​Standard-Definitions20169theditionfinal.pdf.
Benstead, Lindsay J. 2014. “Does Interviewer Religious Dress Affect Survey Responses? Evidence from Morocco.” Politics and Religion 7 (4): 1–27. https:/​/​doi.org/​10.1017/​s1755048314000455.
Google Scholar
Bregje, Holleman, Naomi Kamoen, André Krouwel, Jasper van de Pol, and Claes de Vreese. 2016. “Positive vs. Negative: The Impact of Question Polarity in Voting Advice Applications.” Edited by Alexander N. Sokolov. PloS One 11 (10): e0164184. https:/​/​doi.org/​10.1371/​journal.pone.0164184.
Google ScholarPubMed CentralPubMed
Davis, R. E., M. P. Couper, N. K. Janz, C. H. Caldwell, and K. Resnicow. 2010. “Interviewer Effects in Public Health Surveys.” Health Education Research 25 (1): 14–26. https:/​/​doi.org/​10.1093/​her/​cyp046.
Google ScholarPubMed CentralPubMed
Flood, B. 2020. Fox News. https:/​/​www.foxnews.com/​media/​rush-limbaugh-democrats-mask-wearing-freaks-for-politicizing-coronavirus-face-covers.
Hippler, H., and N. Schwarz. 1986. “Not Forbidding Isn’t Allowing: The Cognitive Basis of the Forbid-Allow Asymmetry.” Public Opinion Quarterly 50 (1): 87–96. https:/​/​doi.org/​10.1086/​268961.
Google Scholar
Kim, A., S. Andrew, and F. Froio. 2020. Cable News Network. https:/​/​www.cnn.com/​2020/​06/​19/​us/​states-face-mask-coronavirus-trnd/​index.html.
Lavrakas, P.J. 2008. “Interviewer Effects.” In Encyclopedia of Survey Research Methods. Thousand Oaks, CA: Sage Publications, Inc. https:/​/​doi.org/​10.4135/​9781412963947.n242.
Stanley-Becker, I. 2020. “Mask or No Mask? Face Covering Become Tool in Partisan Combat.” Washington Post, 2020. https:/​/​www.washingtonpost.com/​politics/​in-virus-response-riven-by-politics-masks-are-latest-rorschach-test/​2020/​05/​12/​698477d4-93e6-11ea-91d7-cf4423d47683_story.html.
West, Brady T., and Annelies G. Blom. 2016. “Explaining Interviewer Effects: A Research Synthesis.” Journal of Survey Statistics and Methodology 5 (2): 175–211. https:/​/​doi.org/​10.1093/​jssam/​smw024.
Google Scholar
Wilson, R. 2020. “Politicians, Scientists Back Masks: They Work.” The Hill, 2020. https:/​/​thehill.com/​policy/​healthcare/​public-global-health/​505453-politicians-scientists-back-masks-they-work.

This website uses cookies

We use cookies to enhance your experience and support COUNTER Metrics for transparent reporting of readership statistics. Cookie data is not sold to third parties or used for marketing purposes.

Powered by Scholastica, the modern academic journal management system