{"id":14659,"date":"2020-12-10T14:57:19","date_gmt":"2020-12-10T13:57:19","guid":{"rendered":"https:\/\/surveyinsights.org\/?p=14659"},"modified":"2023-07-13T06:53:05","modified_gmt":"2023-07-13T05:53:05","slug":"multi-mode-question-pretesting-using-traditional-cognitive-interviews-and-online-testing-as-complementary-methods","status":"publish","type":"post","link":"https:\/\/surveyinsights.org\/?p=14659","title":{"rendered":"Multi-mode question pretesting: Using traditional cognitive interviews and online testing as complementary methods"},"content":{"rendered":"<h1>Introduction<\/h1>\n<p>Ensuring that survey questions are consistently understood, and that respondents are able and willing to answer them, is of critical importance for reducing measurement error and enhancing data quality. \u00a0Survey researchers have multiple question pretesting methods available (Groves et al., 2009), each with their own strengths and limitations (D\u2019Ardenne &amp; Collins, 2019). Traditionally, pretesting studies have been conducted using interviewer-administered cognitive testing. But with the increased use of web-data collection, pretesting in self-administered, online modes has emerged (Behr, 2016). However, few pretesting studies have used both traditional interviewer-administered cognitive testing and online, self-administered testing as complementary approaches to address a single research question. This research describes a series of studies using both approaches, demonstrating how web-based pretesting can complement and expand on traditional pretesting.<\/p>\n<h1><strong>Questionnaire Pretesting <\/strong><\/h1>\n<p>Survey organizations often use traditional, interviewer-administered cognitive testing for questionnaire pretesting (Willis, 2005), collecting rich, qualitative data to gain insight into the response process and potential questionnaire problems. Recent technological advances have led to new and innovative, online testing approaches to obtain similar data to traditional cognitive interviews via self-administered web surveys (Behr, 2016), while Internet panels have made access to respondents readily available (Fowler &amp; Willis, 2019). Similar to the advantages gained when surveys moved from interviewer-administered to online, self-administered modes, online testing allows for questionnaire evaluation and pretesting with less cost and effort and more geographically diverse samples (Scanlon, 2019). Online testing also allows for analysis of paradata, which can provide additional insight into the response process (McCalin et al., 2018), and larger sample sizes required for experiments that are not feasible with traditional, interviewer-administered cognitive interviews.<\/p>\n<p>A number of considerations must be taken into account when designing pretesting research.\u00a0 People who participate in traditional cognitive interviews are often more cooperative than online panel respondents (Ross et al., 2010). Further, online panels may vary in their data quality (e.g., Hillyguys et al.. 2014; Matthijsse et al., 2015; Toepoel et al., 2008; Kennedy et al., 2016; Cornesse et al., 2020), but they tend to be more geographically diverse than samples recruited for traditional cognitive interviews (Berinsky et al., 2012; Scanlon, 2019).\u00a0 Both generally rely on convenience samples, which are appropriate for pretesting efforts seeking to gain insight into the respondents\u2019 thought processes rather than trying to predict survey outcomes or produce general-population estimates.<\/p>\n<p>A number of methodological choices also need to be considered when selecting a pretesting mode. In traditional cognitive interviewing, researchers have the flexibility to use unscripted probes to ask specific follow-up questions about the response processes. Online testing limits researchers to scripted probes that can be programmed into the instrument in advance and responses are sometimes shorter in online testing (Meitinger &amp; Behr, 2016; Lenzner &amp; Neuert, 2017). \u00a0These factors must be carefully considered when designing studies to best leverage the strengths of each method (Kaplan et al., 2019; McCarthy, 2019).<\/p>\n<h1><strong>Motivation<\/strong><\/h1>\n<p>The goal of this research was to use multiple pretesting modes (traditional cognitive interviews and online testing) in an iterative, multi-stage research design. Each study added value to the prior, while addressing limitations due to the mode or design, leading to a well-supported conclusion that would not have been possible had only one method been used.<\/p>\n<p>In 2015, the Federal Cybersecurity Enhancement Act was passed, requiring U.S. Federal government agencies to include language about cybersecurity screening of survey responses. Before implementation, BLS needed to understand respondents\u2019 reactions to and comprehension of the new confidentiality language and its potential impact on survey response. Although confidentiality language is technical and written in legalistic language, respondents do expect to see such disclaimers before participating in U.S. Federal surveys (Landreth et al., 2008). Additionally, this research was of critical importance, as confidentiality language can sometimes raise suspicion about responding to a survey (Singer et al., 1995), increase item and unit nonresponse (Reamer, 1979; Keusch et al., 2019), and overall trust in federal statistics (Childs et al., 2019).<\/p>\n<p>Study 1 used traditional, interviewer-administered cognitive interviews to assess qualitative reactions to, and comprehension of the pledge language. This approach is particularly effective to gauge baseline reactions of how former respondents perceived the pledge language. Study 2 used self-administered, online testing to validate and expand on the findings of Study 1 with a larger sample, examining completion rate and item nonresponse by pledge type, but could only assess hypothetical respondent behavior. Study 3 used an experimental design to validate and add value to the previous studies by assessing the impact of different pledge versions on participation over time. Results were combined to draw a well-supported conclusion about the potential impact of the new language. We discuss how using multiple pretesting modes and the addition of online testing can provide a fuller and more complete evaluation in survey pretesting.<\/p>\n<h1>Study 1<\/h1>\n<h2><strong>Overview<\/strong><\/h2>\n<p>Study 1 used traditional, interviewer-administered cognitive interviews to collect qualitative data from former BLS survey respondents. This study was exploratory, without formal hypotheses, and instead aimed to understand former respondents\u2019 recollection of how BLS keeps data confidential, and to gauge their reactions to, and comprehension of the original and revised pledge language (see Appendix A for the complete language).<\/p>\n<h2><strong>Methods<\/strong><\/h2>\n<p>A total of 23 former BLS survey respondents participated in semi-structured cognitive interviews via telephone lasting an average of 30 minutes, administered by two trained interviewers conducted over one month. Respondents were contacted from a list of prior BLS respondents who had recently responded to a BLS survey, and were asked to participate in a debriefing interview. Demographic or other quotas were not used, but efforts were made to get a mix of respondents who participated in various BLS surveys. Respondents were asked what they recalled about their previous participation in BLS surveys, the confidentiality assurances given, and whether they had any concerns about providing their information. Afterward, they were provided with the original pledge language verbally and visually (via email) followed by the revised pledge language. Concurrent probes were used to gauge respondents\u2019 initial reactions and comprehension of the language, whether the language would affect their decision to participate in a BLS survey, and any concerns with the language.<\/p>\n<h2><strong>Results<\/strong><\/h2>\n<h3><em>Reaction to pledge language<\/em><\/h3>\n<p>The open-ended nature of traditional cognitive testing allowed respondents to express their reactions to the pledge language. Most respondents (21 of 23) had no concerns about either the original or revised pledges, stating \u201cit sounds fine\u201d or \u201cit sounds like the standard language that always accompanies surveys or websites.\u201d When asked for their reactions to the revised pledge, the majority of respondents had a neutral or positive reaction. Many commented that additional cybersecurity protection was a good thing, that the data would be more secure. The few respondents who expressed concerns mentioned how broad and vague they found the language to be: \u201cIt\u2019s a pretty broad statement, you could do just about anything with it.\u201d<\/p>\n<h3><em>Comprehension of pledge language<\/em><\/h3>\n<p>Most respondents understood the pledges correctly, as assessed using a series of questions designed to evaluate participants\u2019 understanding of the data protection being offered (see Appendix B). All respondents correctly understood that their survey responses could be combined with survey answers from other respondents to create summary statistics and most understood that their data could not be shared with private companies.<\/p>\n<h3><em>Recollection of pledge language<\/em><\/h3>\n<p>Respondents were asked an open-ended question about what they recalled about their previous participation in BLS surveys. Most did not recall much about their previous participation. About half recalled the general topic of the survey, but none recalled the name of the survey.\u00a0 Most admitted not remembering what they were told about confidentiality, with only a few noting they thought the interviewer said something about keeping the data confidential.<\/p>\n<h3><em>Impact of pledge language <\/em><\/h3>\n<p>All participants indicated that they did not have any concerns about how BLS would keep their data confidential at the time they provided data. When asked about whether the pledge language provided would affect their decision to participate in a future BLS survey, about half felt it would not affect their decision either way, \u00a0and half felt it would have a positive impact on the decision to participate (e.g., \u201cI\u2019d feel more comfortable giving my data with the added security\u201d).<\/p>\n<h2>Discussion<\/h2>\n<p>Using traditional cognitive interviews allowed for the collection of qualitative data revealing that most respondents correctly understood and did not have serious concerns about the confidentiality language. Use of open-ended probes provided additional insight into the potential impact of the pledge language on future response to BLS surveys. \u00a0However, since these respondents had not only previously responded to a BLS survey, but also agreed to this study, these findings may indicate an overall comfort with BLS rather than providing insight into the effects of the pledge language.<\/p>\n<h1><strong>Study 2<\/strong><\/h1>\n<h2><strong>Overview <\/strong><\/h2>\n<p>Study 2 used online testing to expand on the findings from Study 1 with a larger sample who had not previously participated in a BLS survey. The larger sample and online mode allowed for experiments to assess paradata, such as differences between pledge versions on response to survey items and completion rates. For instance, if one pledge version was associated with lower completion rates and higher item nonresponse, it could be indicative of heightened concerns.\u00a0 Based on Study 1, it was hypothesized that participants would display similar reactions to and comprehension by pledge version, but since respondents may lack familiarity with BLS in general, they may have more concerns about confidentiality overall.<\/p>\n<h2><strong>Method<\/strong><\/h2>\n<p>Study 2 adapted and expanded Study 1 to an online, self-administered survey. Participants were recruited from Amazon\u2019s Mechanical Turk (MTurk) panel in the U.S. and compensated $0.75 for their participation. MTurk was used because it is a robust source of recruiting participants for evaluation and pretesting research quickly and efficiently (Mullinix et al., 2016), is comparable to other convenience samples typically used for pretesting and to quota samples limited to participants in a particular geographic region (Berinsky et al., 2012; Scanlon, 2019). Although panels such as MTurk may overrepresent some demographic groups (Fahimi et al., 2018; Chandler &amp; Shapiro, 2016), MTurk participants have been shown to put substantial effort into their responses and provide high quality data, though they only receive small token incentives (Anson, 2018). MTurk is also unique in that available tasks are typically categorization, coding, and transcribing, reducing the probability that MTurk workers are \u201cprofessional survey respondents\u201d (e.g., Hillygus et al., 2014).<\/p>\n<p>Participants completed questions about their prior survey experiences and were randomly assigned to either the original or revised pledge. Then they completed survey questions that appear on typical surveys (e.g., education, income) to explore the potential impact of the pledge on survey response, with the option to skip questions or select \u2018Prefer not to say.\u2019 Afterward, participants completed debriefing questions similar to those from Study 1.<\/p>\n<h2><strong>Results<\/strong><\/h2>\n<h3><em>Participants<\/em><\/h3>\n<p>A total of 1,128 (Mean age = 36 years; 78.0% White; 63.1% with a college degree or higher) MTurk participants started the study; 264 participants partially completed the study (less than 80% of the survey), and 864 completed the study. The survey took an average of 14 minutes to complete.<\/p>\n<h3><em>Reaction to pledge language<\/em><\/h3>\n<p>Like Study 1, a majority (78.8%) of participants indicated they would have no concerns with how BLS would keep their information confidential, regardless of pledge version (<em>p <\/em>&gt; .05).<\/p>\n<h3><em>Comprehension of pledge language<\/em><\/h3>\n<p>Using a series of closed-ended questions mirroring those from Study 1, participants\u2019 understanding of the data protection being offered by each pledge was assessed using a series of questions (similar to those in Appendix B). We found no differences between pledge versions on beliefs about who could access information participants provided on a BLS survey (all <em>p<\/em> &gt; .05). Participants who read the original pledge (46.0%) were more likely than those who read the revised pledge (38.3%) to incorrectly think that their data from a BLS survey could be given to other statistical agencies for policy making, (\u03c72, = 6.43, df = 2, <em>p<\/em> = .04). No other differences were found (all <em>p<\/em> &gt; .05). Thus, most participants understood the intended meaning of both pledges.<\/p>\n<h3><em>Recollection of pledge language<\/em><\/h3>\n<p>To mirror the Study 1 question about what prior BLS respondents recalled about their previous survey response, participants in Study 2 completed a free recall task where they listed as many words or phrases from the pledge as they could remember. On average, participants reported 5.8 words and a median of 4 words, with no difference by pledge version (<em>p<\/em> &gt; .05). Of the words reported, 62.0% were actually included in the pledge.<\/p>\n<p>Participants then completed a recognition task where they read a list of 15 words and marked whether each word was part of the pledge they read. Some of the words were in both pledges (e.g., \u201cConsent\u201d) and others were not (e.g., \u201cRisk\u201d). Overall, participants correctly recognized an average of 75.0% of the words from the original pledge and 43.0% of the words from the revised pledge, with no difference in correct or incorrect recognition by pledge version (all <em>p<\/em> &gt; .05).<\/p>\n<h3><em>Impact of pledge language<\/em><\/h3>\n<p>A majority of participants (59.8%) across both pledge versions indicated that how BLS kept their data confidential would play a role in their decision to participate in BLS surveys. On average, participants spent 31.4 seconds on the page with the pledge language, with no difference between pledge versions (<em>p<\/em> &gt; .05) despite the revised pledge being one sentence longer.<\/p>\n<p>No effect of pledge version on survey completions was observed (<em>p<\/em> &gt;.05), with the overwhelming majority of participants who started the survey completing it, regardless of pledge version. Item nonresponse rates were extremely low; over 98% of participants answered each item on the survey, with no difference by pledge version (<em>p<\/em> &gt; .05). This was even the case for income questions (about 1% in this study), which often have very high item nonresponse rates of up to 20% (Bollinger et al., 2013).<\/p>\n<h2>Discussion<\/h2>\n<p>Using an online platform, Study 2 confirmed findings from Study 1. In both studies, participants understood the pledge correctly and did not have strong concerns about the pledge. \u00a0Study 2 added value via a larger sample size and collection of paradata.\u00a0 The former allowed for statistical comparison of the two pledge versions, and the latter allowed us to confirm that the pledge version did not impact survey completion, item nonresponse, or memory for the pledge language. However, both studies only collected data at one point in time and could not assess the actual impact of pledge language on survey participation over time. This is crucial, as many BLS and other surveys ask respondents to complete multiple waves of a survey. A third study was designed to address these limitations.<\/p>\n<h1><strong>Study 3<\/strong><\/h1>\n<h2><strong>Overview<\/strong><\/h2>\n<p>Study 3 used an online design with similar research questions as Study 2, but expanded to a longitudinal design to assess the impact of pledge language and survey participation over time. Additionally, given that the item nonresponse rate in Study 2 was lower than typically observed in federal surveys (Bollinger et al., 2013), language was added to the instrument to encourage participants to provide honest responses and use \u2018Don\u2019t know\u2019 or \u2018Prefer not to answer\u2019 options rather than give incorrect responses (de Leeuw et al., 2015). The instrument also asked for participants\u2019 commitment to provide quality data (Betts, 2016). Study 3 also assessed whether priming, or reminding participants about confidentiality concerns prior to or after reading the pledge, would affect survey outcomes (Joinson et al., 2008; Singer et al., 2005). Together, these interventions would increase the ability to detect any differences in survey outcomes by pledge version.<\/p>\n<h2><strong>Method<\/strong><\/h2>\n<p>Study 3 followed the Study 2 protocol closely, asking about prior survey experience and demographic questions. Participants were recruited from MTurk and compensated $1.50 per survey in a longitudinal design with two surveys administered one week apart. In the first survey, participants were randomly assigned to read either the original or revised pledge. Half of the participants completed a confidentiality concern questionnaire designed to remind them of prior confidentiality concerns (see Buchanan et al., 2007).\u00a0 The priming questionnaire was randomly assigned to be located either before or after they read the pledge. One week later, the same participants were recontacted to participate in a second survey. They were given a brief reminder of the first survey, then asked to complete the same recollection and comprehension tasks used in Study 2.<\/p>\n<h2><strong>Results <\/strong><\/h2>\n<p>A total of 610 MTurk participants (Mean age = 38.4 years; 51.7% male; 88.4% White; 66.8% with a college degree or higher) from the U.S. completed the first survey and 463 of those completed the second survey (Mean age = 39.3 years; 52.0% male), for a 76.0% retention rate between the two surveys. Survey 1 took 10.5 minutes and Survey 2 took 7.3 minutes on average.<\/p>\n<h3><em>Reaction to pledge language<\/em><\/h3>\n<p>Like Studies 1 and 2, the majority of participants (65.7%) indicated that they had no or very little concern about the confidentiality pledge, while 69.2% of participants indicated that the pledge would play a role in their decision to respond to a BLS survey, with no differences based on priming location or pledge version (all <em>p<\/em> &gt; .05).<\/p>\n<h3><em>Comprehension of pledge language<\/em><\/h3>\n<p>Like Studies 1 and 2, participants generally understood the pledge language. However, pledge comprehension differed based on priming location. Participants reminded about confidentiality concerns prior to reading the pledge (30.0%) were less likely than those reminded about confidentiality concerns after reading the pledge (40.0%) to incorrectly think that data from a BLS survey could be given to other statistical agencies for policy making, F(3, 410) = 4.53, <em>p<\/em> = .03.<\/p>\n<h3><em>Recollection of pledge language<\/em><\/h3>\n<p>In the free recall task, participants reported an average of 2.8 words with a median of 3 words, with no difference by pledge version (<em>p<\/em> &gt; .05). Participants recalled an average of 19.8% of words they viewed in the pledge, with no difference by pledge version or priming location (all <em>p<\/em> &gt; .05). In the recognition task, participants correctly recognized an average of 58.8% of the words from either pledge, with no differences by pledge version (all <em>p<\/em> &gt;.05). However, controlling for pledge version, participants primed about confidentiality before reading the pledge correctly recognized a greater proportion of pledge words (62.1%) than those primed at the end (55.6%), <em>t<\/em>(417) = 2.62, <em>p<\/em> = .01. Thus, participants\u2019 memory of the pledge language declined over the week, while those primed about confidentiality prior to reading the pledge correctly recognized more of the pledge language.<\/p>\n<h3><em>Impact of pledge language<\/em><\/h3>\n<p>Participants spent an average of 27 versus 35 seconds reading the original pledge and revised pledge, respectively (<em>p<\/em> &gt; .05). There was no interaction or main effect of priming location and pledge version on overall item nonresponse rates (all <em>p<\/em> &gt; .05). Figure 1 shows the item nonresponse rates (i.e., \u201cdon\u2019t know\u201d, \u201cprefer not to say\u201d, or skips) for individual survey items. Compared to 1% overall item nonresponse in Study 2, overall item nonresponse was 11% in Study 3, suggesting that respondents may have provided better quality data (i.e., opting to not answer a question rather than guess or providing most honest responses). No interaction or main effects of priming location and pledge language were found on retention between surveys (all <em>p<\/em> &gt; .05).<\/p>\n<p><a href=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Figure_1.png\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-15052\" src=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Figure_1-1024x793.png\" alt=\"\" width=\"500\" height=\"387\" srcset=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Figure_1-1024x793.png 1024w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Figure_1-300x232.png 300w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Figure_1-768x595.png 768w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Figure_1.png 1412w\" sizes=\"auto, (max-width: 500px) 100vw, 500px\" \/><\/a><\/p>\n<p>Table 1 displays results of a series of logistic regressions to determine if priming location and pledge version affected the odds of responding to demographic questions. Participants primed before reading the pledge had a 1.9 lower odds of responding to personal income and a 2.4 lower odds of responding to household income. Participants who read the revised pledge had a 1.7 lower odds of responding to household income. No interaction between priming location and pledge type was found and no other items were significant (all <em>ps<\/em> &gt; .05) These findings are consistent with the idea that reminding respondents about confidentiality, and longer confidentiality pledges, can increase item nonresponse to sensitive questions (Singer et al., 1995).<\/p>\n<p><strong>Table 1. Impact of Priming Location and Pledge Version on Odds of <\/strong><strong>Responding to Demographic Questions. <\/strong><strong>Priming location (0 = before pledge, 1 = after pledge), <\/strong><strong>Pledge type (0 = original, 1 = revised)<\/strong><\/p>\n<p><a href=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_1.png\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-15053\" src=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_1-678x1024.png\" alt=\"\" width=\"500\" height=\"755\" srcset=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_1-678x1024.png 678w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_1-199x300.png 199w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_1.png 710w\" sizes=\"auto, (max-width: 500px) 100vw, 500px\" \/><\/a><\/p>\n<p><a href=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_2.png\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-15054\" src=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_2.png\" alt=\"\" width=\"500\" height=\"296\" srcset=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_2.png 716w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Item_2-300x178.png 300w\" sizes=\"auto, (max-width: 500px) 100vw, 500px\" \/><\/a><\/p>\n<h1><strong>Discussion<\/strong><\/h1>\n<p>Study 3 built on the previous studies by examining survey response over time, allowing for an examination of whether pledge version affected survey attrition. It also added value to the prior studies by including analysis of the impact of pledge version and confidentiality concerns on participants\u2019 comprehension and recollection of the pledges, and the ability to detect differences in item nonresponse rates to sensitive questions. This provided valuable insight into the impact of the pledge version on response.<\/p>\n<h2>General Discussion<\/h2>\n<p>Researchers have a variety of approaches available for conducting questionnaire development and pretesting, including traditional, interviewer-administered cognitive interviews and self-administered, online testing.\u00a0 Several studies have compared these two pretesting modes with varying conclusions as to whether they should be considered interchangeable or complementary (Edgar et al., 2016; Behr, 2016). This research supports the approach of using both as complementary.\u00a0 By using both methods sequentially and designing studies to leverage the strengths of each while addressing the limitations of the prior study, we were able to draw conclusions that would not have been possible if only one method had been used.\u00a0 Table 2 outlines how the three studies built upon each other and added value to provide a more complete picture in evaluating the confidentiality language than would have been possible with a single mode alone.<\/p>\n<p>Starting with traditional, interviewer-administered cognitive interviews, it was possible to explore BLS respondents\u2019 reactions, comprehension, recollection, and potential impact of the confidentiality pledge language through open-ended probes. This study provided useful baseline-level results, showing that respondents had few concerns with BLS confidentiality language, but was limited in its small sample size, highly cooperative respondents, and measurement of only hypothetical respondent behaviour.<\/p>\n<p>To address these limitations, a second study was conducted online to capture information from a larger number of participants who had no prior relationship with BLS surveys. The larger sample size allowed for statistical evaluation of the two pledge versions to determine its impact on survey completion rates, while also collecting information about participant reactions, comprehension, recollection, and paradata including time spent reading the pledges.\u00a0 The study confirmed findings obtained from former BLS survey respondents, and added new insights about the impact of pledge wording on survey completion. However, participants were still more cooperative than typical survey respondents, as evidenced by low item nonresponse.\u00a0 Additionally, since the first two studies were one-time collections, no information was available about the impact of pledge wording on participation in subsequent collection waves.<\/p>\n<p>To address these limitations, a third study was designed using similar methods to the previous ones to assess the effect of pledge version on reactions, comprehension, recollection, and impact. The study added value by using a longitudinal design to assess actual survey response over time, and added interventions to improve data quality (eliciting more honest responses) and the impact of heightening confidentiality concerns on response. These interventions made it possible to detect the impact of pledge version on survey attrition and item nonresponse rates to sensitive items that was not possible in the previous studies. The longitudinal design allowed for analysis of actual participation rates in a subsequent survey rather than predicted participation.<\/p>\n<p>One limitation of this research was that each study used different recruitment strategies, samples (both prior BLS respondents and from an online, nonprobability panel), and instructions\/interventions. Thus, the conclusions may not be fully comparable, and should be interpreted as complementary. Further, the results are not likely generalizable to a real world environment. This limitation is common amongst most pretesting research, and would ideally be addressed through a large-scale field test to assess actual respondent reactions and response rates using each pledge version. \u00a0Additionally, this research assessed attrition over only two data collection waves, and the findings may not generalize to longitudinal surveys with more waves. Resource constraints prevented a field test in this study, but is recommended in the future.<\/p>\n<p>In sum, using multiple pretesting modes iteratively allowed for a fuller evaluation of the new BLS confidentiality pledge wording. The addition of online testing to complement traditional methods expanded knowledge of the impact of the pledge language on response. By leveraging the strengths of each method, and carefully designing studies to address limitations, we were able to gain insight into the impact of the pledge wording in a way that would not have been possible had only one study or mode been used.<\/p>\n<p><strong>Table 2. Summary of Design Results, and Conclusions across Studies<\/strong><\/p>\n<h1><a href=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_1.png\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-15056\" src=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_1-1002x1024.png\" alt=\"\" width=\"600\" height=\"613\" srcset=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_1-1002x1024.png 1002w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_1-294x300.png 294w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_1-768x784.png 768w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_1.png 1118w\" sizes=\"auto, (max-width: 600px) 100vw, 600px\" \/><\/a><\/h1>\n<h1><a href=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_2.png\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-15057\" src=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_2-1024x442.png\" alt=\"\" width=\"600\" height=\"259\" srcset=\"https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_2-1024x442.png 1024w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_2-300x130.png 300w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_2-768x332.png 768w, https:\/\/surveyinsights.org\/wp-content\/uploads\/2020\/09\/Study_2.png 1116w\" sizes=\"auto, (max-width: 600px) 100vw, 600px\" \/><\/a><\/h1>\n<h1><strong>Appendix A \u2013 Confidentiality pledge language<\/strong><\/h1>\n<p><strong>Original pledge:<\/strong><\/p>\n<p>The Bureau of Labor Statistics, its employees, agents, and partner statistical agencies, will use the information you provide for statistical purposes only and will hold the information in confidence to the full extent permitted by law. In accordance with the Confidential Information Protection and Statistical Efficiency Act of 2002 (Title 5 of Public Law 107-347) and other applicable Federal laws, your responses will not be disclosed in identifiable form without your informed consent.<\/p>\n<p><strong>Revised pledge:<\/strong><\/p>\n<p>[Original pledge with the following sentence added at the end]: Per the Cybersecurity Enhancement Act of 2015, Federal information systems are protected from malicious activities through cybersecurity screening of transmitted data.<\/p>\n<p><strong>\u00a0<\/strong><\/p>\n<h1><strong>Appendix B &#8211; Comprehension questions from Study 1 (Similar questions used in Studies 2-3)<\/strong><\/h1>\n<p>I\u2019m going to read a list of groups, tell me if they\u2019d be able to see the information you provided.<\/p>\n<ol>\n<li>Bureau of Labor Statistics (BLS)<\/li>\n<li>Department of Labor (DOL)<\/li>\n<li>Internal Review Service (IRS)<\/li>\n<li>Social Security Administration (SSA)<\/li>\n<li>Department of Homeland Security (DHS)<\/li>\n<li>The Census Bureau<\/li>\n<li>All federal government agencies<\/li>\n<li>Congress<\/li>\n<li>Other<\/li>\n<li>Anyone else who would be able to access your data under this language?<\/li>\n<\/ol>\n<p>&nbsp;<\/p>\n<p>I\u2019m going to read a list of activities, tell me if you think they\u2019d be allowed under this language?<\/p>\n<ol>\n<li>It could be combined with other survey answers to create summary statistics<\/li>\n<li>It could be published exactly as you provided it<\/li>\n<li>It could be given to other statistical agencies<\/li>\n<li>It could be given to the IRS<\/li>\n<li>It could be given to other federal government agencies<\/li>\n<li>It could be given to private companies<\/li>\n<li>\u00a0What else could be done under this language?<\/li>\n<\/ol>\n","protected":false},"excerpt":{"rendered":"<p>Introduction Ensuring that survey questions are consistently understood, and that respondents are able and willing to answer them, is of critical importance for reducing measurement error and enhancing data quality. \u00a0Survey researchers have multiple question pretesting methods available (Groves et al., 2009), each with their own strengths and limitations (D\u2019Ardenne &amp; Collins, 2019). Traditionally, pretesting [&hellip;]<\/p>\n","protected":false},"author":1156,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"footnotes":""},"categories":[630],"tags":[720,61,722,512,647,721],"class_list":["post-14659","post","type-post","status-publish","format-standard","hentry","category-advancements-in-online-and-mobile-survey-methods","tag-cognitive-interviewing","tag-item-nonresponse","tag-multi-method-studies","tag-survey-methods","tag-survey-pretesting","tag-web-pretesting"],"acf":[],"_links":{"self":[{"href":"https:\/\/surveyinsights.org\/index.php?rest_route=\/wp\/v2\/posts\/14659","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/surveyinsights.org\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/surveyinsights.org\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/surveyinsights.org\/index.php?rest_route=\/wp\/v2\/users\/1156"}],"replies":[{"embeddable":true,"href":"https:\/\/surveyinsights.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=14659"}],"version-history":[{"count":35,"href":"https:\/\/surveyinsights.org\/index.php?rest_route=\/wp\/v2\/posts\/14659\/revisions"}],"predecessor-version":[{"id":18820,"href":"https:\/\/surveyinsights.org\/index.php?rest_route=\/wp\/v2\/posts\/14659\/revisions\/18820"}],"wp:attachment":[{"href":"https:\/\/surveyinsights.org\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=14659"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/surveyinsights.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=14659"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/surveyinsights.org\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=14659"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}