Skip to main content
Glama
brockwebb

Open Census MCP Server

by brockwebb
0e42e41012d6e0a50c3b937804fc9a65518abd4c.txt156 kB
September 7, 2017 2017 AMERICAN COMMUNITY SURVEY RESEARCH AND EVALUATION REPORT MEMORANDUM SERIES # ACS17-RER-09 MEMORANDUM FOR Victoria Velkoff Chief, American Community Survey Office From: Prepared by: Subject: David Waddington Chief, Social, Economic, and Housing Statistics Division (SEHSD) Jamie Lewis Social, Economic, and Housing Statistics Division (SEHSD) 2016 American Community Survey Content Test Evaluation Report: Computer and Internet Use Attached is the final report for the 2016 American Community Survey (ACS) Content Test for Computer and Internet Use. This report describes the results of the test for the revised versions of the Types of Computer, Internet Access, and Internet Subscription questions. If you have any questions about this report, please contact Kurt Bauman at 301-763-6171 or Jamie Lewis at 301-763-4535. Attachment cc: Agnes Kee (ACSO) Jennifer Ortman (ACSO) Edward Porter (CSRM) Dorothy Barth (DSSD) Patrick Cantwell (DSSD) Asaph Young Chun (DSSD) Elizabeth Poehler (DSSD) Anthony Tersine (DSSD) Kurt Bauman (SEHSD) Nicole Scanniello (SEHSD) Intentionally Blank American Community Survey Research and Evaluation Program September 7, 2017 2016 American Community Survey Content Test Evaluation Report: Computer and Internet Use FINAL REPORT Jamie M. Lewis Social, Economic, and Housing Statistics Division Dorothy A. Barth Decennial Statistical Studies Division Intentionally Blank TABLE OF CONTENTS EXECUTIVE SUMMARY ........................................................................................................... iv 1 BACKGROUND ........................................................................................................................ 1 1.1 Justification for Inclusion of Computer and Internet Use in the Content Test ................... 1 1.2 Question Development........................................................................................................ 2 1.3 Question Content ................................................................................................................ 4 1.4 Research Questions ............................................................................................................. 5 1.4.1 Item Missing Data Rates .......................................................................................... 5 1.4.2 Response Proportions ............................................................................................... 5 1.4.3 Response Error ......................................................................................................... 6 2 METHODOLOGY ..................................................................................................................... 6 2.1 Sample Design .................................................................................................................... 6 2.2 Data Collection ................................................................................................................... 7 2.3 Content Follow-Up ............................................................................................................. 8 2.4 Analysis Metrics ................................................................................................................. 9 2.4.1 Unit Response Rates and Demographic Profile of Responding Households ........... 9 2.4.2 Item Missing Data Rates ........................................................................................ 11 2.4.3 Response Proportions ............................................................................................. 12 2.4.4 Response Error ....................................................................................................... 13 2.4.5 Standard Error Calculations ................................................................................... 16 3 KEY RESEARCH CRITERIA FOR COMPUTER AND INTERNET USE .......................... 17 4 LIMITATIONS ........................................................................................................................ 18 5 RESEARCH QUESTIONS AND RESULTS .......................................................................... 20 5.1 Unit Response Rates and Demographic Profile of Responding Households ................... 20 5.1.1 Unit Response Rates for the Original Content Test Interview ............................... 20 5.1.2 Unit Response Rates for the Content Follow-Up Interview ................................... 22 5.1.3 Demographic and Socioeconomic Profile of Responding Households .................. 22 5.2 Item Missing Data Rates ................................................................................................... 24 5.3 Response Proportions........................................................................................................ 26 5.4 Response Error .................................................................................................................. 31 6 CONCLUSIONS AND RECOMMENDATIONS ................................................................... 35 7 ACKNOWLEDGEMENTS ..................................................................................................... 36 8 REFERENCES ......................................................................................................................... 37 Appendix A. Control and Test Questions in CATI, CAPI, and CFU ........................................... 39 i Appendix B. Unit Response Rates Supplemental Table ............................................................... 41 Appendix C. Benchmarks ............................................................................................................. 42 C.1. Research Questions ........................................................................................................... 42 C.2. Methodology ..................................................................................................................... 42 C.3. Results ............................................................................................................................... 44 List of Tables Table 1. Interview and Reinterview Counts for Each Response Category Used for Calculating the Gross Difference Rate and Index of Inconsistency ................................................ 14 Table 2. Key Research Criteria for Types of Computers Question .............................................. 17 Table 3. Key Research Criteria for Internet Access Question ...................................................... 17 Table 4. Key Research Criteria for Internet Subscription Question ............................................. 18 Table 5. Original Interview Unit Response Rates for Control and Test Treatments, Overall and by Mode .................................................................................................... 21 Table 6. Mail Response Rates by Designated High (HRA) and Low (LRA) Response Areas ............................................................................................................................. 22 Table 7. Content Follow-Up Interview Unit Response Rates for Control and Test Treatments, Overall and by Mode of Original Interview ................................................................. 22 Table 8. Response Distributions: Test versus Control Treatment ................................................ 23 Table 9. Comparison of Average Household Size ........................................................................ 23 Table 10. Comparison of Language of Response ......................................................................... 24 Table 11. Item Missing Data Rates for Control and Test Treatments, Types of Computers Question ..................................................................................................... 24 Table 12. Item Missing Data Rates for Control and Test Treatments, Internet Access Question ........................................................................................................................ 25 Table 13. Proportion of Households with Multiple Responses on Mail Questionnaire, Internet Access Question ........................................................................................................... 25 Table 14. Item Missing Data Rates for Control and Test Treatments, Internet Subscription Type Question ......................................................................................... 26 Table 15. Response Proportions for Control and Test Treatments, Types of Computers Question ........................................................................................................................ 27 Table 16. Response Proportions for Control and Test Treatments, Internet Access Question ........................................................................................................................ 28 Table 17. Proportion of Households with Smartphone or Tablet Reporting Access with a Subscription ............................................................................................................... 29 Table 18. Response Proportions for Control and Test Treatments, Internet Subscription Type Question ........................................................................................................................ 29 Table 19. Proportion of Households with Smartphone or Tablet Reporting Mobile Broadband . 31 Table 20. Gross Difference Rates (GDRs) for Control and Test Treatments, Types of Computers Question ..................................................................................................... 31 Table 21. Indexes of Inconsistency (IOIs) for Control and Test Treatments, Types of Computers Question ..................................................................................................... 32 ii Table 22. Gross Difference Rates (GDRs) for Control and Test Treatments, Internet Access Question ........................................................................................................... 32 Table 23. Indexes of Inconsistency (IOIs) for Control and Test Treatments, Internet Access Question ........................................................................................................... 33 Table 24. Gross Difference Rates (GDRs) for Control and Test Treatments, Internet Subscription Type Question ......................................................................................... 34 Table 25. Indexes of Inconsistency (IOIs) for Control and Test Treatments, Internet Subscription Type Question ......................................................................................... 34 Table B1. Unit Response Rates by Designated High (HRA) and Low (LRA) Response Areas....41 Table C1. Benchmark Estimates, Types of Computer Question……………………………....…44 Table C2. Benchmark Estimates, Internet Access Question………………………………….….45 Table C3. Benchmark Estimates, Internet Subscription Type Question…..……………………..45 List of Figures Figure 1. Control (left) and Test (right) Versions of the Types of Computers Question ............... 4 Figure 2. Control (left) and Test (right) Versions of the Internet Access Question ....................... 4 Figure 3. Control (left) and Test (right) Versions of the Internet Subscription Question .............. 4 Figure A1. CATI/CFU and CAPI Versions of the Control and Test Questions…………………38 iii EXECUTIVE SUMMARY Overview From February to June of 2016, the U.S. Census Bureau conducted the 2016 American Community Survey (ACS) Content Test, a field test of new and revised content. The primary objective was to test whether changes to question wording, response categories, and definitions of underlying constructs improve the quality of data collected. Both new and revised versions of existing questions were tested to determine if they could provide data of sufficient quality compared to a control version as measured by a series of metrics including item missing data rates, response distributions, and response error. The results of this test will be used to help determine the future ACS content and to assess the expected data quality of revised questions and new questions added to the ACS. The 2016 ACS Content Test consisted of a nationally representative sample of 70,000 residential addresses in the United States, independent of the production ACS sample. The sample universe did not include group quarters, nor did it include housing units in Alaska, Hawaii, or Puerto Rico. The test was a split-panel experiment with one-half of the addresses assigned to the control treatment and the other half assigned to the test treatment. As in production ACS, the data collection consisted of three main data collection operations: 1) a six-week mailout period, during which the majority of self-response via internet and mailback were received; 2) a one- month Computer-Assisted Telephone Interview period for nonresponse follow-up; and 3) a one- month Computer-Assisted Personal Interview period for a sample of the remaining nonresponse. For housing units that completed the original Content Test interview, a Content Follow-Up telephone reinterview was conducted to measure response error. Computer and Internet Use This report discusses the topic of Computer and Internet Use, which was first introduced to the ACS in 2013. Because of the rapid change in technology and terminology, it was evident that the questions regarding this topic needed to be revised. Specific concerns included the relatively low percentage of handheld-owning households reporting an internet subscription or a mobile broadband subscription. For the question on computer usage, the number of response categories increased from three to four (with a new category for tablet computers) and the wording of each category was revised for clarity, such as replacing “Handheld computer” with “Smartphone.” The wording was revised for the internet access question to address the rapid change in how people access the internet and the terminology we use to describe internet access, asking about payment (rather than subscription) to a cell phone company in addition to an internet service provider. For the internet subscription question, the number of response options was reduced from seven to five (dropping “DSL,” “Cable modem,” and “Fiber-optic” as separate categories), wording was revised for clarity, and the phrase “Mobile broadband plan” was replaced with “Cellular data plan.” The options were also presented in a different order. Although the test versions of the computer and internet use questions were implemented in 2016 production ACS (Reichert, iv 2015), the topic was included in the Content Test in order to conduct analysis to validate the early implementation decision. Research Questions and Results This research was guided by several research questions concerning missing data rates, differences in the reports of computer usage and internet subscriptions by treatment, and response error. Although not part of the key research, comparisons were also made to benchmark estimates from the Current Population Survey (CPS) Computer and Internet Use Supplement and surveys conducted by the Pew Research Center. Research questions, methodology, and results on benchmarks can be found in Appendix C. Item Missing Data Rates Results indicate that the item missing data rates are not significantly different between treatments for the types of computers question as a whole, as well as for the individual computer categories. For the internet access question, the item missing data rate is significantly lower in the test treatment than in the control treatment, indicating that the test version of the question performed better. For mail responses, any response that indicates more than one type of internet access (a response with more than one box being marked) is considered “missing” data. The rate at which this type of response occurs for the internet access question is not significantly different between treatments, indicating that the changes to the question did not affect this indicator. Finally, for the internet subscription type question overall, the item missing data rates are not significantly different between treatments. Of the five categorical comparisons made, the only item missing data rate that shows a significantly lower value in the test treatment compared with the control treatment is the rate for the “Cellular data plan” category. Response Proportions Findings for the types of computers question reveal that the proportion of “Yes” responses for the “Desktop or laptop” category is lower in the test treatment than in the control treatment. A possible explanation is the introduction of a separate “Tablet” category to the test version of the question. In the absence of this category, some control respondents owning tablets (but not desktops or laptops) may have marked the category for “Desktop, laptop, netbook, or notebook computer.” A larger proportion of test households reported owning or using a smartphone or tablet, compared with the share of control households reporting a handheld computer. A smaller proportion of households in the test treatment indicated that they owned or used some other computer, compared with the control treatment. Regarding the question on internet access, among all households overall and households with a smartphone or tablet, the proportion reporting an internet subscription is higher in the test treatment than in the control treatment. Similarly, reporting of no internet subscription is lower among test households overall. For the final item on internet subscription type, reports of mobile broadband are strikingly higher in the test treatment (about 80 percent) than in the control treatment (about 40 percent), whether v looking at households overall or focusing on households with a smartphone or tablet. The proportion of households reporting a broadband service such as DSL, cable, or fiber-optic is lower in the test treatment than in the control treatment. While the difference is significant, the magnitude is fairly small and the results are close to what we were expecting. The difference likely reflects the number of categories measuring this type of service. Respondents had three categories of this type in the control version of the question, but only a single category in the test version. There is no significant difference in the share of households reporting a dial-up subscription, satellite internet service, or some other service in the test versus control treatments. Response Error The test version of the types of computers question is more reliable than the control version for the categories of smartphone and tablet use and use of some other computer. There were no significant differences between test and control for the other computer categories. For the internet access question, the inconsistency in reports of access with a subscription and access without a subscription is lower in the test treatment than in the control treatment. No other significant differences between treatments were detected for the reliability metrics. For the internet subscription type question, the cellular data plan category in test has greater response reliability than the mobile broadband category in control. The high speed and satellite internet categories in test did not perform as well as control for one of the response reliability metrics. There were no significant differences between control and test for the remaining internet subscription categories. Conclusion Overall, results indicate that data quality improved when using the revised questions. All of the key research criteria for the internet access question were met, and four of five key research criteria were met for both the types of computers and internet subscription type questions. In each case, the key criterion not met was of lowest priority. Item missing data rates in the test treatment were not significantly different from those in the control treatment across the board. Results for the response proportions analysis, in general, were as expected. Particularly noteworthy is the substantial increase in the share of households reporting a cellular data plan in the test treatment versus a mobile broadband plan in the control treatment. Whether looking at all households or specifically at households with a smartphone or tablet (handheld in control), the test proportion is about double the control proportion. Finally, although the reliability of the high speed and satellite internet categories was better for the control version, for the most part the test version of the Computer and Internet Use questions was more reliable or not significantly different from the control version. Altogether, the 2016 ACS Content Test and analyses presented here validate the decision to implement the revised question wording on the 2016 production ACS. The revised question wording will be reflected in the 2016 ACS data release, scheduled to begin in September 2017. vi 1 BACKGROUND From February to June of 2016, the Census Bureau conducted the 2016 American Community Survey (ACS) Content Test, a field test of new and revised content. The primary objective was to test whether changes to question wording, response categories, and definitions of underlying constructs improve the quality of data collected. Both revised versions of existing questions and new questions were tested to determine if they could provide data of sufficient quality compared to a control version as measured by a series of metrics including item missing data rates, response distributions, and response error. The results of this test will be used to help determine the future ACS content and to assess the expected data quality of revised questions and new questions added to the ACS. The 2016 ACS Content Test included the following topics:  Relationship  Race and Hispanic Origin  Telephone Service  Computer and Internet Use  Health Insurance Coverage  Health Insurance Premium and Subsidy (new questions)  Journey to Work: Commute Mode  Journey to Work: Time of Departure for Work  Number of Weeks Worked  Class of Worker   Retirement, Survivor, and Disability Income Industry and Occupation This report discusses the topic questions involving Computer and Internet Use. 1.1 Justification for Inclusion of Computer and Internet Use in the Content Test The questions collecting information on computer availability, internet access, and internet subscriptions were first introduced on the ACS in 2013. Given the rapid rate at which technology is growing and changing, it became apparent that a revision to these questions was already needed. As an example, prior to January 2016, the question about computer ownership did not specifically ask about tablets, but the rate of tablet ownership has grown dramatically in recent years, with recent estimates indicating that 51 percent of adults in the country own a tablet (Pew Research Center, 2016). Preliminary data from 2013 showed that the wording of the questions needed to be revised for a variety of reasons. One finding that raised concerns was the somewhat low percentage of handheld-owning households who reported having an internet subscription. For owners of desktops and laptops, internet subscription was reported to be 91.1 percent. For owners of handheld devices, the internet subscription rate was 76.3 percent. We anticipate that the inclusion of the phrase “cell phone company” will encourage handheld-owning households to think about their data plans as internet subscriptions. 1 Preliminary data also showed that the question wording needed to be revised because of low reports of mobile broadband subscriptions (File & Ryan, 2014). In 2013, for example, only 33 percent of households reported having mobile broadband subscriptions though 64 percent of households reported having a handheld device. Thus, among households with handheld devices and internet subscriptions, only 54 percent reported a mobile broadband subscription. If the estimate of mobile broadband is correct, then half of the households with a handheld device are using them without a data plan. We expect the proportion of households with mobile broadband to increase with the new question wording in the test version. 1.2 Question Development Initial versions of the new and revised questions were proposed by federal agencies participating in the U.S. Office of Management and Budget (OMB) Interagency Committee for the ACS. The initial proposals contained a justification for each change and described previous testing of the question wording, the expected impact of revisions to the time series and the single-year as well as five-year estimates, and the estimated net impact on respondent burden for the proposed revision.1 For proposed new questions, the justification also described the need for the new data, whether federal law or regulation required the data for small areas or small population groups, if other data sources were currently available to provide the information (and why any alternate sources were insufficient), how policy needs or emerging data needs would be addressed through the new question, an explanation of why the data were needed with the geographic precision and frequency provided by the ACS, and whether other testing or production surveys had evaluated the use of the proposed questions. The Census Bureau and the OMB, as well as the Interagency Council on Statistical Policy Subcommittee, reviewed these proposals for the ACS. The OMB determined which proposals moved forward into cognitive testing. After OMB approval of the proposals, topical subcommittees were formed from the OMB Interagency Committee for the ACS, which included all interested federal agencies that use the data from the impacted questions. These subcommittees further refined the specific proposed wording that was cognitively tested. The Census Bureau contracted with Westat to conduct three rounds of cognitive testing. The results of the first two rounds of cognitive testing informed decisions on specific revisions to the proposed content for the stateside Content Test (Stapleton and Steiger, 2015). In the first round, 208 cognitive interviews were conducted in English and Spanish and in two modes (self- administered on paper and interviewer-administered on paper). In the second round of testing, 120 cognitive interviews were conducted for one version of each of the tested questions, in English and Spanish, using the same modes as in the first round. A third round of cognitive testing involved only the Puerto Rico Community Survey (PRCS) and Group Quarters (GQ) versions of the questionnaire (Steiger, Anderson, Folz, Leonard, & Stapleton, 2015). Cognitive interviews in Puerto Rico were conducted in Spanish; GQ cognitive 1 The ACS produces both single and five-year estimates annually. Single year estimates are produced for geographies with populations of 65,000 or more and five-year estimates are produced for all areas down to the block-group level, with no population restriction. 2 interviews were conducted in English. The third round of cognitive testing was carried out to assess the revised versions of the questions in Spanish and identify any issues with questionnaire wording unique to Puerto Rico and GQ populations.2 The proposed changes identified through cognitive testing for each question topic were reviewed by the Census Bureau, the corresponding topical subcommittee, and the Interagency Council on Statistical Policy Subcommittee for the ACS. The OMB then provided final overall approval of the proposed wording for field testing.3 The development of the computer device question came about as a result of a need to keep up with technological updates and changes in computer terminology. The terminology “netbook” and “notebook” computer were excluded from the test version of the question, because they are outdated terms. We know from recent Pew numbers that around 51 percent of adults in the country own tablets (Pew Research Center, 2016), so it was apparent that the word “tablet” needed to appear in a category. Also, new technology, such as smart watches and Google glasses, are worn and not carried by hand, so the word “handheld” became outdated. As a result, a new category “Tablet or other portable wireless computer” was created. “Smartphone” was created as its own category because they have become so widely used and owned. The development of the internet access question came about as a result of changes in technology, terminology, and the way people access the internet. During the cognitive testing phase, concerns were raised about the confusing nature of the word “subscription” (Stapleton and Steiger, 2015). Thus, this term was excluded in the final test version of the question. Also, the words “by paying a cell phone company” were added to help respondents realize that their data plans are equivalent to a paid internet service. This addition, in combination with “Smartphone” having its own category as a computer device, should increase the quality of data collected regarding mobile internet access. Also, during cognitive testing, some respondents answered the question thinking about their habits of internet use at home rather than focusing on their actual ability to access the internet at their house. This will undoubtedly become more of an issue as internet access technology grows increasingly mobile. The phrase “access the internet” was changed to “have access to the internet” to more accurately convey the intent of the question. The question involving types of internet subscriptions was revised in order to address changes in internet use and terminology. The first round of cognitive testing included two categories that used the word broadband: “Mobile broadband” and “Broadband (high speed).” Some respondents answered incorrectly because of their misinterpretation of the terms. Several changes were made to the question to address this problem. The phrase “At this house, apartment, or mobile home” was removed from the question and “installed in this household” was added to the end of the “Broadband (high speed)”, “Satellite”, and “Dial-up” categories. This change should help respondents more easily differentiate between smartphone data plans, which are not tied specifically to a place, and other ways of having access to the internet that are tied to a place. This change also enabled a redesign of the mobile broadband category to put less emphasis on the duplicative use of the term “broadband” by changing “Mobile broadband” to “Cellular data plan.” Three categories were collapsed and used as examples to better describe “Broadband (high speed)” internet service. 2 Note that the field testing of the content was not conducted in Puerto Rico or in GQs. See the Methodology section for more information. 3 A cohabitation question and domestic partnership question were included in cognitive testing but ultimately we decided not to move forward with field testing these questions. 3 1.3 Question Content Control and test versions of each question are shown as they appeared on the mail questionnaire. Automated versions of the questionnaire had the same content formatted accordingly for each mode. Examples of the versions used for Computer Assisted Telephone Interviews (CATI) and Computer Assisted Personal Interviews (CAPI) can be found in Appendix A. The internet instrument is very similar in appearance to the mail version. Figure 1. Control (left) and Test (right) Versions of the Types of Computers Question Figure 2. Control (left) and Test (right) Versions of the Internet Access Question Figure 3. Control (left) and Test (right) Versions of the Internet Subscription Question 4 1.4 Research Questions The following research questions were formulated to guide the analyses of the questions involving Computer and Internet Use. The analyses assess how the test version of the questions perform compared to the control version in the following ways: how often the respondents answered the question, how the responses affect the resulting estimates, and the consistency and accuracy of the responses.4 1.4.1 Item Missing Data Rates 1. Is the item missing data rate for the types of computers question as a whole lower for the test treatment than for the control treatment? 2. Is the item missing data rate for each individual computer type lower for the test treatment than for the control treatment? 3. Is the item missing data rate for the internet access question lower for the test treatment than for the control treatment? 4. In the mail mode, is the proportion of households with multiple responses to the internet access question different between the test and control treatments? 5. Is the item missing data rate for the internet subscription type question as a whole lower for the test treatment than for the control treatment? 6. Is the item missing data rate for each individual subscription type lower for the test treatment than for the control treatment? 1.4.2 Response Proportions 7. Is the proportion of “Yes” responses for the first computer category (Desktop/Laptop) in the test treatment the same as the control treatment proportion? 8. Is the combined proportion of “Yes” responses for the second and third computer categories in test treatment (Smartphone/Tablet) greater than the proportion of “Yes” responses for the control treatment second category (Handheld computer)? 9. Do the changes to the types of computers question decrease the proportion in the “Some other” type of computer category? 10. Is the estimated proportion of households with internet access with a subscription higher in the test treatment than in the control treatment? 11. Is the estimated proportion of households without a subscription (“Access without an internet subscription” combined with “No internet access”) lower in the test treatment than in the control treatment? 12. Among households that reported having a handheld device (“Smartphone” plus “Tablet” categories in test) on the types of computers question, is the proportion of those who also reported having access with a paid internet subscription higher in the test treatment than in the control treatment? 13. Is the proportion of “Dial-up” internet service the same for both treatments? 4 Although not part of the key research, comparisons were also made to benchmark estimates from the Current Population Survey (CPS) Computer and Internet Use Supplement and surveys conducted by the Pew Research Center. Research questions, methodology, and results on benchmarks can be found in Appendix C. 5 14. Is the proportion of “Yes” responses obtained by collapsing the control categories of “DSL,” “Cable,” and “Fiber-optic” the same as the proportion of “Yes” responses for the test treatment category of “Broadband (high speed)?” 15. Is the proportion of “Cellular data” higher in the test treatment than “Mobile broadband plan” is in control? 16. Is the proportion of “Satellite” internet services the same for both treatments? 17. Is the proportion of “Some other service” in the test treatment less than or equal to the proportion in the control treatment? 18. Among households that reported having a smartphone or tablet computer in the types of computers question, is the proportion reporting “Yes” to “Mobile broadband” higher in test than in control? 1.4.3 Response Error 19. Are the measures of response reliability (gross difference rate and index of inconsistency) for each computer type category better for the test treatment than for the control treatment? 20. Are the measures of response reliability (gross difference rate and index of inconsistency) better for the test treatment than for the control treatment for the internet access question? 21. Are the measures of response reliability (gross difference rate and index of inconsistency) for each internet subscription type better for the test treatment than for the control treatment? 2 METHODOLOGY 2.1 Sample Design The 2016 ACS Content Test consisted of a nationally representative sample of 70,000 residential addresses in the United States, independent of the production ACS sample. The Content Test sample universe did not include GQs, nor did it include housing units in Alaska, Hawaii, or Puerto Rico.5 The sample design for the Content Test was largely based on the ACS production sample design with some modifications to better meet the test objectives.6 The modifications included adding an additional level of stratification by stratifying addresses into high and low self-response areas, oversampling addresses from low self-response areas to ensure equal response from both strata, and sampling units as pairs.7 The high and low self-response strata were defined based on ACS self-response rates at the tract level. Sampled pairs were formed by first systematically sampling an address within the defined sampling stratum and then pairing that address with the address listed next in the geographically sorted list. Note that the pair was likely not neighboring addresses. One member of the pair was randomly assigned to receive the 5 Alaska and Hawaii were excluded for cost reasons. GQs and Puerto Rico were excluded because the sample sizes required to produce reliable estimates would be overly large and burdensome, as well as costly. 6 The ACS production sample design is described in Chapter 4 of the ACS Design and Methodology report (U.S. Census Bureau, 2014). 7 Tracts with the highest response rate based on data from the 2013 and 2014 ACS were assigned to the high response stratum in such a way that 75 percent of the housing units in the population (based on 2010 Census estimates) were in the high response areas; all other tracts were designated in the low response strata. Self-response rates were used as a proxy for overall cooperation. Oversampling in low response areas helps to mitigate larger variances due to CAPI subsampling. This stratification at the tract level was successfully used in previous ACS Content Tests, as well as the ACS Voluntary Test in 2003. 6 control version of the question and the other member was assigned to receive the test version of the question, thus resulting in a sample of 35,000 control cases and 35,000 test cases. As in the production ACS, if efforts to obtain a response by mail or telephone were unsuccessful, attempts were made to interview in person a sample of the remaining nonresponding addresses (see Section 2.2 Data Collection for more details). Addresses were sampled at a rate of 1-in-3, with some exceptions that were sampled at a higher rate.8 For the Content Test, the development of workload estimates for CATI and CAPI did not take into account the oversampling of low response areas. This oversampling resulted in a higher than expected workload for CATI and CAPI and therefore required more budget than was allocated. To address this issue, the CAPI sampling rate for the Content Test was adjusted to meet the budget constraint. 2.2 Data Collection The field test occurred in parallel with the data collection activities for the March 2016 ACS production panel, using the same basic data collection protocol as production ACS with a few differences as noted below. The data collection protocol consisted of three main data collection operations: 1) a six-week mailout period, during which the majority of internet and mailback responses were received; 2) a one-month CATI period for nonresponse follow-up; and 3) a one- month CAPI period for a sample of the remaining nonresponse. Internet and mailback responses were accepted until three days after the end of the CAPI month. As indicated earlier, housing units included in the Content Test sample were randomly assigned to a control or test version of the questions. CATI interviewers were not assigned specific cases; rather, they worked the next available case to be called and therefore conducted interviews for both control and test cases. CAPI interviewers were assigned Content Test cases based on their geographic proximity to the cases and therefore could also conduct both control and test cases. The ACS Content Test’s data collection protocol differed from the production ACS in a few significant ways. The Content Test analysis did not include data collected via the Telephone Questionnaire Assistance (TQA) program since those who responded via TQA used the ACS production TQA instrument. The Content Test excluded the telephone Failed Edit Follow-Up (FEFU) operation.9 Furthermore, the Content Test had an additional telephone reinterview operation used to measure response reliability. We refer to this telephone reinterview component as the Content Follow-Up, or CFU. The CFU is described in more detail in Section 2.3. ACS production provides Spanish-language versions of the internet, CATI, and CAPI instruments, and callers to the TQA number can request to respond in Spanish, Russian, Vietnamese, Korean, or Chinese. The Content Test had Spanish-language automated instruments; however, there were no paper versions of the Content Test questionnaires in 8 The ACS production sample design for CAPI follow-up is described in Chapter 4, Section 4.4 of the ACS Design and Methodology report (U.S. Census Bureau, 2014). 9 In ACS production, paper questionnaires with an indication that there are more than five people in the household or questions about the number of people in the household, and self-response returns that are identified as being vacant or a business or lacking minimal data are included in FEFU. FEFU interviewers call these households to obtain any information the respondent did not provide. 7 Spanish.10 Any case in the Content Test sample that completed a Spanish-language internet, CATI, or CAPI response was included in analysis. However, if a case sampled for the Content Test called TQA to complete an interview in Spanish or any other language, the production interview was conducted and the response was excluded from the Content Test analysis. This was due to the low volume of non-English language cases and the operational complexity of translating and implementing several language instruments for the Content Test. CFU interviews for the Content Test were conducted in either Spanish or English. The practical need to limit the language response options for Content Test respondents is a limitation to the research, as some respondents self-selected out of the test. 2.3 Content Follow-Up For housing units that completed the original interview, a CFU telephone reinterview was also conducted to measure response error.11 A comparison of the original interview responses and the CFU reinterview responses was used to answer research questions about response error and response reliability. A CFU reinterview was attempted with every household that completed an original interview for which there was a telephone number. A reinterview was conducted no sooner than two weeks (14 calendar days) after the original interview. Once the case was sent to CFU, it was to be completed within three weeks. This timing balanced two competing interests: (1) conducting the reinterview as soon as possible after the original interview to minimize changes in truth between the two interviews, and (2) not making the two interviews so close together that the respondents were simply recalling their previous answers. Interviewers made two call attempts to interview the household member who originally responded, but if that was not possible, the CFU reinterview was conducted with any other eligible household member (15 years or older). The CFU asked basic demographic questions and a subset of housing and detailed person questions that included all of the topics being tested, with the exception of Telephone Service, and any questions necessary for context and interview flow to set up the questions being tested.12 All CFU questions were asked in the reinterview, regardless of whether or not a particular question was answered in the original interview. Because the CFU interview was conducted via telephone, the wording of the questions in CFU followed the same format as the CATI nonresponse interviews. Housing units assigned to the control version of the questions in the original interview were asked the control version of the questions in CFU; housing units assigned to the test version of the questions in the original interview were asked the test version of the questions in CFU. The only exception was for retirement, survivor, and disability income, for which a different set of questions was asked in CFU.13 10 In the 2014 ACS, respondents requested 1,238 Spanish paper questionnaires, of which 769 were mailed back. From that information, we projected that fewer than 25 Spanish questionnaires would be requested in the Content Test. 11 Throughout this report the “original interview” refers to responses completed via paper questionnaire, internet, CATI, or CAPI. 12 Because the CFU interview was conducted via telephone, the Telephone Service question was not asked. We assume that CFU respondents have telephone service. 13 Refer to the 2016 ACS Content Test report on Retirement Income for a discussion on CFU questions for survivor, disability, and retirement income. 8 2.4 Analysis Metrics This section describes the metrics used to assess the revised versions of the computer and internet use question. The metrics include the item missing data rate, response distributions, response error, and other metrics. This section also describes the methodology used to calculate unit response rates and standard errors for the test. All Content Test data were analyzed without imputation due to our interest in how question changes or differences between versions of new questions affected “raw” responses, not the final edited variables. Some editing of responses was done for analysis purposes, such as collapsing response categories or modes together or calculating a person’s age based on his or her date of birth. All estimates from the ACS Content Test were weighted. Analysis involving data from the original interviews used the final weights that take into account the initial probability of selection (the base weight) and CAPI subsampling. For analysis involving data from the CFU interviews, the final weights were adjusted for CFU nonresponse to create CFU final weights. The significance level for all hypothesis tests is α = 0.1. Since we are conducting numerous comparisons between the control and test treatments, there is a concern about incorrectly rejecting a hypothesis that is actually true (a “false positive” or Type I error). The overall Type I error rate is called the familywise error rate and is the probability of making one or more Type I errors among all hypotheses tested simultaneously. When adjusting for multiple comparisons, the Holm-Bonferroni method was used (Holm, 1979). 2.4.1 Unit Response Rates and Demographic Profile of Responding Households The unit response rate is generally defined as the proportion of sample addresses eligible to respond that provided a complete or sufficient partial response.14 Unit response rates from the original interview are an important measure to look at when considering the analyses in this report that compare responses between the control and test versions of the survey questionnaire. High unit response rates are important in mitigating potential nonresponse bias. For both control and test treatments, we calculated the overall unit response rate (all modes of data collection combined) and unit response rates by mode: internet, mail, CATI, and CAPI. We also calculated the total self-response rate by combining internet and mail modes together. Some Content Test analyses focused on the different data collection modes for topic-specific evaluations, thus we felt it was important to include each mode in the response rates section. In addition to those rates, we calculated the response rates for high and low response areas because analysis for some Content Test topics was done by high and low response areas. Using the Census Bureau’s Planning Database (U.S. Census Bureau, 2016), we defined these areas at the tract level based on the low response score. 14 A response is deemed a “sufficient partial” when the respondent gets to the first question in the detailed person questions section for the first person in the household. 9 The universe for the overall unit response rates consists of all addresses in the initial sample (70,000 addresses) that were eligible to respond to the survey. Some examples of addresses ineligible for the survey were a demolished home, a home under construction, a house or trailer that was relocated, or an address determined to be a permanent business or storage facility. The universe for self-response (internet and mail) rates consists of all mailable addresses that were eligible to respond to the survey. The universe for the CATI response rate consists of all nonrespondents at the end of the mailout month from the initial survey sample that were eligible to respond to the survey and for whom we possessed a telephone number. The universe for the CAPI response rates consists of a subsample of all remaining nonrespondents (after CATI) from the initial sample that were eligible to respond to the survey. Any nonresponding addresses that were sampled out of CAPI were not included in any of the response rate calculations. We also calculated the CFU interview unit response rate overall and by mode of data collection of the original interview and compared the control and test treatments because response error analysis (discussed in Section 2.4.4) relies upon CFU interview data. Statistical differences between CFU response rates for control and test treatments will not be taken as evidence that one version is better than the other. For the CFU response rates, the universe for each mode consists of housing units that responded to the original questionnaire in the given mode (internet, mail, CATI, or CAPI) and were eligible for the CFU interview. We expected the response rates to be similar between treatments; however, we calculated the rates to verify that assumption. Another important measure to look at in comparing experimental treatments is the demographic profile of the responding households in each treatment. The Content Test sample was designed with the intention of having respondents in both control and test treatments exhibit similar distributions of socioeconomic and demographic characteristics. Similar distributions allow us to compare the treatments and conclude that any differences are due to the experimental treatment instead of underlying demographic differences. Thus, we analyzed distributions for data from the following response categories: age, sex, educational attainment, and tenure. The topics of race, Hispanic origin, and relationship are also typically used for demographic analysis; however, those questions were modified as part of the Content Test, so we could not include them in the demographic profile. Additionally, we calculated average household size and the language of response for the original interview.15 For response distributions, we used Rao-Scott chi-square tests of independence to determine statistical differences between control and test treatments (Rao & Scott, 1987). If the distributions were significantly different, we performed additional testing on the differences for each response category. To control for the overall Type I error rate for a set of hypotheses tested simultaneously, we performed multiple-comparison procedures with the Holm-Bonferroni method (Holm, 1979). A family for our response distribution analysis was the set of p-values for the overall characteristic categories (age, sex, educational attainment, and tenure) and the set of p-values for a characteristic’s response categories if the response distributions were found to have statistically significant differences. To determine statistical differences for average household size and the language of response of the original interview we performed two-tailed hypothesis tests. 15 Language of response analysis excludes paper questionnaire returns because there was only an English questionnaire. 10 For all response-related calculations mentioned in this section, addresses that were either sampled out of the CAPI data collection operation or that were deemed ineligible for the survey were not included in any of the universes for calculations. Unmailable addresses were also excluded from the self-response universe. For all unit response rate estimates, differences, and demographic response analysis, we used replicate base weights adjusted for CAPI sampling (but not adjusted for CFU nonresponse). 2.4.2 Item Missing Data Rates Respondents leave items blank for a variety of reasons including not understanding the question (clarity), their unwillingness to answer a question as presented (sensitivity), and their lack of knowledge of the data needed to answer the question. The item missing data rate (for a given item) is the proportion of eligible units, housing units for household-level items or persons for person-level items, for which a required response (based on skip patterns) is missing. We calculated and compared the item missing data rates between control and test for all of the Computer and Internet Use questions. All respondents were required to answer the types of computers and internet access questions. Only those units that responded that they had internet access “with a subscription to an internet service” for control or “by paying a cell phone company or internet service provider” for test were required to answer the question about types of internet subscriptions. Statistical significance of differences between versions was determined using two-tailed t-tests. Types of Computers The percentage of eligible housing units that did not provide a response in the control treatment was compared to the corresponding percentage from the test treatment. In addition to evaluating the overall question, missing data rates for the new test categories were compared individually to the control categories, resulting in three tests of item missing data rates on individual computer types. On mail and internet questionnaires, missing responses were those where no boxes were marked. In CATI and CAPI instruments, a response of either “Don’t Know” or “Refused” was considered missing. Responses where “Some other type of computer” was marked but no write- in was provided were not considered missing. If one type of computer was marked “Yes,” any other type of computer that was left blank was considered to be a “No” instead of a missing answer. Internet Access The percentage of eligible housing units that did not provide a response to this question in the control treatment was compared to the corresponding percentage from the test treatment. On mail and internet questionnaires, missing responses were those where no boxes were marked. In the CATI and CAPI instruments, a response of either “Don’t Know” or “Refused” was considered missing. A limitation of the mail questionnaire version of the internet access question is that a respondent may erroneously mark more than one box as an answer to the question. If more than one box was marked then the answer was considered missing, since we cannot assume which answer is the correct one. We also calculated the number of times a respondent checked multiple boxes for the internet access question. We compared the proportions of responses with multiple marks, using 11 adjusted weights, between control and test. We expected the percentage of multiple responses of the test version to be the same as or lower than the control version. Internet Subscription The percentage of eligible housing units that did not provide a response to this question in the control treatment was compared to the corresponding percentage from the test treatment. As with the computer question, we needed an assessment of overall nonresponse as well as nonresponse for individual components. On mail and internet questionnaires, missing responses were those where no boxes were marked. In CATI and CAPI instruments, a response of either “Don’t Know” or “Refused” was considered missing. Responses where “Some other service” was marked but no write-in was provided were not considered missing. If one type of internet subscription was marked “Yes,” any other type that was left blank was considered to be a “No” instead of a missing answer. 2.4.3 Response Proportions Comparing the proportion of the response categories between the control version of a question and the test version of a question allows us to assess whether the question change affects the resulting estimates. Proportion estimates were calculated as: Types of Computers The control category “Desktop, laptop, netbook, or notebook computer” was compared to the test category “Desktop or laptop” using a two-tailed t-test as the percentages were not expected to differ. The control category for “Handheld computer” was compared to the combined test categories for “Smartphone” and “Tablet or other portable wireless computer.” We used a one- tailed t-test because we expected the test to show a greater percentage of households with smartphone and tablet ownership, due to the updated changes to the categories. A straight comparison was made between control and test in the category of “Some other type of computer.” This analysis only involved the checkbox and did not check for the presence or content of a write-in. We compared the “some other type of computer” category between treatments using a two-tailed t-test. Ideally, we expected to see a lower percentage of “Some other type of computer” responses in test than in control, as the new version added “Tablet” and isolated “Smartphone” to its own category; however, similar proportions were also acceptable. Internet Access Both internet access categories were compared with two-tailed t-tests. Although we expected the test treatment to have a greater percentage of respondents with internet access due to the inclusion of “paying a cell phone company” in the question, we considered an outcome of similar proportions to be acceptable. Also, it was expected that the test version would have a lower percentage of respondents reporting that they had access without a subscription or that they did not have internet access; however we considered it acceptable if the percentages were similar. 12 Category proportion= weighted count of valid responses in categoryweighted count of all valid responses When comparing internet access among households reporting a handheld device, the control universe included all households with a handheld device from the types of computers question while the test universe included all households with either a smartphone or tablet. We compared the percentage of each universe reporting a paid internet subscription (the first box in each version of the internet access question). We compared the proportions using a two-tailed t-test. Internet Subscription The control categories of “Dial-up” and “Satellite” were compared to the corresponding test categories using two-tailed t-tests. Since these categories did not change in the test version, they were expected to have similar percentages of “Yes” responses. The control categories of “DSL,” “Cable,” and “Fiber-optic” were combined and compared to the test category of “Broadband (high speed)” using a two-tailed t-test. This percentage comparison was also expected to be about the same for control and test. The control category of “Mobile broadband plan” was compared to the test category of “Cellular data plan” using a one-tailed t-test. Due to the change in terminology, the test version was expected to result in a higher percentage of mobile broadband subscribers. The category of “Some other service” was compared between control and test using a two-tailed t-test. Because of the clarity of the new categories in the test version, we expected to receive a similar or lower percentage of respondents reporting in the “Some other service” category. When assessing mobile broadband among households reporting a handheld device, similar to the analysis for internet access, we compared control households with a handheld device to test households with either a smartphone or tablet. We compared the proportions using a one-tailed t-test. 2.4.4 Response Error Response error occurs for a variety of reasons, such as flaws in the survey design, misunderstanding of the questions, misreporting by respondents, or interviewer effects. There are two components of response error: response bias and simple response variance. Response bias is the degree to which respondents consistently answer a question incorrectly. Simple response variance is the degree to which respondents answer a question inconsistently. A question has good response reliability if respondents tend to answer the question consistently. Re-asking the same question of the same respondent (or housing unit) allows us to measure response variance. We measured simple response variance by comparing valid responses to the CFU reinterview with valid responses to the corresponding original interview.16 The Census Bureau has frequently used content reinterview surveys to measure simple response variance for large demographic data collection efforts, including the 2010 ACS Content Test, and the 1990, 2000, and 2010 decennial censuses (Dusch & Meier, 2012). 16 A majority of the CFU interviews were conducted with the same respondent as the original interview (see the Limitations section for more information). 13 The following measures were used to evaluate consistency:  Gross difference rate (GDR)  Index of inconsistency (IOI)  L-fold index of inconsistency (IOIL) The first two measures – GDR and IOI – were calculated for individual response categories. The L-fold index of inconsistency was calculated for questions that had three or more mutually exclusive response categories, as a measure of overall reliability for the question. The GDR, and subsequently the simple response variance, are calculated using the following table and formula. Table 1. Interview and Reinterview Counts for Each Response Category Used for Calculating the Gross Difference Rate and Index of Inconsistency Original Interview “Yes” Original Interview “No” Reinterview Totals CFU Reinterview “Yes” CFU Reinterview “No” Original Interview Totals A C a + c b d b + d a + b c + d n Where a, b, c, d, and n are defined as follows: a = weighted count of units in the category of interest for both the original interview and reinterview b = weighted count of units NOT in the category of interest for the original interview, but in the category for the reinterview c = weighted count of units in the category of interest for the original interview, but NOT in the category for the reinterview d = weighted count of units NOT in the category of interest for either the original interview or the reinterview n = total units in the universe = a + b + c + d. The GDR for a specific response category is the percent of inconsistent answers between the original interview and the reinterview (CFU). We calculate the GDR for a response category as Statistical significance between the GDR for a specific response category between the control and test treatments is determined using a two-tailed t-test. In order to define the IOI, we must first discuss the variance of a category proportion estimate. If we are interested in the true proportion of a total population that is in a certain category, we can use the proportion of a survey sample in that category as an estimate. Under certain reasonable assumptions, it can be shown that the total variance of this proportion estimate is the sum of two 14 GDR= (b+c)n × 100 components, sampling variance (SV) and simple response variance (SRV). It can also be shown that an unbiased estimate of SRV is half of the GDR for the category (Flanagan, 1996). SV is the part of total variance resulting from the differences among all the possible samples of size n one might have selected. SRV is the part of total variance resulting from the aggregation of response error across all sample units. If the responses for all sample units were perfectly consistent, then SRV would be zero, and the total variance would be due entirely to SV. As the name suggests, the IOI is a measure of how much of the total variance is due to inconsistency in responses, as measured by SRV and is calculated as: Per the Census Bureau’s general rule, index values of less than 20 percent indicate low inconsistency, 20 to 50 percent indicate moderate inconsistency, and over 50 percent indicate high inconsistency. An IOI is computed for each response category and an overall index of inconsistency, called the L-fold index of inconsistency, is reported for the entire distribution. The L-fold index is a weighted average of the individual indexes computed for each response category. When the sample size is small, the reliability estimates are unstable. Therefore, we do not report the IOI and GDR values for categories with a small sample size, as determined by the following formulas: 2a + b + c < 40 or 2d + b + c < 40, where a, b, c, and d are unweighted counts as shown in Table 1 above (see Flanagan 1996, p. 15). The measures of response error assume that those characteristics in question did not change between the original interview and the CFU interview. To the extent that this assumption is incorrect, we assume that it is incorrect at similar rates between the control and test treatments. An example of this could be a question on ownership of computer devices. For instance, a household that did not report having a tablet originally might have acquired one before the CFU interview and then accurately reported a different response than the original. In calculating the IOI reliability measures, the assumption is that the expected value of the error in the original interview is the same as in the CFU reinterview. This assumption of parallel measures is necessary for the SRV and IOI to be valid. In calculating the IOI measures for this report, we found this assumption was not met for the response categories specified in the limitations section (see Section 4). Biemer (2011, pp. 56-58) provides an example where the assumption of parallel measures is not met, but does not provide definitive guidelines for addressing it. In Biemer’s concluding remarks, he states, “...both estimates of reliability are biased to some extent because of the failure of the parallel assumptions to hold.” 15 IOI= n(b+c) a+c c+d +(a+b)(b+d)×100 Flanagan (2001) addresses this bias problem and offers the following adjustment to the IOI formula: This formula was tested on selected topics in the 2016 ACS Content Test. The IOItestimate resulted in negligible reduction in the IOI values. For this reason, we did not recalculate the IOI values using IOItestimate. Similar to Biemer (2011, p. 58), we acknowledge that for some cases, the estimate of reliability is biased to some extent. For the Computer and Internet Use content, analysis examined the reliability—GDRs and IOIs— of each category of the types of computers, internet access, and internet subscription questions. When analyzing the types of computers question, categories for “Smartphone” and “Tablet” in the test version were combined for comparison with the “Handheld” category in the control version. For the internet subscription item, categories for “DSL”, “Cable”, and “Fiber-optic” in the control version were aggregated for comparison with the “Broadband (high speed)” category in the test version. The specific content of the write-in fields for “Some other computer” and “Some other service” were not assessed for reliability. In addition, the IOIL for the internet access item was determined to estimate overall reliability for the question as a whole. It is not appropriate to calculate the IOIL for the types of computers or internet subscription questions, as the categories for these items are not mutually exclusive. For all Computer and Internet Use items, statistical significance between the GDRs and IOIs of each version were determined using two-tailed t-tests. 2.4.5 Standard Error Calculations We estimated the variances of the estimates using the Successive Differences Replication (SDR) method with replicate weights, the standard method used in the ACS (see U.S. Census Bureau, 2014, Chapter 12). We calculated the variance for each rate and difference using the formula below. The standard error of the estimate (X0) is the square root of the variance: where: 𝑋0 = the estimate calculated using the full sample, 𝑋𝑟 = the estimate calculated for replicate 𝑟. 16 IOItestimate= n2 b+c −n(c−b)2n−1 a+c c+d +(a+b)(b+d)×100 Var(X0)= 480 (Xr80r=1−X0)2 3 KEY RESEARCH CRITERIA FOR COMPUTER AND INTERNET USE Before fielding the 2016 ACS Content Test, we identified which of the metrics would be given higher importance in determining which version of the question yielded the best quality of data for each topic. The following tables identify the research questions and associated metrics and criteria in priority order. Table 2. Key Research Criteria for Types of Computers Question Research Questions Research Criteria In Order of Priority 19 8 1, 2 9 7 The reliability for the test version should be the same as or greater than the control version, especially for smartphone and tablet users as compared to handheld device users. The proportion of responses indicating smartphone or tablet use should be greater in the test treatment than the proportion of responses from control that indicate handheld device use. The item missing data rates for the test treatment should be lower than or the same as the control treatment. The proportion of responses indicating use of some other type of computer for the test treatment should be the same as or lower than the control treatment. Additionally, the proportion of responses indicating desktop or laptop use should be the same between control and test treatments. Table 3. Key Research Criteria for Internet Access Question Research Questions 12 20 3 10, 11 4 Research Criteria In Order of Priority Among households with a smartphone or tablet (handheld in control), the proportion having an internet subscription for the test treatment should be the same as or higher than the control treatment. The reliability for the test version should be the same as or greater than the control version. The item missing data rates for the test version should be lower than or the same as the control version. Among all households, the proportion having an internet subscription for the test treatment should be the same as or higher than the control treatment. Similarly, the proportion without a subscription in the test should be the same as or lower than the control proportion. In the mail mode, the proportion of households with multiple responses to the internet Access question in the test should be the same as or lower than the control proportion. 17 Table 4. Key Research Criteria for Internet Subscription Question Research Questions 18 21 13, 16, 17, 15 5, 6 14 Research Criteria In Order of Priority Among households with a smartphone or tablet (handheld in control), the proportion having a cellular data subscription in the test treatment should be higher than the proportion with mobile broadband in the control treatment. The reliability for the test version should be the same as or greater than the control version, when aggregating categories appropriately. The proportion of “Dial-up” and “Satellite” responses should be the same in the test as in the control. Similarly, the proportion of “Some other service” responses for the test version should be the same as or lower than the control version. Finally, the proportion of “Cellular data” responses in the test should be higher than “Mobile broadband” responses in the control. The item missing data rates for the test treatment should be the same as or lower than the control treatment, when measured as the failure to mark any element of the question. Similarly, the item missing data rates in the test for each individual subscription type should be the same as or lower than the control rates. The proportion of “Yes” responses obtained by collapsing the control categories “DSL,” “Cable,” and “Fiber-optic” should be the same as the proportion of “Yes” responses for the test treatment category of “Broadband (high speed).” 4 LIMITATIONS CATI and CAPI interviewers were assigned control and test treatment cases, as well as production cases. The potential risk of this approach is the introduction of a cross-contamination or carry-over effect due to the same interviewer administering multiple versions of the same question item. Interviewers are trained to read the questions verbatim to minimize this risk, but there still exists the possibility that an interviewer may deviate from the scripted wording of one question version to another. This could potentially mask a treatment effect from the data collected. Interviews were only conducted in English and Spanish. Respondents who needed language assistance in another language were not able to participate in the test. Additionally, the 2016 ACS Content Test was not conducted in Alaska, Hawaii, or Puerto Rico. Any conclusions drawn from this test may not apply to these areas or populations. For statistical analysis specific to the mail mode, there may be bias in the results because of unexplained unit response rate differences between the control and test treatments. We were not able to conduct demographic analysis by relationship status, race, or ethnicity because these topics were tested as part of the Content Test. The CFU reinterview was not conducted in the same mode of data collection for households that responded by internet, by mail, or by CAPI in the original interview since CFU interviews were only administered using a CATI mode of data collection. As a result, the data quality measures derived from the reinterview may include some bias due to the differences in mode of data collection. 18 To be eligible for a CFU reinterview, respondents needed to either provide a telephone number in the original interview or have a telephone number available to the Census Bureau through reverse address look up. As a result, 2,284 of the responding households (11.8 percent with a standard error of 0.2) from the original control interviews and 2,402 of the responding households (12.4 percent with a standard error of 0.2) from the original test interviews were not eligible for the CFU reinterview. The difference between the control and test treatments is statistically significant (p-value=0.06). Although we reinterviewed the same person who responded in the original interview when possible, we interviewed a different member of the household in the CFU for 7.5 percent (standard error of 0.4) of the CFU cases for the control treatment and 8.4 percent (standard error of 0.5) of the CFU cases for the test treatment.17 The difference between the test and control treatments is not statistically significant (p-value=0.26). This means that differences in results between the original interview and the CFU for these cases could be due in part to having different people answering the questions. However, those changes were not statistically significant between the control and test treatments and should not impact the conclusions drawn from the reinterview. The 2016 ACS Content Test does not include the production weighting adjustments for seasonal variations in ACS response patterns, nonresponse bias, and under-coverage bias. As a result, any estimates derived from the Content Test data do not provide the same level of inference as the production ACS and cannot be compared to production estimates. In developing initial workload estimates for CATI and CAPI, we did not take into account the fact that we oversampled low response areas as part of the Content Test sample design. Therefore, workload and budget estimates were too low. In order to stay within budget, the CAPI workload was subsampled more than originally planned. This caused an increase in the variances for the analysis metrics used. An error in addressing and assembling the materials for the 2016 ACS Content Test caused some Content Test cases to be mailed production ACS questionnaires instead of Content Test questionnaires. There were 49 of these cases that returned completed questionnaires, and they were all from the test treatment. These cases were excluded from the analysis. Given the small number of cases affected by this error, there is very little effect on the results. Questionnaire returns were expected to be processed and keyed within two weeks of receipt. Unfortunately, a check-in and keying backlog prevented this requirement from being met, thereby delaying eligible cases from being sent to CFU on a schedule similar to the other modes. Additionally, the control treatment questionnaires were processed more quickly in keying than the test treatment questionnaires resulting in a longer delay for test mail cases to be eligible for CFU. On average, it took 18 days for control cases to become eligible for CFU; it took 20 days for test cases. The difference is statistically significant. This has the potential to impact the response reliability results. 17 This is based on comparing the first name of the respondent between the original interview and the CFU interview. Due to a data issue, we were not able to use the full name to compare. 19 The assumption of parallel measures for the GDR and IOI calculations was not met for the following categories: some other type of computer, access with a subscription, access without a subscription, and mobile broadband internet service. For these categories, the GDR and IOI estimates are biased to some extent. 5 RESEARCH QUESTIONS AND RESULTS This section presents the results from the analyses of the 2016 ACS Content Test data for the Computer and Internet Use questions. An analysis of unit response rates is presented first followed by topic-specific analyses. For the topic-specific analyses, each research question is restated, followed by corresponding data and a brief summary of the results. 5.1 Unit Response Rates and Demographic Profile of Responding Households This section provides results for unit response rates for both control and test treatments for the original Content Test interview and for the CFU interview. It also provides results of a comparison of socioeconomic and demographic characteristics of respondents in both control and test treatments. 5.1.1 Unit Response Rates for the Original Content Test Interview The unit response rate is generally defined as the proportion of sample addresses eligible to respond that provided a complete or sufficient partial response. We did not expect the unit response rates to differ between treatments. This is important because the number of unit responses should also affect the number of item responses we receive for analyses done on specific questions on the survey. Similar item response universe sizes allow us to compare the treatments and conclude that any differences are due to the experimental treatment instead of differences in the populations sampled for each treatment. Table 5 shows the unit response rates for the original interview for each mode of data collection (internet, mail, CATI, and CAPI), all modes combined, and both self-response modes (internet and mail combined) for the control and test treatments. When looking at the overall unit response rate (all modes combined), the difference between control (93.5 percent) and test (93.5 percent) is less than 0.1 percentage points and is not statistically significant. 20 Mode All Modes Self-Response Internet Mail Table 5. Original Interview Unit Response Rates for Control and Test Treatments, Overall and by Mode Test Interviews Test Percent Control Interviews Control Percent Test minus Control P-Value 19,400 93.5 (0.3) 19,455 93.5 (0.3) <0.1 (0.4) 0.98 13,131 8,168 4,963 872 5,397 52.9 (0.5) 34.4 (0.4) 18.4 (0.3) 8.7 (0.4) 83.5 (0.7) 13,284 8,112 5,172 880 5,291 53.7 (0.5) 34.1 (0.4) 19.6 (0.3) 9.2 (0.4) 83.6 (0.6) -0.8 (0.6) 0.4 (0.6) -1.2 (0.5) -0.4 (0.6) <0.1 (0.9) 0.23 0.49 0.01* 0.44 0.96 CATI CAPI Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test at the α=0.1 level. The weighted response rates account for initial sample design as well as CAPI subsampling. When analyzing the unit response rates by mode of data collection, the only modal comparison that shows a statistically significant difference is the mail response rate. The control treatment had a higher mail response (19.6 percent) than the test treatment (18.4 percent) by 1.2 percentage points. As a result of this difference, we looked at how mail responses differed in the high and low response areas. Table 6 shows the mail response rates for both treatments in high and low response areas.18 The difference in mail response rates appears to be driven by the difference of rates in the high response areas. It is possible that the difference in the mail response rates between control and test is related to the content changes made to the test questions. There are some test questions that could be perceived as being too sensitive by some respondents (such as the test question relating to same- sex relationships) and some test questions that could be perceived to be too burdensome by some respondents (such as the new race questions with added race categories). In the automated modes (internet, CATI, and CAPI) there is a higher likelihood of obtaining a sufficient partial response (obtaining enough information to be deemed a response for calculations before the respondent stops answering questions) than in the mail mode. If a respondent is offended by the questionnaire or feels that the questions are too burdensome they may just throw the questionnaire away, and not respond by mail. This could be a possible explanation for the unit response rate being lower for test than control in the mail mode. We note that differences between overall and total self-response response rates were not statistically significant. As most analysis was conducted at this level, we are confident the response rates were sufficient to conduct topic-specific comparisons between the control and test treatments and that there are no underlying response rate concerns that would impact those findings. 18 Table B-1 (including all modes) can be found in Appendix B. 21 Table 6. Mail Response Rates by Designated High (HRA) and Low (LRA) Response Areas Test Interviews 2,082 2,881 - Test Percent 20.0 (0.4) 13.8 (0.3) 6.2 (0.5) Control Interviews 2,224 2,948 - HRA LRA Difference Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test at the α=0.1 level. The weighted response rates account for initial sample design as well as CAPI subsampling. 0.02* 0.43 0.11 Test minus Control -1.5 (0.6) -0.3 (0.4) -1.1 (0.7) Control Percent 21.5 (0.4) 14.1 (0.3) 7.4 (0.4) P-Value 5.1.2 Unit Response Rates for the Content Follow-Up Interview Table 7 shows the unit response rates for the CFU interview by mode of data collection of the original interview and for all modes combined, for control and test treatments. Overall, the differences in CFU response rates between the treatments are not statistically significant. The rate at which CAPI respondents from the original interview responded to the CFU interview is lower for test (34.8 percent) than for control (37.7 percent) by 2.9 percentage points. While the protocols for conducting CAPI and CFU were the same between the test and control treatments, we could not account for personal interactions that occur in these modes between the respondent and interviewer. This can influence response rates. We do not believe that the difference suggests any underlying CFU response issues that would negatively affect topic-specific response reliability analysis for comparing the two treatments. Table 7. Content Follow-Up Interview Unit Response Rates for Control and Test Treatments, Overall and by Mode of Original Interview Test Percent Test Interviews Control Interviews Control Percent Original Interview Mode Test minus Control P-Value All Modes 7,867 44.8 (0.5) Internet Mail CATI CAPI 4,078 2,202 369 1,218 51.9 (0.6) 46.4 (0.9) 48.9 (1.9) 34.8 (1.2) 7,903 4,045 2,197 399 1,262 45.7 (0.6) -0.8 (0.8) 52.5 (0.7) 44.2 (0.9) 51.5 (2.5) 37.7 (1.1) -0.6 (0.8) 2.1 (1.3) -2.5 (2.9) -2.9 (1.6) 0.30 0.49 0.11 0.39 0.07* Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test at the α=0.1 level. 5.1.3 Demographic and Socioeconomic Profile of Responding Households One of the underlying assumptions of our analyses in this report is that the sample for the Content Test was selected in such a way that responses from both treatments would be comparable. We did not expect the demographics of the responding households for control and test treatments to differ. To test this assumption, we calculated distributions for respondent data for the following response categories: age, sex, educational attainment, and tenure.19 The 19 We were not able to conduct demographic analysis by relationship status, race, or ethnicity because these topics were tested as part of the Content Test. 22 response distribution calculations can be found in Table 8. Items with missing data were not included in the calculations. After adjusting for multiple comparisons, none of the differences in the categorical response distributions shown below is statistically significant. Table 8. Response Distributions: Test versus Control Treatment Test Percent (n=43,236) 5.7 (0.2) 17.8 (0.3) 8.6 (0.3) 25.1 (0.3) 26.8 (0.4) 16.0 (0.3) (n=43,374) 48.8 (0.3) 51.2 (0.3) (n=27,482) 1.3 (0.1) 8.1 (0.3) 1.7 (0.1) 21.7 (0.4) 3.5 (0.2) 21.0 (0.4) 8.8 (0.3) 20.9 (0.4) 13.1 (0.3) (n=17,190) 43.1 (0.6) 21.1 (0.4) 33.8 (0.6) 1.9 (0.2) Item AGE Under 5 years old 5 to 17 years old 18 to 24 years old 25 to 44 years old 45 to 64 years old 65 years old or older SEX Male Female EDUCATIONAL ATTAINMENT# No schooling completed Nursery to 11th grade 12th grade (no diploma) High school diploma GED† or alternative credential Some college Associate’s degree Bachelor’s degree Advanced degree TENURE Owned with a mortgage Owned free and clear Rented Occupied without payment of rent Control Percent (n=43,325) 6.1 (0.2) 17.6 (0.3) 8.1 (0.3) 26.2 (0.3) 26.6 (0.4) 15.4 (0.3) (n=43,456) 49.1 (0.3) 50.9 (0.3) (n=27,801) 1.2 (0.1) 8.0 (0.3) 1.6 (0.1) 22.3 (0.4) 3.6 (0.2) 20.2 (0.4) 9.1 (0.3) 20.3 (0.4) 13.7 (0.3) (n=17,236) 43.2 (0.5) 21.2 (0.4) 34.0 (0.5) 1.7 (0.1) Source: U.S. Census Bureau, 2016 American Community Survey Content Test #For ages 25 and older †General Educational Development Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. Significance testing done at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. Adjusted P-Value 0.34 - - - - - - 1.00 - - 1.00 - - - - - - - - - 1.00 - - - - We also analyzed two other demographic characteristics shown by the responses from the survey: average household size and language of response. The results for the remaining demographic analyses can be found in Table 9 and Table 10. Table 9. Comparison of Average Household Size Test (n=17,608) Control (n=17,694) Test minus Control Topic Average Household Size (Number of People) Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Significance was tested based on a two-tailed t-test at the α=0.1 level. >-0.01 (<0.1) 2.52 (<0.1) 2.51 (<0.1) 0.76 P-value 23 Table 10. Comparison of Language of Response Control Percent (n=17,694) Language of Response 96.2 (0.2) English 2.6 (0.2) Spanish 1.2 (0.1) Undetermined Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Significance was tested based on a two-tailed t-test at the α=0.1 level. Test Percent (n=17,608) 96.1 (0.2) 2.7 (0.2) 1.2 (0.1) Test minus Control <0.1 (0.3) <0.1 (0.2) <0.1 (0.2) 0.52 0.39 0.62 P-value The Content Test was available in two languages, English and Spanish, for all modes except the mail mode. However, the language of response variable was missing for some responses, so we created a category called undetermined to account for those cases. There are no detectable differences between control and test for average household size or language of response. There are also no detectable differences for any of the response distributions that we calculated. As a result of these analyses, it appears that respondents in both treatments do exhibit comparable demographic characteristics since none of the resulting findings is significant, which verifies our assumption of demographic similarity between treatments. 5.2 Item Missing Data Rates Is the item missing data rate for the types of computers question as a whole lower for the test treatment than for the control treatment? The first row of Table 11 shows the item missing data rates for the types of computers question as a whole. There are no significant differences in the items missing data rates between treatments for any of the computer type categories. This suggests that the changes made to the question do not affect item nonresponse. Table 11. Item Missing Data Rates for Control and Test Treatments, Types of Computers Question Item Entire question Test Percent (n=17,588) 1.3 (0.1) Control Percent (n=17,688) 1.4 (0.1) Test minus Control -0.1 (0.2) Adjusted P-Value 1.00 Desktop or laptop Smartphone or tablet vs. Handheld Other computer Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. Significance was tested based on a two-tailed t-test (test≠control) at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. <0.1 (0.2) -0.2 (0.2) -0.3 (0.2) 1.5 (0.1) 1.7 (0.1) 2.0 (0.1) 1.5 (0.1) 1.5 (0.1) 1.7 (0.1) 1.00 0.91 0.23 24 Is the item missing data rate for each individual computer type lower for the test treatment than for the control treatment? Item missing data rates for each computer type category are displayed above in Table 11. Similar to what was observed for the question overall, the item missing data rate for individual categories is not significantly different for test versus control, indicating that the test version does not reduce or increase item nonresponse. Is the item missing data rate for the internet access question lower for the test treatment than for the control treatment? Table 12 contains information on item missing data rates for the internet access question. Item missingness is significantly lower in the test treatment (2.0 percent) than in the control treatment (2.3 percent), indicating that the test version of the question performed better in terms of item missingness. Omitting the confusing term “subscription” from the test version of the question likely made it easier for some respondents to answer. Table 12. Item Missing Data Rates for Control and Test Treatments, Internet Access Question Test Percent (n=17,588) Item Control Percent (n=17,688) Test minus Control P-Value Entire question Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test (test≠control) at the α=0.1 level. -0.3 (0.2) 2.3 (0.2) 2.0 (0.2) 0.07* In the mail mode, is the proportion of households with multiple responses to the internet access question different between the test and control treatments? The share of households providing multiple responses to the internet access question in the mail mode is found in Table 13. We include the results of multiple responses in this section, as the internet access item is considered missing for cases marking more than one box. There is no significant difference between treatments in the proportion of households with multiple responses, indicating that the changes to the question did not affect this indicator. Table 13. Proportion of Households with Multiple Responses on Mail Questionnaire, Internet Access Question Control Percent (n=5,062) Test minus Control Test Percent (n=4,859) P-Value 0.5 (0.1) 0.5 (0.1) 0.1 (0.2) 0.75 Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Significance was tested based on a two-tailed t-test at the α=0.1 level. 25 Is the item missing data rate for the internet subscription type question as a whole lower for the test treatment than for the control treatment? The first row of Table 14 displays item missing data rates for the internet subscription type question as a whole. Note that whereas the universe for the types of computers and internet access questions is all eligible housing units, the universe for the internet subscription type question is all eligible housing units that have internet access with a subscription. The item missing data rate is not significantly different between the control and test treatments. Table 14. Item Missing Data Rates for Control and Test Treatments, Internet Subscription Type Question Item Entire question Test Percent (n=14,033) 2.3 (0.2) Control Percent (n=13,624) 1.8 (0.2) Test minus Control 0.5 (0.3) Adjusted P-Value 0.25 Dial-up High speed vs. DSL/Cable/Fiber-optic Cellular data plan vs. Mobile broadband Satellite Other service Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test (test ≠ control) at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. -0.2 (0.1) <0.0 (0.1) -0.4 (0.1) -0.3 (0.1) -0.3 (0.1) 1.1 (0.1) 0.9 (0.1) 1.2 (0.1) 1.2 (0.1) 1.3 (0.1) 1.0 (0.1) 1.0 (0.1) 0.8 (0.1) 1.0 (0.1) 1.0 (0.1) 0.29 0.78 0.01* 0.25 0.25 Is the item missing data rate for each individual subscription type lower for the test treatment than for the control treatment? Finally, information on missingness for individual subscription types is found above in Table 14. Of the five categorical comparisons made, the only significant difference detected in the item missing data rates is the rate for the “Cellular data plan” category. Missingness is lower in the test treatment (0.8 percent) than in the control treatment (1.2 percent), suggesting that respondents understand the phrase “Cellular data plan” better than the phrase “Mobile broadband plan.” 5.3 Response Proportions For all Computer and Internet Use questions, the universe for the response proportion analysis is households with a nonmissing response to the item of concern. Is the proportion of "Yes" responses for the first computer category (Desktop/Laptop) in the test treatment the same as the control treatment proportion? Table 15 displays the response proportions for each category of the types of computers question. Although we expected the same share of households to report owning or using a desktop or laptop in each treatment, results indicate that this proportion is lower in the test treatment (78.6 percent) than in the control treatment (80.7 percent). A possible explanation for the observed 26 difference is the introduction of the “Tablet” category to the test version of the question. In the absence of this category, some control respondents owning or using tablets (but not desktops or laptops) may have marked the category for “Desktop, laptop, netbook, or notebook computer.” Table 15. Response Proportions for Control and Test Treatments, Types of Computers Question Control Percent Category (n=17,387) Desktop or laptop 80.7 (0.4) Smartphone or tablet vs. Handheld 79.8 (0.4) 7.9 (0.3) Other computer Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference between the two rates at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. Question allows for multiple categories to be marked, so columns will not sum to 100 percent. Test Percent (n=17,329) 78.6 (0.4) 82.4 (0.4) 4.6 (0.2) Test minus Control -2.1 (0.6) 2.6 (0.6) -3.3 (0.4) Alternative Hypothesis T≠C T>C T≠C <0.01* <0.01* <0.01* Adjusted P-Value Is the combined proportion of “Yes” responses for the second and third computer categories in test treatment (Smartphone/Tablet) greater than the proportion of “Yes” responses for the control treatment second category (Handheld computer)? Looking once more at Table 15 above, as expected, results reveal that a larger proportion of test households reported owning or using a smartphone or tablet (82.4 percent), compared with the share of control households reporting a handheld computer (79.8 percent). Under the old (control) question wording, some smartphone and/or tablet owners may not have recognized the category for “Handheld computer, smart mobile phone, or other handheld wireless computer” as applying to them. Specific categories for “Smartphone” and “Tablet or other portable wireless computer” found in the new (test) wording likely are better understood by those with these devices. Do the changes to the types of computers question decrease the proportion in the “Some other” category? The final row of Table 15 above shows the share of households reporting some other computer in the test and control treatments. As predicted, a smaller proportion of households in the test treatment indicated that they owned or used some other computer (4.6 percent), compared with the control treatment (7.9 percent). This change is also likely due to replacing the “Handheld computer, smart mobile phone, or other handheld wireless computer” category with specific options for “Smartphone” and “Tablet or other portable wireless computer.” Under the old (control) wording, some smartphone and/or tablet users may have marked the “other” category, but the new (test) wording makes it easier to find the relevant descriptive category(ies). 27 Is the estimated proportion of households with internet access with a subscription higher in the test treatment than in the control treatment? Table 16 contains response proportions for the internet access question. As expected, results show that the proportion of households reporting internet access with a subscription is higher in the test treatment, at 83.8 percent, than in the control treatment, at 82.3 percent. As suggested by earlier cognitive testing, respondents likely find the term “paying,” used in the test version of the question, clearer than the term “subscription,” used in the control version. Also important, adding the phrase “cell phone company” likely resonated with respondents who receive internet through a cell phone provider instead of or in addition to a conventional internet service provider. Table 16. Response Proportions for Control and Test Treatments, Internet Access Question Control Percent (n=17,188) 82.3 (0.4) 17.7 (0.4) Category Access with subscription No subscription Total Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test (test ≠ control) at the α=0.1 level. Test minus Control 1.5 (0.6) -1.5 (0.6) N/A Test Percent (n=17,171) 83.8 (0.4) 16.2 (0.4) 0.01* 0.01* N/A P-Value 100.0 100.0 Is the estimated proportion of households without a subscription (“Access without an internet subscription” combined with “No internet access”) lower in the test treatment than in the control treatment? We see (in Table 16) that at the same time that reporting of internet subscriptions was higher for test households, reporting of no internet subscription was lower among test households. Only 16.2 percent of households in the test treatment indicated having internet access without a subscription or no internet access, compared with 17.7 percent of households in the control treatment. Among households that reported having a handheld device (“Smartphone” plus “Tablet” categories in test) on the types of computers question, is the proportion of those who also reported having access with a paid internet subscription higher in the test treatment than in the control treatment? shows the share of households reporting access with a subscription, looking specifically at households owning a device such as a smartphone or tablet. As was observed for all households overall, those with a smartphone or tablet are more likely to report a subscription when receiving the test version of the question (92.4 percent) than when seeing the control version (90.5 percent). Thus, the revised question wording better captures internet access among portable device owners as well as for the general population. 28 Table 17. Proportion of Households with Smartphone or Tablet Reporting Access with a Subscription Test Percent (n=13,976) 92.4 (0.4) Control Percent (n=13,437) 90.5 (0.4) Test minus Control P-Value 1.9 (0.5) <0.01* Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test (test ≠ control) at the α=0.1 level. Is the proportion of “Dial-up” internet service the same for test and control treatments? Response proportions for the various types of internet subscriptions can be found in Table 18. Once again, please note that the universe for the subscriptions question is households that access the internet with a subscription. Starting with dial-up, we see that there is no significant difference in the share of households reporting this type of subscription in the test versus control treatments. This is as expected, given similar wording for this category in the two versions of the subscription type question. Table 18. Response Proportions for Control and Test Treatments, Internet Subscription Type Question Category Dial-up Test Percent (n=14,037) 2.3 (0.2) Control Percent (n=13,476) 2.7 (0.2) Test minus Control -0.4 (0.2) High speed vs. DSL/Cable/Fiber-optic 81.4 (0.5) 85.0 (0.5) -3.6 (0.6) Cellular data plan vs. Mobile broadband 79.9 (0.4) 39.7 (0.6) 40.2 (0.8) Satellite 6.5 (0.3) 6.0 (0.3) 0.5 (0.4) Alternative Hypothesis Adjusted P-Value T≠C T≠C T>C T≠C 0.23 <0.01* <0.01* 0.44 0.61 Other service Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference between the two rates at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. Question allows for multiple categories to be marked, so columns will not sum to 100 percent. 0.1 (0.2) 1.7 (0.2) 1.6 (0.1) T≠C Is the proportion of “Yes” responses obtained by collapsing the control categories of “DSL,” “Cable,” and “Fiber-optic” the same as the proportion of “Yes” responses for the test treatment category of “Broadband (high speed)?” The second row of Table 18 above displays the proportion of households reporting a broadband service such as DSL, cable, or fiber-optic. The share of households reporting this type of internet service is lower in the test treatment, at 81.4 percent, than in the control treatment, at 85.0 percent. While the difference is significant, the results are close to what we were expecting. This difference likely reflects the number of categories measuring this type of service. Respondents had three categories of this type in the control version of the question, but a single category in the test version. We are unable to determine whether the difference indicates overreporting for 29 the control version or underreporting for the test version. However, this does indicate an unintended consequence of streamlining the question. Is the proportion of “Cellular data” higher in the test treatment than “Mobile broadband plan” is in the control? Looking once more at Table 18 above, we see a striking result for the share of households reporting cellular or mobile internet service. Reports of this type of service are about twice as high in the test treatment, at 79.9 percent, compared with the control treatment, at 39.7 percent. This finding suggests that respondents understand the phrase “Cellular data plan” more clearly than “Mobile broadband plan.” The movement of the category to the first position under the question stem in the test treatment may also have made the choice more visible to respondents. Is the proportion of “Satellite” internet service the same for test and control treatments? Results for satellite service in Table 18 indicate that there is no significant difference in the proportion of households reporting satellite internet service in the test versus control treatments. These results were expected as there was no change to the wording for satellite internet service category. Is the proportion of “Some other service” in the test treatment less than or equal to the proportion in the control treatment? The final row of Table 18 above contains results for the share of households reporting “some other service”. There is no significant difference between the test treatment and the control treatment. This was expected due to the fact that there was no difference in question wording between control and test for this response category. Among households that reported having a smartphone or tablet computer in the computers question, is the proportion reporting “Yes” to “Mobile broadband” higher in test than in control? Finally, Table 19 displays the proportion of households reporting mobile broadband, focusing on households owning a device such as a smartphone or tablet. Similar to what was seen for all households, the share reporting mobile broadband is strikingly higher for households in the test treatment (85.4 percent) than those in the control treatment (43.3 percent). This result indicates that the new question wording improves measurement of mobile broadband not only for households overall, but also for those owning or using handheld devices. 30 Table 19. Proportion of Households with Smartphone or Tablet Reporting Mobile Broadband Control Percent (n=11,818) 43.3 (0.7) Test Percent (n=12,758) 85.4 (0.4) Alternative Hypothesis T>C Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. P-values with an asterisk (*) indicate a significant difference between the two rates at the α=0.1 level. Minor additive discrepancies are due to rounding. Test minus Control 42.2 (0.8) P-Value <0.01* 5.4 Response Error Are the measures of response reliability (GDR and IOI) for each computer type category better for the test treatment than for the control treatment? Table 20 displays the Gross Difference Rates (GDRs) from the control and test treatments for each category of the types of computers question. The reliability of responses on ownership of a desktop or laptop is not significantly different between treatments. However, as expected, the test treatment shows greater reliability regarding both smartphone or tablet use and use of some other computer. Seven percent of answers on smartphone or tablet use are inconsistent between the original interview and CFU for the test treatment, whereas 10.8 percent of answers on handheld use are inconsistent in the control treatment. Inconsistency in reports of owning some other computer is lower in the test treatment, at 11.2 percent, than in the control treatment, at 19.0 percent. Greater reliability for the test treatment is likely due to the addition of a category clarifying how tablets should be classified, as well as a category allowing respondents to report ownership of a smartphone specifically. Under the old (control) wording, some respondents may have reported their smartphone or tablet using the “Handheld” category in one interview, and under the “other computer” category in the other interview. Table 20. Gross Difference Rates (GDRs) for Control and Test Treatments, Types of Computers Question Category Desktop or laptop Smartphone or tablet vs. Handheld Other computer Test Sample Size Test GDR Percent Control Sample Size Control GDR Percent Test minus Control Adjusted P-Value 7,766 6.0 (0.4) 7,799 6.0 (0.4) <0.1 (0.6) 0.94 7,746 7.0 (0.4) 7,771 7,728 11.2 (0.5) 7,748 10.8 (0.5) 19.0 (0.6) -3.8 (0.7) <0.01* -7.9 (0.8) <0.01* Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test (test ≠ control) at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. Indexes of Inconsistency (IOIs) for each category of the types of computers item are found in Table 21. Similar to results from GDRs, the IOIs indicate greater reliability in the reporting of smartphones or tablets for the test treatment. The IOI estimate for the test treatment (23.8 percent) is significantly lower than that for the control treatment (32.9 percent). Inconsistency in reports of using a desktop or laptop is not significantly different in test versus control, exhibiting 31 a low value in each treatment. Nor is inconsistency in answers for the “other computer” category significantly different in the test treatment compared with the control treatment. Values are high across treatments. Once more, results suggest that the new “Tablet” and specific “Smartphone” categories increase reliability. High inconsistency in reports of other computers is likely due to the inherent vagueness of “other” response options. Table 21. Indexes of Inconsistency (IOIs) for Control and Test Treatments, Types of Computers Question Category Desktop or laptop Test Sample Size Test IOI Percent Control Sample Size Control IOI Percent Test minus Control Adjusted P-Value 7,766 18.3 (1.2) 7,799 20.5 (1.5) -2.1 (1.8) 0.46 Smartphone or tablet vs. Handheld Other computer Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test (test ≠ control) at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. 7,771 32.9 (1.5) 7,748 89.7 (2.3) 7,746 23.8 (1.3) 7,728 88.0 (2.4) -9.0 (2.2) -1.6 (3.3) <0.01* 0.63 Are the measures of response reliability (GDR and IOI) better for the test treatment than for the control treatment for the internet access question? GDRs for the control and test versions of the internet access question are presented in Table 22. Inconsistency in reports of access with a subscription is lower in the test treatment (9.4 percent) than in the control treatment (11.3 percent). Reliability for access without a subscription also improves under the revised question wording. About 4.2 percent of responses on access without a subscription are inconsistent between the original interview and reinterview for the test treatment, compared with 9.1 percent of responses in the control treatment. Reliability for the “No internet access” category is not significantly different between test and control. Respondents likely interpret the term “paying,” used in the test version of the question, in a more consistent way than the term “subscription,” used in the control version. Also, adding the phrase “cell phone company” likely increases reliability for respondents who receive internet through a cell phone service instead of or in addition to a conventional internet service provider. Table 22. Gross Difference Rates (GDRs) for Control and Test Treatments, Internet Access Question Category Test GDR Percent (n=7,669) Control GDR Percent (n=7,641) Test minus Control Adjusted P-Value Access with subscription Access without subscription No internet access Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based a two-tailed t-test (test ≠ control) at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. 11.3 (0.5) 9.1 (0.5) 5.7 (0.4) -2.0 (0.7) -4.9 (0.6) 0.9 (0.6) 9.4 (0.5) 4.2 (0.4) 6.6 (0.4) <0.01* <0.01* 0.13 32 Table 23 contains the IOIs for each category of the internet access question, as well as the L-fold index of inconsistency (IOIL) capturing reliability for the overall question. Starting with the IOIL, we see that the reliability of estimates of internet access is not significantly different between test and control treatments. As a whole, the internet access question demonstrates moderate levels of inconsistency. Similarly, the IOIs for the individual access categories are not significantly different in the test treatment. Levels of inconsistency are moderate for the “access with a subscription” and “no internet access” categories, and high for the “access without a subscription” response option. The high inconsistency of the “access without a subscription” category likely relates to its status as a residual category. Because the legislation governing this topic in ACS specifies that internet subscriptions be measured, this response option is needed to make the internet access question exhaustive. However, respondents may interpret this category differently at various points in time. For example, respondents whose apartment building provides internet service could initially say that they have access without a subscription, since they do not directly subscribe. But at a later point, they could report access with a subscription, thinking that they do pay for internet through higher rent. Table 23. Indexes of Inconsistency (IOIs) for Control and Test Treatments, Internet Access Question Category Test IOI Percent (n=7,669) Control IOI Percent (n=7,641) Test minus Control Adjusted P-Value Entire question (IOIL) 34.9 (1.8) 39.5 (1.5) -4.7 (2.5) 0.23 Access with subscription 33.4 (1.8) Access without subscription 71.8 (5.2) 27.6 (1.8) No internet access Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. Significance was tested based on a two- tailed t-test (test ≠ control) at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. 36.1 (1.6) 78.9 (3.2) 24.6 (1.5) -2.8 (2.6) -7.1 (6.0) 3.0 (2.5) 0.69 0.69 0.69 Are the measures of response reliability (GDR and IOI) for each internet subscription type better for the test treatment than for the control treatment? Turning to the internet subscription type question, the GDRs for the control and test treatments are found in Table 24. As a final reminder, the universe for the subscription question is households that access the internet with a subscription. For “dial-up” and “other service” subscription types there are no significant differences between test and control. The GDR for the “Satellite” category is higher in the test than in the control treatment. In contrast, inconsistency between the original interview and reinterview is lower for the test item on cellular data plans (17.4 percent) than for the control item on mobile broadband (38.1 percent). Thus, respondents interpret the phrase “Cellular data plan” more consistently than the phrase “Mobile broadband.” Inconsistency is higher for the test version of high speed internet versus the control version of the combined categories of DSL, Cable, and Fiberoptic. The need to combine categories to make a straight comparison between treatments may have contributed to the lower gross difference rate for control as the probability of consistency is higher for three combined categories than for one category on its own. 33 Table 24. Gross Difference Rates (GDRs) for Control and Test Treatments, Internet Subscription Type Question Category Dial-up High speed vs. DSL/Cable/Fiber-optic Cellular data plan vs. Mobile broadband Test Sample Size Test GDR Percent Control Sample Size Control GDR Percent Test minus Control Adjusted P-Value 5,950 4.6 (0.4) 5,527 3.9 (0.4) 0.8 (0.6) 0.40 5,927 13.0 (0.8) 5,531 9.9 (0.5) 3.1 (0.9) <0.01* 5,965 17.4 (0.8) 5,494 38.1 (1.0) -20.8 (1.3) <0.01* Satellite 5,954 9.4 (0.7) 5,523 6.5 (0.4) 2.9 (0.8) <0.01* Other service Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test (test ≠ control) at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. 0.7 (0.6) 4.9 (0.5) 4.2 (0.5) 5,945 5,511 0.40 Finally, Table 25 contains the IOIs for each internet subscription type. These results indicate that reliability for dial-up, broadband (high speed), satellite, or other service is not significantly different when comparing test versus control. Again, we find evidence of greater reliability for estimates of cellular data plans from the test treatment compared with estimates of mobile broadband from the control treatment. The IOI test estimate, at 52.3 percent, is significantly lower than the IOI control estimate, at 76.5 percent. In general, levels of inconsistency for the various subscription types are high, with index values over 50 percent. Once more, these findings suggest that respondents more reliably understand the phrase “Cellular data plan” than the phrase “Mobile broadband.” Table 25. Indexes of Inconsistency (IOIs) for Control and Test Treatments, Internet Subscription Type Question Test Sample Size Category Test IOI Percent Control Sample Size Control IOI Percent Test minus Control Adjusted P-Value Dial-up High speed vs. DSL/Cable/Fiber-optic Cellular data plan vs. Mobile broadband Satellite 5,950 85.5 (3.6) 5,527 85.4 (4.8) 0.1 (5.9) 5,927 53.2 (2.8) 5,531 53.2 (2.3) -0.1 (3.3) 1.00 1.00 5,965 52.3 (2.1) 5,494 76.5 (2.1) -24.2 (3.1) <0.01* 5,954 65.6 (3.4) 5,523 58.2 (2.9) 7.4 (3.9) 0.23 Other service Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are shown in parentheses. Minor additive discrepancies are due to rounding. P-values marked with an asterisk (*) indicate a significant difference based on a two-tailed t-test (test ≠ control) at the α=0.1 level. P-values have been adjusted for multiple comparisons using the Holm-Bonferroni method. 96.7 (2.0) 93.5 (4.3) -3.2 (4.5) 5,511 5,945 1.00 34 6 CONCLUSIONS AND RECOMMENDATIONS Questions on Computer and Internet Use were first introduced to the ACS in 2013. Considering the brisk rate at which technology develops and changes, question revisions were already needed. Specific concerns included the relatively low percentage of handheld-owning households reporting an internet subscription, as well as low reports of mobile broadband subscriptions (File & Ryan, 2014). The 2016 ACS Content tested several changes to the Computer and Internet Use questions. The primary change to the types of computers question involved the replacement of the “Handheld computer” category with a specific “Smartphone” category and a new category for “Tablet or other portable wireless computer.” For the internet access question, the main changes involved replacing the term “subscription” with “paying,” and asking about payment to a cell phone company in addition to an internet service provider. Substantial changes to the subscription type question involved replacing the phrase “Mobile broadband plan” with “Cellular data plan,” and moving this category to the top position. In addition, the individual categories for “DSL,” “Cable,” and “Fiber-optic” were combined into a single “Broadband (high speed)” category. Overall, results indicate that data quality improved when using the revised questions. All of the key research criteria for the internet access question were met, and four of five key criteria were met for both the types of computers and internet subscription type questions. In each case, the key criterion not met was the question of lowest priority. Item missing data rates in the test treatment were either significantly lower than or not significantly different from those in the control treatment across the board. Revised wording showed improvements in nonresponse for the internet access question and for the “Cellular data plan” subscription category. Results for the response proportions analysis, in general, were as expected. Particularly noteworthy is the substantial increase in the share of households reporting a cellular data plan in the test treatment versus a mobile broadband plan in the control treatment. Whether looking at all households or specifically at households with a smartphone or tablet (handheld in control), the test proportion is about double the control proportion. Contrary to expectations, the share of households owning a desktop or laptop is lower in the test treatment compared with the control treatment. This is likely due to some owners of tablets (but not desktops or laptops) in the control treatment marking the “Desktop, laptop, netbook, or notebook computer” category, due to the lack of a specific category for tablets. The second unexpected result and unmet key research criteria involves the share of households reporting a DSL, cable, or fiber-optic subscription in the control treatment versus a broadband (high speed) subscription in the test treatment. This proportion is lower in the test treatment. Once more, there is a likely explanation for this difference, as respondents were offered three categories of this type in the control version of the question but only one category in the test version. Finally, findings from the response error analysis indicate that, across most Computer and Internet Use items, reliability is better or not significantly different for the test treatment compared with the control treatment. Worth noting is reduced inconsistency for the new “Tablet” and specific “Smartphone” categories, compared with the old “Handheld” category. For internet 35 subscriptions, there was an improvement for the “cellular data plan” category in the test treatment as compared to the “mobile broadband” category in the control treatment. As measured by the GDR, however, response reliability was less favorable in the test treatment than the control treatment for the “satellite internet” category and for the “high speed” category when compared to the combined “DSL/Cable/Fiberoptic” category. Even though these contrasts were not significant when using the IOI as the measure of reliability, the significant differences found in the GDRs provides evidence that respondent confusion may still be a problem with the test version. Due to the large improvement in reliability for the “cellular data plan” category in the test treatment, along with the other improvements to reliability in the computer use and internet access questions, the evidence suggests that in general the test questions performed better in terms of consistency of responses. Altogether, the 2016 ACS Content Test and analyses presented here validate the decision to implement the revised question wording on the 2016 production ACS. Whether considering item missing data rates, response proportions, or response error; in general data quality is not significantly different or is improved given changes to the questionnaire. Especially promising is the higher share of households that indicates owning a smartphone or tablet reporting an internet subscription, and much higher reports of mobile broadband subscriptions. The revised question wording will be reflected in the 2016 ACS data release, scheduled to begin in September 2017. 7 ACKNOWLEDGEMENTS The 2016 ACS Content Test would not have been possible without the participation and assistance of many individuals from the Census Bureau and other agencies. Their contributions are sincerely appreciated and gratefully acknowledged.  Census Bureau staff in the American Community Survey Office, Application Development and Services Division, Decennial Information Technology Division, Decennial Statistical Studies Division, Field Division, National Processing Center, Population Division, and Social, Economic, and Housing Statistics Division.  Representatives from other agencies in the Federal statistical system serving on the Office of Management and Budget’s Interagency Working Group for the ACS and the Topical Subcommittees formed by the Working Group for each topic tested on the 2016 ACS Content Test.  Staff in the Office of Management and Budget’s Statistical and Science Policy Office. The authors would like to thank the following individuals for their contributions to the analysis and review of this report: Kurt Bauman, Nicole Scanniello, Jason Lee, Broderick Oliver, and Elizabeth Poehler. 36 8 REFERENCES Biemer, P. (2011). Latent Class Analysis of Survey Error. Wiley, New York. Dusch, G., & Meier, F. (2012). 2010 Census Content Reinterview Survey Evaluation Report. U.S. Census Bureau, June 13, 2012. Retrieved January 13, 2017 from http://www.census.gov/2010census/pdf/2010_Census_Content_Reinterview_Survey_Eva luation_Report.pdf File, T., & Ryan, C. (November 2014). Computer and Internet Use in the United States: 2013. U.S. Census Bureau. Retrieved January 13, 2017 from http://www.census.gov/content/dam/Census/library/publications/2014/acs/acs-28.pdf Flanagan, P. (1996). Survey Quality & Response Variance (Unpublished Internal Document). U.S. Census Bureau. Demographic Statistical Methods Division. Quality Assurance and Evaluation Branch. Flanagan, P. (2001). Measurement Errors in Survey Response. University of Maryland Baltimore County, Baltimore, Maryland. Holm, S. (1979). “A Simple Sequentially Rejective Multiple Test Procedure,” Scandinavian Journal of Statistics, Vol. 6, No. 2: 65-70. Retrieved January 11, 2017 from https://www.jstor.org/stable/4615733?seq=1#page_scan_tab_contents National Telecommunications and Information Administration (October 2016). Digital Nation Data Explorer. Retreived June 12, 2017, from https://www.ntia.doc.gov/data/digital-nation-data-explorer#sel=internetUser&disp=map Pew Research Center. (January 2014). E-Reading Rises as Device Ownership Jumps. Retrieved January 13, 2017 from http://www.pewinternet.org/files/2014/01/PIP_E-reading_011614.pdf Pew Research Center. (October 2015). Technology Device Ownership: 2015. Retrieved December 29, 2016 from http://www.pewinternet.org/files/2015/10/PI_2015-10-29_device-ownership_FINAL.pdf Pew Research Center. (December 2015). Home Broadband 2015. Retrieved December 29, 2016 from http://www.pewinternet.org/files/2015/12/Broadband-adoption-full.pdf Pew Research Center. (September 2016). Book Reading 2016. Retrieved December 29, 2016 from http://assets.pewresearch.org/wp-content/uploads/sites/14/2016/08/PI_2016.09.01_Book- Reading_FINAL.pdf Pew Research Center. (November 2016). Device ownership 2016. Retrieve June 12, 2017 from http://www.pewresearch.org/data-trend/media-and-technology/device-ownership/ 37 Rao, J. N. K.; Scott, A. J. (1987). “On Simple Adjustments to Chi-Square Tests with Sample Survey Data,” The Annal of Statistics, Vol. 15, No. 1, 385-397. Retrieved on January 31, 2017 from http://projecteuclid.org/euclid.aos/1176350273 Reichert, J. (2015, July 7). 2016 Mandated Content Changes. U.S. Census Bureau. Stapleton, M., & Steiger, D. (2015). Cognitive Testing of the 2016 American Community Survey Content Test Items: Summary Report for Round 1 and Round 2 Interviews. Westat, Rockville, Maryland, January 2015. Steiger, D., Anderson, J., Folz, J., Leonard, M., & Stapleton, M. (2015). Cognitive Testing of the 2016 American Community Survey Content Test Items: Briefing Report for Round 3 Interviews. Westat, Rockville, Maryland, June 2015. U.S. Census Bureau. (2014). American Community Survey Design and Methodology (January 2014). Retrieved January 13, 2017 from http://www2.census.gov/programssurveys/acs/methodology/design_and_methodology/ac s_design_methodology_report_2014.pdf U.S. Census Bureau (2016). 2015 Planning Database Tract Data [Data file]. Retrieved January 11, 2017 from http://www.census.gov/research/data/planning_database/2015/ 38 Appendix A. Control and Test Questions in CATI, CAPI, and CFU Figure A1. CATI/CFU and CAPI Versions of the Control and Test Questions Control Version Test Version [LAPTOP] For the next few questions about computers, EXCLUDE GPS devices, digital music players, and devices with only limited computing capabilities, for example: household appliances. At this <house/apartment/mobile home/unit>, do you or any member of this household own or use a desktop, laptop, netbook, or notebook computer? Yes No [HANDHELD] At this <house/apartment/mobile home/unit>, Do you or any member of this household own or use a handheld computer, smart mobile phone, or other handheld wireless computer? Yes No [COMPOTH] At this <house/apartment/mobile home/unit>, Do you or any member of this household own or use some other type of computer? Yes No (Skip to internet access question) [COMPOTHW] What is this other type of computer? ________ ---------------------------------------------------------------- [WEB] At this <house/apartment/mobile home/unit>, do you or any member of this household access the Internet? Yes No (Skip to vehicle question) [SUBSCRIBE] At this <house/apartment/mobile home/unit>, Do you or any member of this household access the Internet with or without a subscription to an Internet service? With a subscription to an Internet service Without a subscription to an Internet service (Skip to vehicle question) ------------------------------------------------------- [DIALUP] At this <house/apartment/mobile home/unit>, do you or any member of this household subscribe to the Internet using a dial-up service? Yes No 39 [LAPTOP] At this <house/apartment/mobile home/unit>, do you or any member of this household own or use a desktop or laptop computer? Yes No [SMARTPHONE] At this <house/apartment/mobile home/unit>, Do you or any member of this household own or use a smartphone? Yes No [TABLET] At this <house/apartment/mobile home/unit>, Do you or any member of this household own or use a tablet or other portable wireless computer? Yes No [COMPOTH] At this <house/apartment/mobile home/unit>, Do you or any member of this household own or use some other type of computer? Yes No (Skip to Internet access question) [COMPOTHW] What is this other type of computer? ________ ------------------------------------------------------- **[ACCESS] – Internet mode **[WEB] – CATI/CAPI/CFU At this <house/apartment/mobile home/unit>, do you or any member of this household have access to the Internet? Yes No (Skip to vehicle question) [SUBSCRIBE] At this <house/apartment/mobile home/unit>, Do you or any member of this household pay a cell phone company or Internet service provider to access the Internet? Yes No (Skip to vehicle question) ------------------------------------------------------- Figure A1. (continued). CATI/CFU and CAPI Versions of the Control and Test Questions Control Version Test Version [DSL] At this <house/apartment/mobile home/unit>, Do you or any member of this household subscribe to the Internet using a DSL service? Yes No [BROADBND] Do you or any member of this household access the Internet using a cellular data plan for a smartphone or other mobile device? Yes No [MODEM] At this <house/apartment/mobile home/unit>, Do you or any member of this household subscribe to the Internet using a cable-modem service? Yes No [FIBEROP] At this <house/apartment/mobile home/unit>, Do you or any member of this household subscribe to the Internet using a fiber-optic service? Yes No [BROADBND] At this <house/apartment/mobile home/unit>, Do you or any member of this household subscribe to the Internet using a mobile broadband plan for a computer or a cell phone? Yes No [SATELLITE] At this <house/apartment/mobile home/unit>, Do you or any member of this household subscribe to the Internet using a satellite Internet service? Yes No [OTHSVCE] At this <house/apartment/mobile home/unit>, Do you or any member of this household subscribe to the Internet using some other service? Yes No (Skip to vehicle question) [OTHSVCEW] What is this other type of Internet service? ________ [HISPEED] Do you or any member of this household access the Internet using broadband or high speed Internet service such as cable, fiber optic, or DSL service installed in this <house/apartment/mobile home/unit>? Yes No [SATELLITE] Do you or any member of this household access the Internet using a satellite Internet service installed in this <house/apartment/mobile home/unit>? Yes No [DIALUP] Do you or any member of this household access the Internet using a dial-up Internet service installed in this <house/apartment/mobile home/unit>? Yes No [OTHSVCE] Do you or any member of this household access the Internet using some other service? Yes No (Skip to vehicle question) [OTHSVCEW] What is this other type of Internet service? ________ 40 Appendix B. Unit Response Rates Supplemental Table Table B1. Unit Response Rates by Designated High (HRA) and Low (LRA) Response Areas Mode Total Response HRA LRA Difference Self-Response HRA LRA Difference Internet HRA LRA Difference Mail HRA LRA Difference CATI HRA LRA Difference CAPI HRA LRA Difference Test Interviews Test Percent Control Interviews Control Percent Test minus Control P-Value 19,400 7,556 11,844 - 13,131 6,201 6,930 - 8,168 4,119 4,049 - 4,963 2,082 2,881 - 872 296 576 - 5,397 1,059 4,338 - - 94.3 (0.4) 91.5 (0.3) 2.7 (0.5) - 59.7 (0.7) 33.2 (0.4) 26.5 (0.8) - 39.6 (0.6) 19.4 (0.3) 20.2 (0.6) - 20.0 (0.4) 13.8 (0.3) 6.2 (0.5) - 9.0 (0.5) 7.9 (0.4) 1.1 (0.6) - 82.2 (1.0) 85.8 (0.5) -3.7 (1.1) 19,455 7,608 11,847 - 13,284 6,272 7,012 - 8,112 4,048 4,064 - 5,172 2,224 2,948 - 880 301 579 - 5,291 1,035 4,256 - - 94.5 (0.3) 91.0 (0.3) 3.5 (0.5) - 60.6 (0.7) 33.6 (0.4) 27.0 (0.8) - 39.1 (0.6) 19.5 (0.3) 19.6 (0.7) - 21.5 (0.4) 14.1 (0.3) 7.4 (0.4) - 9.6 (0.6) 8.0 (0.3) 1.6 (0.7) - 82.7 (0.9) 85.0 (0.4) -2.3 (1.0) - -0.2 (0.6) 0.5 (0.5) -0.7 (0.7) - -0.9 (0.9) -0.4 (0.6) -0.5 (1.2) - 0.5 (0.8) 0.1 (0.4) 0.6 (0.9) - -1.5 (0.6) -0.3 (0.4) -1.1 (0.7) - -0.6 (0.8) -0.1 (0.5) -0.5 (0.9) - -0.5 (1.3) 0.8 (0.7) -1.3 (1.5) - 0.72 0.29 0.33 - 0.31 0.55 0.66 - 0.51 0.87 0.52 - 0.02* 0.43 0.11 - 0.44 0.85 0.58 - 0.69 0.23 0.36 Source: U.S. Census Bureau, 2016 American Community Survey Content Test Note: Standard errors are in parentheses. Minor additive discrepancies are due to rounding. P-values with an asterisk (*) indicate a significant difference based on a two-tailed t-test at the α=0.1 level. The weighted response rates account for initial sample design as well as CAPI subsampling. 41 Appendix C. Benchmarks C.1. Research Questions 1. How do the proportions for each category of computers in each treatment compare with proportions found in the Current Population Survey (CPS) and from surveys done by the Pew Research Center? 2. How do the proportions in each treatment compare with proportions found in the CPS for the Internet access question? 3. How do the proportions of mobile broadband subscribers compare to Pew Research findings as well as the most recent CPS results? C.2. Methodology We compared the 2016 ACS Content Test data from both control and test treatments with the most current version of other surveys available as benchmarks for the comparisons. These comparisons allow us to tell whether our results differ from other reliable sources. As a cautionary note, although the other surveys provide benchmarks, they are not statistically comparable with the Content Test results, given differences in universe, timing, question wording, and survey design between the sources. Useful comparisons can still be made, however, as the overall distributions should be similar between surveys or differ in expected ways. Types of Computers For the topic of types of computers, we compared data from both control and test treatments to information from the July 2015 Current Population Survey (CPS) Computer and Internet Use Supplement and recent Pew Research Center surveys. The CPS, a national household survey, has collected data on computer use since 1984 and internet use since 1997 in an occasional supplement. The July 2015 CPS Supplement included questions about access to desktops, laptops, smartphones, and tablets, as well as wearable technology and smart TVs.20 For our comparison, we looked at CPS estimates on use of 1) a desktop or laptop/notebook, 2) a smartphone, and 3) a tablet/e-book reader or a wearable internet-connected device (such as a smart watch or glasses, with the item offering specific examples). The smartphone estimate is a recode rather than a direct question, as the direct question on the survey asks about both cellular phones and smartphones. Following guidance by the National Telecommunications and Information Administration (NTIA), the supplement sponsor, we created the smartphone recode using the item on cellular phones and smartphones, combined with information on internet use from any location and subscription to a mobile data plan. Similar to the Content Test, the universe for the CPS is households. The Pew Research Center began asking about cellphone ownership in 2000, desktops or laptops in 2004, tablets in 2010, and smartphones in 2011. The most recent data on smartphones is from 2015, and 2016 data on laptops/desktops and tablets are available. Pew respondents receive 20 Complete technical documentation, including question wording, for the 2015 CPS Computer and Internet Use Supplement is available at http://www2.census.gov/programs-surveys/cps/techdocs/cpsjul15.pdf. 42 direct questions about whether they have 1) a desktop or laptop computer, and 2) a tablet (with the item offering specific examples). Note that question wording regarding tablet ownership in the Pew survey differs somewhat from the Content Test wording. The test treatment asks about owning or using a “Tablet or other portable wireless computer,” whereas the comparable Pew item only asks about tablet computers. Smartphone owners are identified through two questions, with the first asking about having a cell phone, and the second asking if the person’s cell phone is a smartphone (with the item noting examples).21 Pew data are typically collected for adults and are therefore not statistically comparable with ACS data on Computer and Internet Use, which represent results for each housing unit (Pew Research Center, 2015). Proportions of desktop/laptop ownership from both the test and control treatments were used in comparisons. However, only test proportions on smartphone and tablet ownership were used, as the control treatment lacks categories specific to these devices. Internet Access For the topic of internet access, we again compared data from both control and test treatments to information from the 2015 CPS Supplement and a recent Pew Research Center survey. The 2015 CPS Supplement asked a series of five questions about how household members connect to the internet at home. Those who stated they used a plan bought from 1) a company, or 2) a public agency, nonprofit, or cooperative were considered to have access with a subscription. As for the Content Test, the universe for the CPS estimate is households. The Pew estimate for access with a subscription was captured in 2015, and includes those with either a smartphone or a home broadband subscription. Pew separately reported on dial-up subscribers. As noted above, Pew estimates have a universe of adults, in contrast with the Content Test universe of households. Internet Subscription For the topic of internet subscription type, we compared data from both control and test treatments to information from the 2015 CPS Supplement. Although most internet subscription types are not captured in Pew Research Center surveys, we were able to use a Pew estimate on dial-up service for comparison. The 2015 CPS questions contain categories similar to the ACS Content Test: 1) mobile internet service or a data plan; 2) high-speed internet such as cable, DSL, or fiber-optic; 3) satellite; 4) dial-up; and 5) some other service. Also similar to the Content Test, the universe for CPS estimates is households with an internet subscription. As with other Pew estimates, the universe for the Pew dial-up estimate is adults aged 18 and over. 21 Pew collects data through phone interviews, sampling both those with landline phones and those with cellular phones. The cellular phone sample is automatically marked as having a cell phone (Pew Research Center, October 2015). 43 C.3. Results How do the proportions for each category of computers in each treatment compare with proportions found in the Current Population Survey (CPS) and from surveys done by the Pew Research Center? Table C1 contains proportions of households owning various types of computers from the test and control treatments. It also contains proportions from the 2015 CPS Computer and Internet Use Supplement and recent surveys by the Pew Research Center. Although CPS and Pew estimates are not statistically comparable to those from the Content Test, useful comparisons can still be made. An important difference to note across surveys is that the Content Test and CPS estimates are for households, whereas Pew estimates are for adults aged 18 and over. Because not all household members in a household with a given type of computer would be expected to report that they own such a computer, the percentage of households with that type of computer should be larger than the percentage of people with that type. In addition, because computer ownership and use have been growing, data collected at a later time would be expected to show higher levels of ownership than those collected at an earlier time. Table C1. Benchmark Estimates, Types of Computer Question Test Percent 79 (± 0.7) 78 (± 0.7) 60 (± 0.8) Item Desktop or laptop Smartphone Tablet Sources: U.S. Census Bureau, 2016 American Community Survey Content Test and 2015 Current Population Survey Computer and Internet Use Supplement; Pew Research Center (Surveys conducted June 10-July 12, 2015 and March 7-April 4, 2016.) Note: N/A indicates not applicable. Ninety percent margins of error are shown in parentheses. Estimates across surveys are not statistically comparable. Content Test and CPS estimates are for households, whereas Pew estimates are for adults aged 18 and over. CPS Percent 71 (± 0.3) 62 (± 0.4) 39 (± 0.4) Pew Percent 74 (± 2.4) 68 (± 2.1) 48 (± 2.4) Control Percent 81 (± 0.7) N/A N/A In general, we see that Content Test estimates of computer ownership—including desktop/laptop, smartphone, and tablet ownership—conform to expectations, with the Content Test showing results that, at surface value, are somewhat higher than Pew and CPS benchmarks. The large difference between CPS and Content Test estimates of tablet use is not easily explained. Nonetheless, these results suggest that the Content Test questions do a reasonably good job of measuring computer ownership. How do the proportions in each treatment compare with proportions found in the Current Population Survey for the Internet access question? Estimates of internet access from the test treatment, control treatment, CPS, and Pew are displayed in Table C2. In summary, Content Test proportions of access with a subscription conform to expectations relative to Pew estimates, adding to our confidence in Content Test estimates of internet access. The apparent difference between Content Test and CPS estimates of internet subscriptions is more problematic. This difference may partially result from CPS issues, 44 in that those data show a (nonsignificant) decline in household internet use between 2012 and 2015 not evident in other data.22 This difference may be addressed in future research. Table C2. Benchmark Estimates, Internet Access Question Test Percent 84 (± 0.7) 16 (± 0.7) Item Access with subscription No subscription Sources: U.S. Census Bureau, 2016 American Community Survey Content Test and 2015 Current Population Survey Computer and Internet Use Supplement; Pew Research Center (Surveys conducted in April, July, and November 2015.) Note: N/A indicates not applicable. Ninety percent margins of error are shown in parentheses. Estimates across surveys are not statistically comparable. Content Test and CPS estimates are for households, whereas Pew estimates are for adults aged 18 and over. Pew estimate represents those with a smartphone or home broadband connection. Another two percent have a dial-up connection. CPS Percent 71 (± 0.3) 29 (± 0.3) Pew Percent 80 (± 1.1) N/A Control Percent 82 (± 0.7) 18 (± 0.7) How do the proportions of mobile broadband subscribers compare to Pew Research findings as well as the most recent CPS results? Table C3 shows estimates of internet subscription type from the test treatment, control treatment, CPS supplement, and Pew. Again, whereas the universe for the types of computers and internet access questions is all eligible households, the Content Test universe for the internet subscriptions question is households with internet access with a subscription. Proportions of dial- up are not that different among the Content Test treatments, CPS, and Pew. Nor are differences great between the test treatment, control treatment, and CPS regarding high speed broadband service such as DSL, cable, or fiber-optic; satellite subscriptions; or subscription to some other service. Benchmarking the test and control estimates for households with a mobile broadband subscription against the CPS provides additional evidence that the new question wording improves measurement of mobile broadband. However, the difference between the test and CPS estimates for mobile broadband should be explored further in future research. Table C3. Benchmark Estimates, Internet Subscription Type Question Item Dial-up Test Percent 2 (± 0.3) Control Percent 3 (± 0.3) CPS Percent 1 (± 0.1) High speed or DSL/Cable/Fiber-optic 81 (± 0.8) 85 (± 0.8) 76 (± 0.4) Cellular data plan or Mobile broadband 80 (± 0.7) 40 (± 1.0) 61 (± 0.4) Pew Percent 2 (± 2.3) N/A N/A Satellite Other service Sources: U.S. Census Bureau, 2016 American Community Survey Content Test and 2015 Current Population Survey Computer and Internet Use Supplement; Pew Research Center (Survey conducted June 10-July 12, 2015.) Note: N/A indicates not applicable. Ninety percent margins of error are shown in parentheses. Estimates across surveys are not statistically comparable. Content Test and CPS estimates are for households with an internet subscription, whereas Pew estimates are for adults aged 18 and over. 6 (± 0.4) 2 (± 0.3) 3 (± 0.2) 1 (± 0.1) 6 (± 0.5) 2 (± 0.2) N/A N/A 22 See the relevant chart from NTIA’s Digital Nation Data Explorer, available at https://www.ntia.doc.gov/data/digital-nation- data-explorer#sel=internetAtHome&demo=&pc=prop&disp=chart. 45

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/brockwebb/open-census-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server