How to Conduct P r e t e s t i n g

Similar documents
Applied Statistics in Business & Economics, 5 th edition

Investigating Similar Triangles and Understanding Proportionality: Lesson Plan

A Pilot Study on Functional Processing: Inferences of Pairwise Relationships in Systems of Three Variables. Kent L. Norman and Benjamin K.

Student Questionnaire (s) Main Survey

Rural Pennsylvania: Where Is It Anyway? A Compendium of the Definitions of Rural and Rationale for Their Use

Detailed Contents. 1. Science, Society, and Social Work Research The Process and Problems of Social Work Research 27

Typical information required from the data collection can be grouped into four categories, enumerated as below.

Green Chemistry Member Survey April 2014

A Better Way to Do R&R Studies

Manipulating Radicals

Lesson Ten. What role does energy play in chemical reactions? Grade 8. Science. 90 minutes ENGLISH LANGUAGE ARTS

2002 HSC Notes from the Marking Centre Geography

CHOOSING THE RIGHT SAMPLING TECHNIQUE FOR YOUR RESEARCH. Awanis Ku Ishak, PhD SBM

Academic Affairs Assessment of Student Learning Report for Academic Year

An Introduction to Scientific Research Methods in Geography Chapter 3 Data Collection in Geography

Intermediate Algebra. Gregg Waterman Oregon Institute of Technology

Data Collection. Lecture Notes in Transportation Systems Engineering. Prof. Tom V. Mathew. 1 Overview 1

Matrix ranking: a means to discussion

Cambridge O Level Additional Mathematics. Paper 1

Donna J. Kain, PhD and Catherine F. Smith, PhD East Carolina University

SPIRITUAL GIFTS. ( ) ( ) 1. Would you describe yourself as an effective public speaker?

F rom the C u s t o m er s Perspective... N o rt h w e s t e rn Memorial Hospital Delivers Quality Care With the S Y S T I M A X GigaSPEED Solution

Determining the Appropriateness of Terminology in Content-Aligned Assessments for Middle School Students: Examples from Plate Tectonics

Qualitative and Quantitative Research Methods

Chemistry SCI1100. Course Description. Rationale. Prerequisite. Measurable Learning Outcomes

Writing Patent Specifications

CONCEPTS OF GENETICS BY ROBERT BROOKER DOWNLOAD EBOOK : CONCEPTS OF GENETICS BY ROBERT BROOKER PDF

Chapter Goals. To introduce you to data collection

DATA DISAGGREGATION BY GEOGRAPHIC

LEGAL REASONING A S A TYPE OF PRACTICAL REASONING. J. David NEWELL

REVIEW OF THE EMERGING ISSUES TASK FORCE

Zimbabwean undergraduate mathematics in-service teachers understanding of matrix operations

Use precise language and domain-specific vocabulary to inform about or explain the topic. CCSS.ELA-LITERACY.WHST D

8 th Arctic Regional Hydrographic Commission Meeting September 2018, Longyearbyen, Svalbard Norway

Variation of geospatial thinking in answering geography questions based on topographic maps

Module 9: Sampling IPDET. Sampling. Intro Concepts Types Confidence/ Precision? How Large? Intervention or Policy. Evaluation Questions

Deep Algebra Projects: Algebra 1 / Algebra 2 Go with the Flow

CLARKSON SECONDARY SCHOOL. Course Name: Grade 12 University Chemistry

Test and Evaluation of an Electronic Database Selection Expert System

Objective: Construct a paper clock by partitioning a circle into halves and quarters, and tell time to the half hour or quarter hour.

An Analysis of Field Test Results for Assessment Items Aligned to the Middle School Topic of Atoms, Molecules, and States of Matter

Br anding Recommendations 5.

Notifications and Accretions: Facility Role

Section 2. Indiana Geographic Information Council: Strategic Plan

Notes 3: Statistical Inference: Sampling, Sampling Distributions Confidence Intervals, and Hypothesis Testing

On Publishing in the AAS Journals (as an author)

Economic and Social Council

Determining Sufficient Number of Imputations Using Variance of Imputation Variances: Data from 2012 NAMCS Physician Workflow Mail Survey *

Measure of Design Thinking-based Regional Innovation: Focusing on Gamcheon Culture Village

Underground Railroad Grades 6-8. Time Frame: 1-3 class periods, depending on the extent of outside research

Limiting Reactants An analogy and learning cycle approach

PAPER EVALUATION OF TRUNCATED ARROW-PER-LANE GUIDE SIGNS. WORD COUNT (5840): Abstract (248) + Text (4092) + Tables (5) + Figures (1) Authors:

Econometric Modelling Prof. Rudra P. Pradhan Department of Management Indian Institute of Technology, Kharagpur

New Zealand Population Distribution

THE REAL ESTATE COUNCIL OF BRITISH COLUMBIA. IN THE MATTER OF THE REAL ESTATE SERVICES ACT S.B.C. 2004, c. 42 as amended AND IN THE MATTER OF

A Journey Back in Time

The Characteristics of States of Matter Concept Attributes of 3 rd to 6 th Grade Elementary School Students

A Note on Methods. ARC Project DP The Demand for Higher Density Housing in Sydney and Melbourne Working Paper 3. City Futures Research Centre

Test about Voltage - A Basic Term in Electricity Results

Machine Learning, Fall 2009: Midterm

Objective: Construct a paper clock by partitioning a circle and tell time to the hour. (10 minutes)

Introduction to the Next Generation Science Standards (NGSS) First Public Draft

from

A S U M M A RY OF Chemigation Applicatio n Studies

MATH 137 : Calculus 1 for Honours Mathematics. Online Assignment #2. Introduction to Sequences

Developing the Next Generation Science Standards

ZecoByte. Panelgranskning. SQCT - Survey Quality Control Tool. technical presentation. 12 april 2012

Dublin City Schools Science Graded Course of Study Physical Science

104 Business Research Methods - MCQs

Business Statistics: A First Course

Use precise language and domain-specific vocabulary to inform about or explain the topic. CCSS.ELA-LITERACY.WHST D

WORLD GEOGRAPHY GRADE 10

Keeping Fit. Be t ween 1980 and 2000, t h e. To d ay, i t s more ch a l l e n g i n g. Overview. By Peter Winkler. Science Background.

The Comprehensive Report

Computational Tasks and Models

GOVERNMENT MAPPING WORKSHOP RECOVER Edmonton s Urban Wellness Plan Mapping Workshop December 4, 2017

BRIDGE CIRCUITS EXPERIMENT 5: DC AND AC BRIDGE CIRCUITS 10/2/13

Regression Analysis in R

CSCI3390-Assignment 2 Solutions

Higher National Unit specification: general information

C A R I B B E A N E X A M I N A T I O N S C O U N C I L REPORT ON CANDIDATES WORK IN THE CARIBBEAN ADVANCED PROFICIENCY EXAMINATION MAY/JUNE 2009

Chemistry. Course Description. Rationale. Prerequisite. Measurable Learning Outcomes SCI1100

ISO/TR TECHNICAL REPORT. Nanotechnologies Methodology for the classification and categorization of nanomaterials

Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi

Algebra Year 9. Language

GIS Institute Center for Geographic Analysis

This Session s Assignments. Once you have completed your Assignment:

Measurement Invariance (MI) in CFA and Differential Item Functioning (DIF) in IRT/IFA

CHAPTER 3: DATA ACQUISITION AND ANALYSIS. The research methodology is an important aspect of research to make a study results

Sun s rotation period. Teacher s Guide - Intermediate Level CESAR s Science Case

HOSTING AN INTERNATIONAL CONFERENCE ON THE HISTORY OF CARTOGRAPHY: GUIDELINES,

Community participation in sustainable tourism - A case study of two indigenous communities

Most people said that they understand force and acceleration. GOOD!

Featured Alumna Sarah Caudill ( 06)

Data Collection: What Is Sampling?

Intro to Info Vis. CS 725/825 Information Visualization Spring } Before class. } During class. Dr. Michele C. Weigle

value of the sum standard units

THE DATA REVOLUTION HAS BEGUN On the front lines with geospatial data and tools

PALS: Neighborhood Identification, City of Frederick, Maryland. David Boston Razia Choudhry Chris Davis Under the supervision of Chao Liu

SURVEY SAMPLING. ijli~iiili~llil~~)~i"lij liilllill THEORY AND METHODS DANKIT K. NASSIUMA

Transcription:

ASA Seri e s What Is a Survey? What Is a Survey? How to Plan a Survey How to Collect Survey Data Judging the Quality of a Survey How to Conduct Pre t e s t i n g What A re Focus Gro u p s? M o re About Mail Survey s What Is a Margin of Erro r? Designing a Questionnaire M o re About Telephone Survey s ASA Seri e s What Is a Survey? How to Conduct P r e t e s t i n g P roduced by Section on Survey Research Methods American Statistical Association 1429 Duke Stre e t Alexandria, VA 22314-3415 USA Section on Survey Research Methods American Statistical Association 8/00 1997 American Statistical Association

How to Conduct Pretesting This pamphlet, How to Conduct Pre t e s t i n g, is the fifth in the ASA series What Is a Surv e y? It looks at how to check out or pretest a questionnaire among the most important planning steps in any s u rv e y. Care at this point will not guarantee success, but a lack of care will almost always lead to extra costs and a lost opportunity to collect the re q u i re d i n f o rm a t i o n. T h e What is a Surv e y? series is written primarily for the general public. Its overall goal is to improve survey literacy among individuals who participate in s u rveys or use survey results. The series is designed to promote a better understanding of what is involved in carrying out sample surv e y s e s p e c i a l l y those aspects that have to be taken into account in evaluating the results of surv e y s. T he systematic checking or pre t e s t- ing of a questionnaire is central to planning a good surv e y. As mentioned earlier in this series, the surv e y sponsors should play a major role in developing the data-collection instruments being proposed including any testing being done. Much of the accuracy and interpretability of the survey re s u l t s hinge on this pretesting step w h i c h should never be omitted. P retesting is critical for identifying questionn a i re problems. These can occur for both Pretesting is respondents and interviewers re g a rd i n g question content, skip p a t t e rns, or form a t- ting. Problems with question content include confusion with the overall meaning of the question, as well as misinterpretation of individual terms or critical for i d e n t i f y i n g p ro blems for both respondents and i n t e r v i e w e r s. concepts. Problems with how to skip or navigate from question to question may result in missing data and frustration for both interv i e w- ers and respondents. Questionnaire form a t t i n g c o n c e rns are particularly relevant to self- 1

a d m i n i s t e red questionn a i res, and if unaddre s s e d, P r e t e s t i n g may lead to loss of vital i n c o r p o ra t e s i n f o rm a t i o n. many diffe r e n t P retesting is a bro a d methods or t e rm that incorporates many d i ff e rent methods or combinations of methods. c o m b i n a t i o n s of methods. This pamphlet briefly describes eight suggested t e c h n i q u e s that can be used to pretest questionnaires. These techniques have diff e rent strengths and weaknesses. They can be invaluable for identifying problems with draft questionnaires and also for evaluating s u rveys in the field. Types of Pretesting P retesting techniques are divided into two major categories p re - f i e l d and f i e l d. Pre - f i e l d techniques are generally used during the pre l i m i- n a ry stages of questionn a i re development. They include re s p o n - dent focus groups and cognitive laboratory i n t e rv i e w s. Six field techniques that test questionnaire s under operational Pretesting is divided into two c a t e g o r i e s p re - fi e l d and fi e l d t e c h n i q u e s. conditions are also covered. These include behavior coding of interv i e w e r / re s p o n d e n t interactions, interviewer debriefings, re s p o n- dent debriefings, split-sample tests, and the analysis of item nonresponse rates and response distributions. used in a variety of ways to assess the questionanswering process. Such groups may gather information about a topic before questionnaire construction begins (for example, to learn how people stru c t u re their thoughts about a topic, their understanding of general concepts or specific term i n o l o g y, or their opinions about the sensitivity or difficulty of the questions). Focus groups help identify variations in language, term i n o l o g y, or interpretation of questions and response options. Self-administere d q u e s t i o n n a i res can be p retested in a focus g roup, to learn about Focus groups the appearance and formatting of the ques- p rovide the t i o n n a i re. In addition, opportunity to knowledge of content observe a large p roblems is gained. amount of interaction in a limited One of the main advantages of focus g roups is the opport u n i- period of time. ty to observe a gre a t deal of interaction on a topic in a limited period of time. They also produce information and insights that may be less accessible without the give and take found in a group. Because of their interactive nature, however, focus g roups do not permit a good test of the n o rmal interviewing process. Researc h e r s also do not have as much control over the p rocess as with other pretesting methods. (For example, one or two people in the group may dominate the discussion and restrict input from other focus group members.) 1. Respondent Focus Groups 2. Cognitive Laboratory Interviews Focus groups a form of in-depth gro u p Cognitive laboratory interviews are also i n t e rv i e w i n g a re conducted early in the generally used early in the questionnaire q u e s t i o n n a i re development cycle and can be 2 3

development cycle. They consist of one-onone interviews using a stru c t u red questionn a i re in which respondents describe their thoughts while answering the survey questions. Think aloud interviews, as this technique is T h i n k a l o u d called, can be conducted either concurrently or re t- rospectively (i.e. the respondents verbalizations of their thought processes can occur either during or after the completion of the q u e s t i o n n a i re). L a b o r a t o ry interv i e w s p rovide an import a n t cognitive i n t e r v i e w s c a n b e conducted either c o n c u r r e n t l y o r r e t ro s p e c t i v e l y. means of finding out directly from re s p o n- dents what their problems are with the quest i o n n a i re. In addition, small numbers of interviews (as few as 15) can yield inform a t i o n about major problems such as re s p o n d e n t s repeatedly identifying the same questions and concepts as sources of confusion. Because sample sizes are not large, re p e a t e d p retesting of an instrument is often possible. After one round of lab interviews is completed, re s e a rchers can diagnose pro b l e m s, revise question wording to resolve these p roblems, and conduct additional interv i e w s to see if the new questions are better. Cognitive interviews can incorporate followup questions by the interviewer in addition to respondents statements of their thoughts. D i ff e rent types of follow-up questions are used. P robing questions a re used when the re s e a rcher wants to focus the respondent on p a rticular aspects of the question-re s p o n s e t a s k. (For example, the interviewer may ask how respondents chose their answers, how they interpreted reference periods, or what they thought a particular term meant.) P a r a p h r a s i n g (i.e., asking the respondents to repeat the question in their own words) permits the re s e a rcher to learn whether the respondent understands the question and interprets it in the manner intended. It may also reveal better word i n g s for questions. 3. Behavior Coding Behavior coding of re s p o n d e n t - i n t e rv i e w e r interactions involves systematic coding of the interaction between interviewers and respondents from live or taped interv i e w s. The emphasis is on specific aspects of how the interviewer asked the question and how the respondent reacted. When used for quest i o n n a i re assessment, the coding highlights i n t e rviewer or respondent behaviors indicative of a problem with the question, the re s p o n s e categories, or the re s p o n d e n t s ability to form an adequate re s p o n s e. For example, if a re s p o n- dent asks for clarification after hearing the question, it is likely that some aspect of the question caused confusion. Likewise, if a respondent interrupts before the i n t e rviewer finishes reading the question, then the respondent may miss inform a t i o n that might be important to giving a corre c t a n s w e r. In contrast to pre-field techniques, behavior 4 5

coding re q u i res a sample size sufficient to a d d ress analytic re q u i rements. For example, if the q u e s t i o n n a i re contains many skip patterns, it is necess a ry to select a large enough sample to permit observ a- tion of various movements through the questionnaire. The determining sample Behavior coding allows s y s t e m a t i c detection of questionnaire p ro bl e m s. sizes for behavior coding should take into account the relevant population gro u p s for which separate analyses a re desire d. The value of behavior coding is that it allows s y s t e m a t i c detection of questions that have large numbers of behaviors that reflect pro b- lems. It is not usually designed to provide answers about the sourc e of the problems. It also may not distinguish which of several similar versions of a question is better. 4. Respondent Debriefings Respondent debriefings involve incorporating stru c t u red follow-up questions at the end of a field test interview to elicit quantitative and qualitative information about re s p o n- dents interpretations of survey questions. For p retesting purposes, the primary objective is to determine whether concepts and questions a re understood by respondents in the same way that the survey sponsors intended. Respondent debriefings can also be used to evaluate other aspects of respondents tasks, such as their use of re c o rds to answer surv e y questions or their understanding of the purpose of the interv i e w. In addition, re s p o n d e n t debriefings can be useful in determining the reason for respondent misunderstandings. Sometimes results of respondent debriefings show a question is superfluous and can be eliminated. Altern a t i v e l y, additional questions may need to be included in the final questionn a i re. Finally, the debriefings may show that concepts or questions cause confusion or misunderstanding as far as the intended meaning is concern e d. Some survey goals may need to be gre a t- ly modified or even d ro p p e d. A critical aspect of a successful re s p o n d e n t debriefing is that question designers and re s e a rchers must R e s p o n d e n t debriefings can be useful in determining the reason for respondent m i s u n d e r s t a n d i n g s. have a clear idea of potential problems so that good debriefing questions can be developed. Ideas about potential problems can come from pre-field techniques conducted prior to the field test, from analysis of data f rom a previous surv e y, from careful review of q u e s t i o n n a i res, or from observation of actual i n t e rv i e w s. Respondent debriefings have the potential to supplement information obtained fro m behavior coding. As previously discussed, 6 7

behavior coding can demonstrate the existence of problems but does not always indicate the source of the problem. When designed pro p e r l y, the results of re s p o n d e n t debriefings can provide information about the p roblem sources and may reveal problems not evident from the response behavior. 5. Interviewer Debriefings I n t e rviewer debriefings traditionally have been the primary method to evaluate field tests. The interv i e w e r s who conduct the surv e y Interviewer debriefings have t ra d i t i o n a l l y been the primary method to evaluate field t e s t s. field tests are queried to use their direct contact with respondents to enrich the questionnaire d e s i g n e r s understanding of questionnaire p ro b l e m s. Although import a n t, i n t e rviewer debriefings a re not adequate as the sole evaluation method. I n t e rviewers may not always be accurate re p o rters of certain types o f q u e s t i o n n a i re problems for several re a s o n s : When interviewers re p o rt a problem it is not known whether it was troublesome for one respondent or for many. I n t e rviewer re p o rts of problem questions may reflect their own pre f e rence for a question rather than respondent confusion. Experienced interviewers sometimes change the wording of problem questions as a matter of course to make them work and may not even realize they have done so. I n t e rviewer debriefings can be conducted in several diff e rent ways: G roup-setting debriefings a re the most common 8 method, involving a focus group with the field test interv i e w e r s. Rating form s obtain more quantitative information by asking interviewers to rate each question in the pretest questionnaire on selected characteristics of interest to the re s e a rchers (whether the interviewer had t rouble reading the question as written and whether the respondent understood the w o rds or ideas in the question, among o t h e r s ). S t a n d a rdized interviewer debriefing questionnaire s collect information about the interviewers perceptions of the problem, prevalence of a pro b- lem, reasons for the problem, and pro p o s e d solutions to a problem. They can also be used to ask about the magnitude of specific types of p roblems and to test an interv i e w e r s knowledge of subject-matter concepts. 6. Split-Panel Tests Split-panel tests refer to controlled experimental testing among questionnaire variants or interviewing modes to determine which is better or to measure Multiple versions of a dra f t questionnaire can be tested under c o n t ro l l e d e x p e r i m e n t a l conditions in a split-panel test. d i ff e rences between them. For pre t e s t i n g multiple versions of a q u e s t i o n n a i re there needs to be a pre v i o u s l y d e t e rmined standard by which to judge the d i ff e re n c e s. Split-panel tests are also used to calibrate the effect of changing q u e s t i o n s p a rt i c u l a r l y i m p o rtant in the redesign and testing of s u rveys where the comparability of the data collected over time is an issue. 9

Split-panel tests can incorporate changes in a single question, a set of questions, or an e n t i re questionnaire. It is important to pro v i d e for adequate sample sizes in a split-panel test so that diff e rences of substantive interest can be measured well. It is also imperative that these tests involve the use of randomized assignment so diff e rences can be attributed to the question or questionnaire, and not to something else. 7. Analysis of Item Nonresponse Rates Analysis of item nonresponse rates from the data collected during a field test (involving one Both response or multiple panels) can and nonresponse p rovide useful inform a- tion about how well the rates provide useful info r m a t i o n This can be done by q u e s t i o n n a i re works. looking at how often about how well a items are missing q u e s t i o n n a i r e (item nonre s p o n s e w o r k s. r a t e s ). These rates can be i n f o rmative in two ways: Don t know rates can determine how difficult a task is for respondents to do. Refusal rates can determine how often respondents find certain questions or versions of a question too sensitive to be answered. 8. Analysis of Response Distributions Analysis of response distributions for an item can be used to determine whether diff e re n t question wordings or question sequences produce diff e rent response patterns. This kind of analysis is most useful when pretesting more than one version of a questionnaire or a single q u e s t i o n n a i re in which some known distribution of characteristics exists for comparative p u r p o s e s. When looking at response distributions in split-panel tests, the results do not necessarily reveal whether one version of a question p roduces a better understanding of what is being asked than another. Knowledge of d i ff e rences in response patterns alone is not s u fficient to decide which question best conveys the concept of intere s t. At times response distribution analysis demonstrates that revised question word i n g has no effect on estimates. Response distribution analyses should not be used alone to evaluate modifications in question wording or sequencing. It is useful only in conjunction with other question evaluation methods such as respondent debriefings, interv i e w e r debriefings, and behavior coding. Combining Methods Both pre-field and field testing should be done when time and funds permit; but, there a re some situations in which it is not feasible to use all methods. Still, it is C o m b i n i n g p a rticularly desirable to pre-field and meld the objective with the subjective methods field methods the respondent centere d p rovides an with the interv i e w e r- c e n t e red. This evaluation of complementarity allows b road scope. for both good pro b l e m identification and pro b l e m resolution and provides an evaluation of b road scope. 10 11

Wh e re Can I Get More Info rm a t i o n? In addition to the pamphlets in this series, ASA also makes other bro c h u res available upon re q u e s t : Ethical Guidelines for Statistical Practice S u rveys and Privacy, p roduced by the ASA Committee on Privacy and Confidentiality. For the above bro c h u res or other pamphlets in the What Is a Surv e y? series, contact: Section on Survey Research Methods American Statistical Association 1429 Duke Stre e t Alexandria, VA 22314-3415 USA (703) 684-1221/fax: (703) 684-2037 Email: asainfo@amstat.org Web site: http://www. a m s t a t. o rg / s e c t i o n s / s rm s / Besides the ASA, there are many other associations that are concerned with the p roper collection and use of survey data: The American Association for Public Opinion Researc h (AAPOR) offers a number of publications perhaps the most re l e v a n t of these is the one entitled Best Practices for S u rvey a n d Public Opinion Research Surv e y Practices AAPOR Condemns. To obtain copies, call (313) 764-1555 or visit their We b site at http://www. a a p o r. o rg. The National Council on Public Polls publishes another useful pamphlet, Tw e n t y Questions a Journalist Should Ask About Poll R e s u l t s. To obtain a copy, call (800) 239-0909. The Research Industry Coalition, Inc., publishes a bro c h u re, Integrity and Good Practice in Marketing and Opinion Researc h. To obtain a copy, call (516) 928-6803. The Council of American Survey Researc h O rg a n i z a t i o n s publishes a pamphlet, S u rveys and Yo u. To obtain a copy, call (516) 928-6954, or email their Web site at h t t p : / / w w w. c a s ro. o rg. This pamphlet is based on a Census Bureau publication entitled P retesting Policy and Options: Demographic S u rveys at the Census Bureau. U.S. Census Bureau, 1993, submitted by T h e resa DeMaio. Information on cost and suggestions on the timing of pretesting can be found in this Census re p o rt. For suggestions about this pamphlet or potential future topics in the What Is a Survey? series, contact Fritz Scheuren, overall series editor and coordinator, at The Urban Institute, Washington, D.C. (scheuren@aol.com). This pamphlet, How to Conduct Pretesting, was prepared under the general direction of Charles Proctor, 1997-98 Publications Officer, ASA Section on Survey Research Methods. 12