/
Curb-stoning, a Too Neglected and Very Embarrassing Survey Curb-stoning, a Too Neglected and Very Embarrassing Survey

Curb-stoning, a Too Neglected and Very Embarrassing Survey - PowerPoint Presentation

pasty-toler
pasty-toler . @pasty-toler
Follow
385 views
Uploaded On 2017-10-07

Curb-stoning, a Too Neglected and Very Embarrassing Survey - PPT Presentation

Comments Jaki S McCarthy Senior Cognitive Research Methodologist US Department of Agriculture National Agricultural Statistics Service WSS Seminar December 2 2014 Another perspective on interviewer falsification ID: 593857

indicators data falsification interviewers data indicators interviewers falsification collect quality motivation rates accurate relevant detection extrinsic work falsify interviews

Share:

Link:

Embed:

Download Presentation from below link

Download Presentation The PPT/PDF document "Curb-stoning, a Too Neglected and Very E..." is the property of its rightful owner. Permission is granted to download and print the materials on this web site for personal, non-commercial use only, and to display it on your personal computer provided you do not modify the materials and that you retain all copyright notices contained in the materials. By downloading content from our website, you accept the terms of this agreement.


Presentation Transcript

Slide1

Curb-stoning, a Too Neglected and Very Embarrassing Survey ProblemComments

Jaki S. McCarthy

Senior Cognitive Research Methodologist

US Department of Agriculture

National Agricultural Statistics Service

WSS Seminar, December 2, 2014Slide2

Another perspective on interviewer falsificationSlide3

Why do interviewers falsify?Understanding why can:

Help identify relevant measures for models

Help

develop strategies to prevent falsificationSlide4

Falsifying to max $/min effortMost indicators have this underlying assumption (i.e. easy answers, shorter answers, rounding, etc.)

Can we use data mining to identify other (less obvious) indicators?Slide5

Falsifying to Meet DeadlinesDo indicators change?Maybe completion dates are important hereSlide6

Other reasons to falsify?Deliberate fabrication/data misrepresentationFatigue

Perceived reduction in respondent burden

Would indicators be the same?Slide7

Do these inform potential indicators/falsification model inputs?

Speed indicators (length of interviews, completed interviews/day, etc.)

Item

nonresponse

rates

Edit rates

Contact histories

GPS trackingSlide8

How do we prevent falsification?How do we change motivation?

Intrinsic versus extrinsic motivation

Employee (and respondent) engagementSlide9

Extrinsic Motivation

Should we pay interviewers more?

How much do you have to pay to ensure data won’t be falsified?

“Because

of what they are paying me

, I‘m going to collect the most accurate data possible.”Slide10

Extrinsic MotivationInterviewers may falsify because they think no one is checking and it doesn’t matter

Knowing that QA procedures are in place, and work is monitored can help here

“Because

I know they are checking my work,

I’m going to collect the most accurate data possible.”Slide11

Intrinsic MotivationHow else can we motivate interviewers?“Because _____________________________, I’m going to collect the most accurate data possible.” Slide12

How to get interviewers invested in the processMinimize Us versus Them

Supervisors/Monitors versus interviewers

HQ versus field

Data collectors versus data providers

Value of the agency

Value of the work

“Because __________________, I’m going to collect the most accurate data possible.”Slide13

This extends to respondents too!Many of the indicators would flag poor quality data provided by

respondents

Why

do respondents want to provide good quality data?

How can we improve the quality of respondents’ inputs

?

Will INTs who are good at gaining cooperation (i.e. getting cooperation from “hard to reach” units) look like they are collecting lower quality data?Slide14

What can we do to get the right answer in that blank?Need to invest in

Training

Employee engagement

Communication up and down the

chain

Ultimate goal is to have only

unintentional

errors to detectSlide15

Comments on Winker’s paperSlide16

Curb-stoning as Fraud Detection ProblemAdvantages to this approach?

Why not a classification problem?

Why not score interviewers using an index of indicators?

How about scoring interviews and verifying cases (not interviewers), or following up interviewers with highest percent of suspicious records?

Is this an outlier detection problem?Slide17

Objective way to narrow focusHow to target scarce resourcesBut as in other fraud detection problems, likely doesn’t go far enough (i.e. need to detect at much lower rates than 20% falsifiers with 70% falsification rate)Slide18

How can this method be extended?Are there other variables beyond indicators that might be useful in classifying falsifiers?

Data relevant indicators (time stamps, edit rates)

Person relevant indicators (INT characteristics – Yes, I realize we are getting into dicey territory!)

This would require “real” data – i.e. cannot be done with simulated data