Conference Agenda

Overview and details of the sessions of this conference. Please select a date or location to show only sessions at that day or location. Please select a single session for detailed view (with abstracts and downloads if available).

 
 
Session Overview
Session
A4.1: Data Quality in Online Surveys
Time:
Thursday, 21/Sept/2023:
5:00pm - 6:00pm

Session Chair: Tobias Rettig, University of Mannehim, Germany
Location: Lecture Hall 3, Room 1135

Kassel University Campus Center Campus Holländischer Platz Moritzstraße 18 34127 Kassel Germany

Show help for 'Increase or decrease the abstract text size'
Presentations

Data Quality Indicators: Some Practical Recommendation

Nivedita Bhaktha, Henning Silber, Clemens Lechner

GESIS, Germany

Relevance & Research Question
It is very common in survey research to collect data on multi-item scales or constructs with Likert or categorical response options. Low quality data (LQD) in such samples has been a well discussed issue but is gaining further traction due to the rise in online data collection. Low quality data can be defined as those responses where the respondents have not put in sufficient thought and effort into responding. Consequently, there has been a surge in literature on methods for identifying LQD, however, the methods and best practices are scattered across discplines. In this study, we aim to demonstrate the use of data quality indicators used across different fields such as psychology, sociology, epidemiology, and public health on multi-item scales. We will discuss various indicators used, the correlation among them, and the best practices associated with flagging low quality data.
Methods & Data
We use publicly available German Longitudinal Election Study cross section wave 21, pre-election data for the demonstration. We use functions and approaches provided in the dataquieR and careless packages in R for data quality analysis. Different rules of thumb and thresholds will be applied to the data quality indicators for flagging problematic data.
Results
We observe that many data quality indicators should be used to flag low quality data. Different indicators pinpoint different underlying reasons for problematic data. The correlation among different indicators are typically low. Different thresholds and cut-offs for the indicators flag different number of observations as problematic responses.
Added Value
We have unified data quality practices from different fields. We have demonstrated the use of data quality indicators in identifying low quality data in multi-item largescale survey data. We have taken up the discussion on setting thresholds for the indicators to flag low quality data. Finally, we provide recommendations on combining different indicators to flag potentially problematic data.



Fielding a long online survey: Evidence from the first Generations and Gender Survey (GGS) in the UK

Olga Maslovskaya, Grace Chang, Brienna Perelli-Harris

University of Southampton, United Kingdom

Relevance & Research Question

Our team has collected the first Generations and Gender Survey (GGS) in the UK. This survey used push-to-web design with online-only mode available for respondents. The approximate length of time for survey completion was specified as being around 50 minutes for the respondents.

The length of the online surveys topic has recently attracted a lot of attention from survey methodologists as many high quality social surveys moved to online data collection or mixed-mode designs in the recent years. The rule of thumb until recently was not to have online surveys of the length exciding 10-20 minutes. Recent experiments conducted by the European Social Survey (ESS) suggest that it is possible to conduct longer (35 minutes or even 55 minutes) online surveys without significant reduction in data quality. However, more evidence is needed to establish the optimal length of online social surveys.

Methods & Data

In this paper, we will present the evidence for fielding a long online probability-based survey. We reflect on the challenges and opportunities of conducting a long probability-based online data collection in the UK by reporting on nonresponse, break-off rates, and quality of responses. We also investigate the de-briefing questions in which respondents were able to reflect on how they felt about the survey. We will compare paradata for the length of time it took respondents to complete the survey with the respondents’ perception on how long the survey was when these paradata become available in April 2023.

Results

Preliminary results suggest that despite the fact that the UK GGS survey was long and complex, 82% of respondents found the survey “not at all difficult”. High proportion of respondents felt that the survey was about as long as they expected (47%) with further 19% felt that the survey was shorter than they expected. Also, another positive outcome of the survey was that 82% of participants gave consent to be recontacted for the second wave of the UK GGS survey.

Added Value

Our findings provide evidence for the optimal length for long and complex online social surveys and have important implications for survey practice.



Ability to identify fakers in online surveys: Comparison of BIDR.Short.24 and MCSD-SF

Vaka Vésteinsdóttir, Ingunn Ros Kristjansdottir, Katrin S. J. Steingrimsdottir

University of Iceland, Iceland

Relevance & Research Question: Socially desirable responding (SDR) is a common problem in self-report measures, as the tendency to present oneself favorably to others can influence the honesty of responders. One facet of SDR is faking, an intentional misrepresentation in self-report. There are, however, two kinds of faking; faking good and faking bad. Faking good involves deliberately presenting oneself favorably to others, whereas faking bad involves deliberately presenting oneself in an undesirable manner. There are several ways to detect faking, one of them being SDR scales. Two of the most widely used scales are the Marlowe-Crowne Social Desirability Scale (MCSDS) and the Balanced Inventory of Desirable Responding (BIDR). To evaluate which SDR measure is better suited to detect SDR, one can compare their ability to detect faking. A previous study comparing the ability of MCSDS and BIDR to detect faking found that MCSDS outperformed BIDR in identifying both types of faking. A limitation of the applicability of those results is the fact that the comparison was between full-length versions of the measures but short-form versions are usually more practical as they reduce response fatigue. For that reason, the current study compared two short-forms’ ability to detect faking, MCSD-SF (short-form version of MCSDS) and IM-Short.24 (short-form of the IM subscale of BIDR).

Methods & Data: Participants were recruited online through a probability-based panel. The final sample consisted of 106 men and 122 women, others chose not to answer the gender question. Participants were randomly assigned to one of three groups: 1) standard instructions, 2) fake good instructions, or 3) fake bad instructions and then asked to complete both the MCSD-SF and the IM-Short.24.

Results: Discriminant function analyses and receiver operating characteristic curve analyses showed that, overall, MCSD-SF outperformed IM-Short.24 in identifying faking good, while IM-Short.24 outperformed MCSD-SF in identifying faking bad.

Added Value: These findings show a clear preference for the use of MCSD-SF for identifying fake good responses and IM-Short.24 for identifying fake bad responses, which should assist researchers in choosing an appropriate measure for their studies, as well as advance the use of SDR measures overall.



 
Contact and Legal Notice · Contact Address:
Privacy Statement · Conference: GOR 23
Conference Software: ConfTool Pro 2.8.101
© 2001–2024 by Dr. H. Weinreich, Hamburg, Germany