On Measurement Validity and Language Models: Increasing Validity and Decreasing Bias with Instructions

Moritz Laurer*, Wouter van Atteveldt, Andreu Casas, Kasper Welbers

*Corresponding author for this work

Research output: Contribution to JournalArticleAcademicpeer-review

Abstract

Language models like BERT or GPT are becoming increasingly popular measurement tools, but are the measurements they produce valid? Literature suggests that there is still a relevant gap between the ambitions of computational text analysis methods and the validity of their outputs. One prominent threat to validity is hidden biases in the training data, where models learn group-specific language patterns instead of the concept researchers want to measure. This paper investigates to what extent these biases impact the validity of measurements created with language models. We conduct a comparative analysis across nine group types in four datasets with three types of classification models, focusing on the robustness of models against biases and on the validity of their outputs. While we find that all types of models learn biases, the effects on validity are surprisingly small. In particular when models receive instructions as an additional input, they become more robust against biases from the fine-tuning data and produce more valid measurements across different groups. An instruction-based model (BERT-NLI) sees its average test-set performance decrease by only 0.4% F1 macro when trained on biased data and its error probability on groups it has not seen during training increases only by 0.8%.

Original languageEnglish
Pages (from-to)46-62
Number of pages17
JournalCommunication Methods and Measures
Volume19
Issue number1
Early online date29 Jul 2024
DOIs
Publication statusPublished - 2025

Bibliographical note

Publisher Copyright:
© 2024 The Author(s). Published with license by Taylor & Francis Group, LLC.

Funding

This work was supported by a Ph.D. scholarship of the Heinrich B\u00F6ll Foundation and with a Snellius compute grant [EINF-3006] by the Dutch Research Council (NWO). We would like to thank our colleagues at the Vrije Universiteit Amsterdam PolCom Group, the participants of the COMPTEXT 2023 conference, and Arthur Spirling and Camille Borrett for their constructive feedback on earlier versions of the manuscript.

FundersFunder number
Nederlandse Organisatie voor Wetenschappelijk Onderzoek
Heinrich Böll StiftungEINF-3006

    Cite this