AI and Research Methods

05 May 2026, Version 1
This content is an early or alternative research output and has not been peer-reviewed at the time of posting.

Abstract

This chapter surveys how large language models (LLMs) are being used in political science research and how they are reshaping established methods. It is organized around six applications: annotation and measurement, experimental treatment generation, silicon sampling, generative agent-based modeling, tool-augmented data collection, and the study of LLMs as objects of inquiry. The chapter traces the methodological lineages from which these approaches emerge and examines their implications for validity, inference, bias, and reproducibility. A central argument is that LLMs require researchers to rethink standards. As these models match or exceed human performance in some tasks, traditional human-coded gold standards become less secure, increasing the importance of construct validity and robustness. At the same time, because LLM-based workflows are sensitive to prompts, versions, and parameters, exact reproducibility is often unattainable. The chapter therefore discusses some of the new – and often still open – questions relating to replication and transparency with LLMs.

Keywords

AI
Methods

Comments

Comments are not moderated before they are posted, but they can be removed by the site moderators if they are found to be in contravention of our Commenting Policy [opens in a new tab] - please read this policy before you post. Comments should be used for scholarly discussion of the content in question. You can find more information about how to use the commenting feature here [opens in a new tab] .
This site is protected by reCAPTCHA and the Google Privacy Policy [opens in a new tab] and Terms of Service [opens in a new tab] apply.