Back to issue
Cover of: How to Use Large Language Models for Empirical Legal Research
Jonathan H. Choi

How to Use Large Language Models for Empirical Legal Research

Section: Conference Article 1
Volume 180 (2024) / Issue 2, pp. 214-233 (20)
Published 16.07.2024
DOI 10.1628/jite-2024-0006
  • article PDF
  • available
  • 10.1628/jite-2024-0006
Summary
Legal scholars have long annotated cases by hand to summarize and learn about developments in jurisprudence. Dramatic recent improvements in the performance of large language models (LLMs) now provide a potential alternative. This article demonstrates how to use LLMs to analyze legal documents. It evaluates best practices and suggests both the uses and potential limitations of LLMs in empirical legal research. In a simple classification task involving Supreme Court opinions, it finds that GPT-4 performs approximately as well as human coders and significantly better than a variety of prior-generation natural language processing (NLP) classifiers, with no improvement from supervised training, finetuning, or specialized prompting.