Skip to main content
eScholarship
Open Access Publications from the University of California

Graded grammatical expectations in transformer models

Abstract

Large language models (LLMs) can be reasonably thought of as models of idealized statistical learners. Thus, the extent to which they grasp the grammar of the language they are trained on suggests how much of it can be learned from memorization, abstraction, and generalization of linguistic input. However, the knowledge of LMMs’ grammar has largely been gleaned from examples of their outputs or datasets not designed to assess how native-like its knowledge is. In this study, we probed the knowledge of an LLM, GPT-3, with a graded grammatical acceptability task previously normed on humans. GPT-3’s ratings were correlated with human ratings, even with minimal examples. Moreover, GPT-3’s deviation from the human norms was predicted by the between-subject variation for each item, and these deviations were rarely outside of the range of human ratings. Follow-up analyses tested the extent to which local probabilistic structure drives these judgments using n-gram models.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View