Overlap in meaning is a stronger predictor of semantic activation in GPT-3 than in humans
被引:0
|
作者:
Jan Digutsch
论文数: 0引用数: 0
h-index: 0
机构:Leibniz Research Centre for Working Environment and Human Factors at the Technical University of Dortmund,Institute of Behavioral Science and Technology
Jan Digutsch
Michal Kosinski
论文数: 0引用数: 0
h-index: 0
机构:Leibniz Research Centre for Working Environment and Human Factors at the Technical University of Dortmund,Institute of Behavioral Science and Technology
Michal Kosinski
机构:
[1] Leibniz Research Centre for Working Environment and Human Factors at the Technical University of Dortmund,Institute of Behavioral Science and Technology
Modern large language models generate texts that are virtually indistinguishable from those written by humans and achieve near-human performance in comprehension and reasoning tests. Yet, their complexity makes it difficult to explain and predict their functioning. We examined a state-of-the-art language model (GPT-3) using lexical decision tasks widely used to study the structure of semantic memory in humans. The results of four analyses showed that GPT-3’s patterns of semantic activation are broadly similar to those observed in humans, showing significantly higher semantic activation in related (e.g., “lime–lemon”) word pairs than in other-related (e.g., “sour–lemon”) or unrelated (e.g., “tourist–lemon”) word pairs. However, there are also significant differences between GPT-3 and humans. GPT-3’s semantic activation is better predicted by similarity in words’ meaning (i.e., semantic similarity) rather than their co-occurrence in the language (i.e., associative similarity). This suggests that GPT-3’s semantic network is organized around word meaning rather than their co-occurrence in text.
机构:
Tech Univ Dortmund, Leibniz Res Ctr Working Environm & Human Factors, Dortmund, Germany
Univ St Gallen, Inst Behav Sci & Technol, St Gallen, SwitzerlandTech Univ Dortmund, Leibniz Res Ctr Working Environm & Human Factors, Dortmund, Germany