论文标题
语言模型显示了类似人类的内容对推理任务的影响
Language models show human-like content effects on reasoning tasks
论文作者
论文摘要
推理是智能系统的关键能力。大型语言模型(LMS)在抽象的推理任务上实现了高度的性能,但表现出许多缺陷。但是,人类抽象推理也不完美。例如,人类的推理受我们的现实知识和信念的影响,并显示出显着的“内容效应”。当问题的语义内容支持正确的逻辑推断时,人类会更可靠地理解。这些内容输入的推理模式在有关人类智能的基本本质的辩论中起着核心作用。在这里,我们调查了语言模型$ \ unicode {x2014} $是否以前的期望捕获了人类知识的某些方面$ \ unicode {x2014} $类似地将内容混合到他们对逻辑问题的答案中。我们在三个逻辑推理任务中探讨了这个问题:自然语言推论,判断三段论的逻辑有效性和ison选择任务。我们评估了最新的大语言模型以及人类的状态,并发现语言模型反映了在这些任务中人类在人类中观察到的许多相同模式,$ \ unicode {x2014} $像人类一样,当任务的语义内容支持逻辑推断时,模型会更准确地回答。这些相似之处既反映在答案模式中,又有较低级别的特征,例如模型答案分布与人类响应时间之间的关系。我们的发现对了解人类的这些认知效应以及有助于语言模型表现的因素具有影响。
Reasoning is a key ability for an intelligent system. Large language models (LMs) achieve above-chance performance on abstract reasoning tasks, but exhibit many imperfections. However, human abstract reasoning is also imperfect. For example, human reasoning is affected by our real-world knowledge and beliefs, and shows notable "content effects"; humans reason more reliably when the semantic content of a problem supports the correct logical inferences. These content-entangled reasoning patterns play a central role in debates about the fundamental nature of human intelligence. Here, we investigate whether language models $\unicode{x2014}$ whose prior expectations capture some aspects of human knowledge $\unicode{x2014}$ similarly mix content into their answers to logical problems. We explored this question across three logical reasoning tasks: natural language inference, judging the logical validity of syllogisms, and the Wason selection task. We evaluate state of the art large language models, as well as humans, and find that the language models reflect many of the same patterns observed in humans across these tasks $\unicode{x2014}$ like humans, models answer more accurately when the semantic content of a task supports the logical inferences. These parallels are reflected both in answer patterns, and in lower-level features like the relationship between model answer distributions and human response times. Our findings have implications for understanding both these cognitive effects in humans, and the factors that contribute to language model performance.