Dissociating language and thought in large language models: a cognitive perspective

01/16/2023
by   Kyle Mahowald, et al.
5

Today's large language models (LLMs) routinely generate coherent, grammatical and seemingly meaningful paragraphs of text. This achievement has led to speculation that these networks are – or will soon become – "thinking machines", capable of performing tasks that require abstract knowledge and reasoning. Here, we review the capabilities of LLMs by considering their performance on two different aspects of language use: 'formal linguistic competence', which includes knowledge of rules and patterns of a given language, and 'functional linguistic competence', a host of cognitive abilities required for language understanding and use in the real world. Drawing on evidence from cognitive neuroscience, we show that formal competence in humans relies on specialized language processing mechanisms, whereas functional competence recruits multiple extralinguistic capacities that comprise human thought, such as formal reasoning, world knowledge, situation modeling, and social cognition. In line with this distinction, LLMs show impressive (although imperfect) performance on tasks requiring formal linguistic competence, but fail on many tests requiring functional competence. Based on this evidence, we argue that (1) contemporary LLMs should be taken seriously as models of formal linguistic skills; (2) models that master real-life language use would need to incorporate or develop not only a core language module, but also multiple non-language-specific cognitive capacities required for modeling thought. Overall, a distinction between formal and functional linguistic competence helps clarify the discourse surrounding LLMs' potential and provides a path toward building models that understand and use language in human-like ways.

READ FULL TEXT
research
08/29/2023

Large Language Models on the Chessboard: A Study on ChatGPT's Formal Language Comprehension and Complex Reasoning Skills

While large language models have made strides in natural language proces...
research
07/20/2023

LLM Cognitive Judgements Differ From Human

Large Language Models (LLMs) have lately been on the spotlight of resear...
research
06/22/2023

From Word Models to World Models: Translating from Natural Language to the Probabilistic Language of Thought

How does language inform our downstream thinking? In particular, how do ...
research
06/11/2023

A blind spot for large language models: Supradiegetic linguistic information

Large Language Models (LLMs) like ChatGPT reflect profound changes in th...
research
02/10/2017

Modeling Semantic Expectation: Using Script Knowledge for Referent Prediction

Recent research in psycholinguistics has provided increasing evidence th...
research
07/01/2016

Throwing fuel on the embers: Probability or Dichotomy, Cognitive or Linguistic?

Prof. Robert Berwick's abstract for his forthcoming invited talk at the ...
research
07/11/2023

Unleashing Cognitive Synergy in Large Language Models: A Task-Solving Agent through Multi-Persona Self-Collaboration

Human intelligence thrives on the concept of cognitive synergy, where co...

Please sign up or login with your details

Forgot password? Click here to reset