From Word Models to World Models: Translating from Natural Language to the Probabilistic Language of Thought

06/22/2023
by   Lionel Wong, et al.
0

How does language inform our downstream thinking? In particular, how do humans make meaning from language–and how can we leverage a theory of linguistic meaning to build machines that think in more human-like ways? In this paper, we propose rational meaning construction, a computational framework for language-informed thinking that combines neural language models with probabilistic models for rational inference. We frame linguistic meaning as a context-sensitive mapping from natural language into a probabilistic language of thought (PLoT)–a general-purpose symbolic substrate for generative world modeling. Our architecture integrates two computational tools that have not previously come together: we model thinking with probabilistic programs, an expressive representation for commonsense reasoning; and we model meaning construction with large language models (LLMs), which support broad-coverage translation from natural language utterances to code expressions in a probabilistic programming language. We illustrate our framework through examples covering four core domains from cognitive science: probabilistic reasoning, logical and relational reasoning, visual and physical reasoning, and social reasoning. In each, we show that LLMs can generate context-sensitive translations that capture pragmatically-appropriate linguistic meanings, while Bayesian inference with the generated programs supports coherent and robust commonsense reasoning. We extend our framework to integrate cognitively-motivated symbolic modules (physics simulators, graphics engines, and planning algorithms) to provide a unified commonsense thinking interface from language. Finally, we explore how language can drive the construction of world models themselves. We hope this work will provide a roadmap towards cognitive models and AI systems that synthesize the insights of both modern and classical computational perspectives.

READ FULL TEXT

page 4

page 13

page 17

page 25

page 27

page 29

page 34

page 40

research
05/01/2022

MRKL Systems: A modular, neuro-symbolic architecture that combines large language models, external knowledge sources and discrete reasoning

Huge language models (LMs) have ushered in a new era for AI, serving as ...
research
10/30/2021

Automatic Knowledge Augmentation for Generative Commonsense Reasoning

Generative commonsense reasoning is the capability of a language model t...
research
12/15/2022

Improving Chess Commentaries by Combining Language Models with Symbolic Reasoning Engines

Despite many recent advancements in language modeling, state-of-the-art ...
research
01/16/2023

Dissociating language and thought in large language models: a cognitive perspective

Today's large language models (LLMs) routinely generate coherent, gramma...
research
03/13/2013

A Symbolic Approach to Reasoning with Linguistic Quantifiers

This paper investigates the possibility of performing automated reasonin...
research
02/03/2021

A Computational Framework for Slang Generation

Slang is a common type of informal language, but its flexible nature and...
research
06/11/2023

A blind spot for large language models: Supradiegetic linguistic information

Large Language Models (LLMs) like ChatGPT reflect profound changes in th...

Please sign up or login with your details

Forgot password? Click here to reset