Random Language Model: a path to principled complexity

09/04/2018
by   E. DeGiuli, et al.
0

Many complex generative systems use languages to create structured objects. We consider a model of random languages, defined by weighted context-free grammars. As the distribution of grammar weights broadens, a transition is found from a random phase, in which sentences are indistinguishable from noise, to an organized phase in which nontrivial information is carried. This marks the emergence of deep structure in the language, and can be understood by a competition between energy and entropy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/20/2019

Emergence of order in random languages

We consider languages generated by weighted context-free grammars. It is...
research
03/10/2016

Zipf's law emerges asymptotically during phase transitions in communicative systems

Zipf's law predicts a power-law relationship between word rank and frequ...
research
03/21/2023

A Tale of Two Circuits: Grokking as Competition of Sparse and Dense Subnetworks

Grokking is a phenomenon where a model trained on an algorithmic task fi...
research
06/26/2020

Dissecting Power of a Finite Intersection of Context Free Languages

Let ^k,α denote a tetration function defined as follows: ^1,α=2^α and ^k...
research
02/11/2013

Toric grammars: a new statistical approach to natural language modeling

We propose a new statistical model for computational linguistics. Rather...
research
05/22/2023

GPT-SW3: An Autoregressive Language Model for the Nordic Languages

This paper details the process of developing the first native large gene...
research
11/04/2021

How Do Neural Sequence Models Generalize? Local and Global Context Cues for Out-of-Distribution Prediction

After a neural sequence model encounters an unexpected token, can its be...

Please sign up or login with your details

Forgot password? Click here to reset