DeepAI AI Chat
Log In Sign Up

The Effect of Efficient Messaging and Input Variability on Neural-Agent Iterated Language Learning

04/15/2021
by   Yuchen Lian, et al.
0

Natural languages commonly display a trade-off among different strategies to convey constituent roles. A similar trade-off, however, has not been observed in recent simulations of iterated language learning with neural network based agents (Chaabouni et al., 2019b). In this work, we re-evaluate this result in the light of two important factors, namely: the lack of effort-based pressure in the agents and the lack of variability in the initial input language.

READ FULL TEXT

page 1

page 2

page 3

page 4

09/15/2015

Dependency length minimization: Puzzles and Promises

In the recent issue of PNAS, Futrell et al. claims that their study of 3...
04/08/2020

Internal and external pressures on language emergence: least effort, object constancy and frequency

In previous work, artificial agents were shown to achieve almost perfect...
09/28/2021

On Homophony and Rényi Entropy

Homophony's widespread presence in natural languages is a controversial ...
12/14/2020

Federated Learning under Importance Sampling

Federated learning encapsulates distributed learning strategies that are...
04/20/2020

Compositionality and Generalization in Emergent Languages

Natural language allows us to refer to novel composite concepts by combi...
03/08/2021

Nondeterminism and Instability in Neural Network Optimization

Nondeterminism in neural network optimization produces uncertainty in pe...