Large language models have an exceptional capability to incorporate new
...
We examine how transformers cope with two challenges: learning basic int...
The standard methodology of evaluating large language models (LLMs) base...
Token embeddings, a mapping from discrete lexical symbols to continuous
...
Premise selection is a fundamental problem of automated theorem proving....
Designing networks capable of attaining better performance with an incre...
Language models (LMs) are becoming the foundation for almost all major
l...
The formalization of existing mathematical proofs is a notoriously diffi...
Prompted models have demonstrated impressive few-shot learning abilities...
The ability to extrapolate from short problem instances to longer ones i...
Language models have achieved remarkable performance on a wide range of ...
Pre-training produces representations that are effective for a wide rang...
Complex reasoning problems contain states that vary in the computational...
Autoformalization is the process of automatically translating from natur...
In theorem proving, the task of selecting useful premises from a large
l...
Generating step-by-step "chain-of-thought" rationales improves language ...
Language models typically need to be trained or finetuned in order to ac...
We introduce the Block-Recurrent Transformer, which applies a transforme...
Transformer models yield impressive results on many NLP and sequence mod...
Our ability to know when to trust the decisions made by machine learning...
Humans excel in solving complex reasoning tasks through a mental process...
Due to spurious correlations, machine learning systems often fail to
gen...
Labeled data for imitation learning of theorem proving in large librarie...
While designing inductive bias in neural architectures has been widely
s...
In this work, we focus on an analogical reasoning task that contains ric...
Propositional model counting or #SAT is the problem of computing the num...
In learning-assisted theorem proving, one of the most critical challenge...
Mathematical proofs can be mechanised using proof assistants to eliminat...
We propose a novel hierarchical agent architecture for multi-agent
reinf...
State-of-the-art meta reinforcement learning algorithms typically assume...
Sparse reward is one of the most challenging problems in reinforcement
l...
Careful tuning of the learning rate, or even schedules thereof, can be
c...
We consider the problem of exploration in meta reinforcement learning. T...
In this technical report, we consider an approach that combines the PPO
...
In this work, we propose to apply trust region optimization to deep
rein...
We propose a simple and general variant of the standard reparameterized
...
The past several years have seen remarkable progress in generative model...
We investigate the parameter-space geometry of recurrent neural networks...
In this paper, we systematically analyze the connecting architectures of...
We introduce a weight update formula that is expressed only in terms of
...