Knowing When to Stop: Evaluation and Verification of Conformity to Output-size Specifications

04/26/2019
by   Chenglong Wang, et al.
6

Models such as Sequence-to-Sequence and Image-to-Sequence are widely used in real world applications. While the ability of these neural architectures to produce variable-length outputs makes them extremely effective for problems like Machine Translation and Image Captioning, it also leaves them vulnerable to failures of the form where the model produces outputs of undesirable length. This behavior can have severe consequences such as usage of increased computation and induce faults in downstream modules that expect outputs of a certain length. Motivated by the need to have a better understanding of the failures of these models, this paper proposes and studies the novel output-size modulation problem and makes two key technical contributions. First, to evaluate model robustness, we develop an easy-to-compute differentiable proxy objective that can be used with gradient-based algorithms to find output-lengthening inputs. Second and more importantly, we develop a verification approach that can formally verify whether a network always produces outputs within a certain length. Experimental results on Machine Translation and Image Captioning show that our output-lengthening approach can produce outputs that are 50 times longer than the input, while our verification approach can, given a model and input domain, prove that the output length is below a certain size.

READ FULL TEXT

page 7

page 8

research
07/02/2018

Dynamic Prediction Length for Time Series with Sequence to Sequence Networks

Recurrent neural networks and sequence to sequence models require a pred...
research
12/06/2018

Verification of deep probabilistic models

Probabilistic models are a critical part of the modern deep learning too...
research
11/01/2019

Sequence Modeling with Unconstrained Generation Order

The dominant approach to sequence generation is to produce a sequence in...
research
09/13/2016

Multimodal Attention for Neural Machine Translation

The attention mechanism is an important part of the neural machine trans...
research
08/24/2018

Approximate Distribution Matching for Sequence-to-Sequence Learning

Sequence-to-Sequence models were introduced to tackle many real-life pro...
research
11/28/2019

Multimodal Machine Translation through Visuals and Speech

Multimodal machine translation involves drawing information from more th...
research
03/27/2023

Graph Sequence Learning for Premise Selection

Premise selection is crucial for large theory reasoning as the sheer siz...

Please sign up or login with your details

Forgot password? Click here to reset