Precise Multi-Neuron Abstractions for Neural Network Certification

03/05/2021
by   Mark Niklas Müller, et al.
0

Formal verification of neural networks is critical for their safe adoption in real-world applications. However, designing a verifier which can handle realistic networks in a precise manner remains an open and difficult challenge. In this paper, we take a major step in addressing this challenge and present a new framework, called PRIMA, that computes precise convex approximations of arbitrary non-linear activations. PRIMA is based on novel approximation algorithms that compute the convex hull of polytopes, leveraging concepts from computational geometry. The algorithms have polynomial complexity, yield fewer constraints, and minimize precision loss. We evaluate the effectiveness of PRIMA on challenging neural networks with ReLU, Sigmoid, and Tanh activations. Our results show that PRIMA is significantly more precise than the state-of-the-art, verifying robustness for up to 16 than prior work on ReLU-, Sigmoid-, and Tanh-based networks, respectively.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset