Interpretability of an Interaction Network for identifying H → bb̅ jets

11/23/2022
by   Avik Roy, et al.
0

Multivariate techniques and machine learning models have found numerous applications in High Energy Physics (HEP) research over many years. In recent times, AI models based on deep neural networks are becoming increasingly popular for many of these applications. However, neural networks are regarded as black boxes – because of their high degree of complexity it is often quite difficult to quantitatively explain the output of a neural network by establishing a tractable input-output relationship and information propagation through the deep network layers. As explainable AI (xAI) methods are becoming more popular in recent years, we explore interpretability of AI models by examining an Interaction Network (IN) model designed to identify boosted H→ bb̅ jets amid QCD background. We explore different quantitative methods to demonstrate how the classifier network makes its decision based on the inputs and how this information can be harnessed to reoptimize the model-making it simpler yet equally effective. We additionally illustrate the activity of hidden layers within the IN model as Neural Activation Pattern (NAP) diagrams. Our experiments suggest NAP diagrams reveal important information about how information is conveyed across the hidden layers of deep model. These insights can be useful to effective model reoptimization and hyperparameter tuning.

READ FULL TEXT

page 1

page 5

research
10/09/2022

A Detailed Study of Interpretability of Deep Neural Network based Top Taggers

Recent developments in the methods of explainable AI (xAI) methods allow...
research
06/14/2022

Explainable AI for High Energy Physics

Neural Networks are ubiquitous in high energy physics research. However,...
research
04/26/2020

ExpDNN: Explainable Deep Neural Network

In recent years, deep neural networks have been applied to obtain high p...
research
06/05/2018

Explainable Neural Networks based on Additive Index Models

Machine Learning algorithms are increasingly being used in recent years ...
research
02/20/2023

SpecXAI – Spectral interpretability of Deep Learning Models

Deep learning is becoming increasingly adopted in business and industry ...
research
02/05/2021

Interpretable Neural Networks based classifiers for categorical inputs

Because of the pervasive usage of Neural Networks in human sensitive app...
research
10/19/2020

A Framework to Learn with Interpretation

With increasingly widespread use of deep neural networks in critical dec...

Please sign up or login with your details

Forgot password? Click here to reset