Interpretability of an Interaction Network for identifying H → b(Equation presented) jets

Research output: Contribution to journalConference articlepeer-review

Abstract

Multivariate techniques and machine learning models have found numerous applications in High Energy Physics (HEP) research over many years. In recent times, AI models based on deep neural networks are becoming increasingly popular for many of these applications. However, neural networks are regarded as black boxes- because of their high degree of complexity it is often quite difficult to quantitatively explain the output of a neural network by establishing a tractable input-output relationship and information propagation through the deep network layers. As explainable AI (xAI) methods are becoming more popular in recent years, we explore interpretability of AI models by examining an Interaction Network (IN) model designed to identify boosted H → b(Equation presented) jets amid QCD background. We explore different quantitative methods to demonstrate how the classifier network makes its decision based on the inputs and how this information can be harnessed to reoptimize the model- making it simpler yet equally effective. We additionally illustrate the activity of hidden layers within the IN model as Neural Activation Pattern (NAP) diagrams. Our experiments suggest NAP diagrams reveal important information about how information is conveyed across the hidden layers of deep model. These insights can be useful to effective model reoptimization and hyperparameter tuning.

Original languageEnglish (US)
Article number223
JournalProceedings of Science
Volume414
StatePublished - 2022
Externally publishedYes
Event41st International Conference on High Energy Physics, ICHEP 2022 - Bologna, Italy
Duration: Jul 6 2022Jul 13 2022

ASJC Scopus subject areas

  • General

Fingerprint

Dive into the research topics of 'Interpretability of an Interaction Network for identifying H → b(Equation presented) jets'. Together they form a unique fingerprint.

Cite this