Generalizing Neural Additive Models via Statistical Multimodal Analysis

Young Kyung Kim, J. Matías Di Martino, Guillermo Sapiro

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Interpretable models are gaining increasing attention in the machine learning community, and significant progress is being made to develop simple, interpretable, yet powerful deep learning approaches. Generalized Additive Models (GAM) and Neural Additive Models (NAM) are prime examples. Despite these methods’ great potential and popularity in critical applications, e.g., medical applications, they fail to generalize to distributions with more than one mode (multimodal1). The main reason behind this limitation is that these "all-fit-one" models collapse multiple relationships by being forced to fit the data unimodally. We address this critical limitation by proposing interpretable multimodal network frameworks capable of learning a Mixture of Neural Additive Models (MNAM). The proposed MNAM learns relationships between input features and outputs in a multimodal fashion and assigns a probability to each mode. The proposed method shares similarities with Mixture Density Networks (MDN) while keeping the interpretability that characterizes GAM and NAM. We demonstrate how the proposed MNAM balances between rich representations and interpretability with numerous empirical observations and pedagogical studies. We present and discuss different training alternatives and provided extensive practical evaluation to assess the proposed framework. The code is available at https://github.com/youngkyungkim93/MNAM.

Original languageEnglish (US)
JournalTransactions on Machine Learning Research
Volume2024
StatePublished - 2024
Externally publishedYes

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence
  • Computer Vision and Pattern Recognition

Fingerprint

Dive into the research topics of 'Generalizing Neural Additive Models via Statistical Multimodal Analysis'. Together they form a unique fingerprint.

Cite this