logo
today local_bar
Poster Session 3 · Thursday, December 4, 2025 11:00 AM → 2:00 PM
#1115

Additive Models Explained: A Computational Complexity Approach

NeurIPS OpenReview

Abstract

Generalized Additive Models (GAMs) are commonly considered interpretable within the ML community, as their structure makes the relationship between inputs and outputs relatively understandable. Therefore, it may seem natural to hypothesize that obtaining meaningful explanations for GAMs could be performed efficiently and would not be computationally infeasible.
In this work, we challenge this hypothesis by analyzing the computational complexity of generating different explanations for various forms of GAMs across multiple contexts. Our analysis reveals a surprisingly diverse landscape of both positive and negative complexity outcomes.
Particularly, under standard complexity assumptions such as PNP, we establish several key findings:
  1. in stark contrast to many other common ML models, the complexity of generating explanations for GAMs is heavily influenced by the structure of the input space;
  2. the complexity of explaining GAMs varies significantly with the types of component models used - but interestingly, these differences only emerge under specific input domain settings;
  3. significant complexity distinctions appear for obtaining explanations in regression tasks versus classification tasks in GAMs; and
  4. expressing complex models like neural networks additively (e.g., as neural additive models) can make them easier to explain, though interestingly, this benefit appears only for certain explanation methods and input domains.
Collectively, these results shed light on the feasibility of computing diverse explanations for GAMs, offering a rigorous theoretical picture of the conditions under which such computations are possible or provably hard.