pith. machine review for the scientific record. sign in

arxiv: 2605.05480 · v1 · submitted 2026-05-06 · 💻 cs.LG · cs.AI· stat.ML

Recognition: unknown

GRALIS: A Unified Canonical Framework for Linear Attribution Methods via Riesz Representation

Authors on Pith no claims yet

Pith reviewed 2026-05-08 17:00 UTC · model grok-4.3

classification 💻 cs.LG cs.AIstat.ML
keywords attribution methodsRiesz representationexplainable AISHAPintegrated gradientsLIMElinear functionalsunified framework
0
0 comments X

The pith

Every additive linear continuous attribution method on square-integrable functions has a unique canonical form via the Riesz theorem.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper develops GRALIS as a representation theory that places several popular neural network explanation techniques on identical mathematical footing. It proves that any attribution method which is additive, linear and continuous as a functional on an L2 space must take the specific form of a measure, a set of weights and a differencing operator. This common structure is required by the Riesz Representation Theorem and therefore applies simultaneously to SHAP, Integrated Gradients, LIME and linearized GradCAM. A reader would care because the single form supplies completeness, convergence rates and interaction-value guarantees that the separate methods previously lacked or could not all satisfy together.

Core claim

GRALIS establishes that every additive, linear, and continuous attribution functional on L^2(Q, μ) admits a unique canonical representation (Q, w, Δ) proved necessary by the Riesz Representation Theorem. This class encompasses SHAP, IG, LIME and linearized GradCAM. Seven theorems deliver exact completeness, Monte Carlo convergence O(1/sqrt(m))+O(1/k), exact Shapley interaction values, Hoeffding ANOVA decomposition, Sobol sensitivity generalization, and a multi-scale extension (MS-GRALIS) with minimum-variance weights.

What carries the argument

The Riesz Representation Theorem applied to linear continuous functionals on L^2(Q, μ), which forces every such attribution to be expressed by the triple (Q, w, Δ).

Load-bearing premise

The methods being unified must be additive, linear, and continuous functionals on the chosen L^2 space.

What would settle it

An attribution method that is additive, linear and continuous yet cannot be written in the (Q, w, Δ) form, or a listed method that violates one of the seven stated theorems.

Figures

Figures reproduced from arXiv: 2605.05480 by Raimondo Fanale.

Figure 1
Figure 1. Figure 1: Conditioned integration paths in GRALIS ( view at source ↗
Figure 2
Figure 2. Figure 2: GRALIS-MC pseudocode. The random permutation samples the view at source ↗
Figure 3
Figure 3. Figure 3: visualizes the reducibility hierarchy. GRALIS (canonical form) KernelSHAP Integrated Gradients LIME MS-GRALIS SHAP GradCAM πx ≡1, ∇F ≈ const. S =F \{i}, πx→δx lin. surrogate on F L levels uniform kernel 1 level cf. Tab. 5 6/14 5.5/14 view at source ↗
read the original abstract

The main XAI attribution methods for deep neural networks -- GradCAM, SHAP, LIME, Integrated Gradients -- operate on separate theoretical foundations and are not formally comparable. We present GRALIS (Gradient-Riesz Averaged Locally-Integrated Shapley), a mathematical framework establishing a representation theory for attributions: every additive, linear, and continuous attribution functional on L^2(Q,mu) admits a unique canonical representation (Q, w, Delta), proved necessary by the Riesz Representation Theorem. This class encompasses SHAP, IG, LIME and linearized GradCAM, but excludes nonlinear functionals such as standard GradCAM or attention maps. Seven formal theorems provide simultaneous guarantees absent in any individual method: (T1) necessary canonical form; (T2) exact completeness; (T3) Monte Carlo convergence O(1/sqrt(m))+O(1/k); (T4) exact Shapley Interaction Values; (T5) Hoeffding ANOVA decomposition; (T6) Sobol sensitivity generalization; (T7) multi-scale extension (MS-GRALIS) with minimum-variance weights. An algebraic appendix justifies the GRALIS-SIV correspondence via the Mobius transform without circularity. GRALIS satisfies 13.5/14 axiomatic properties vs. 2.5-6/14 for individual methods, including completeness, sensitivity, locality, order-k interactions and optimal multi-scale aggregation simultaneously. Preliminary validation on BreaKHis (1,187 histology images, DenseNet-121) reports deletion faithfulness AUC +0.015 (malignant), 96% class-conditional consistency, SAL = 0.762+/-0.109 and sparsity index 0.39. Extended comparison with baseline XAI methods is planned for a companion paper.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 2 minor

Summary. The paper introduces GRALIS, a framework based on the Riesz Representation Theorem asserting that every additive, linear, and continuous attribution functional on the Hilbert space L^2(Q, mu) admits a unique canonical representation (Q, w, Delta). It claims this class includes SHAP, Integrated Gradients, LIME, and linearized GradCAM (but excludes nonlinear methods such as standard GradCAM), and presents seven theorems guaranteeing necessary form (T1), exact completeness (T2), Monte Carlo convergence (T3), exact Shapley Interaction Values (T4), Hoeffding ANOVA decomposition (T5), Sobol sensitivity generalization (T6), and a multi-scale extension MS-GRALIS with minimum-variance weights (T7). An algebraic appendix justifies the GRALIS-SIV link via the Möbius transform without circularity. The framework is said to satisfy 13.5/14 axiomatic properties simultaneously. Preliminary numerical results on the BreaKHis dataset (1,187 images, DenseNet-121) report deletion faithfulness AUC improvements and other metrics.

Significance. If the central claims hold, the work would be significant for XAI by supplying a single representation theory that simultaneously delivers completeness, sensitivity, locality, order-k interactions, and optimal multi-scale aggregation—properties that no individual method satisfies together. The algebraic appendix addressing non-circularity and the parameter-free character inherited from the Riesz theorem are explicit strengths. The Monte Carlo rate O(1/sqrt(m)) + O(1/k) and the Hoeffding/Sobol connections would also be valuable if rigorously established for the listed methods.

major comments (1)
  1. [Abstract (statements of T1–T7) and algebraic appendix] The central claim that SHAP, IG, LIME, and linearized GradCAM belong to the class of continuous linear functionals on the entire L^2(Q, mu) is load-bearing for all seven theorems. Standard definitions of these methods are given only for a fixed model f and a specific input distribution; the manuscript must supply explicit embeddings or constructions showing that the resulting functionals remain linear and bounded (hence continuous) when extended to arbitrary square-integrable functions while exactly reproducing the original numerical values. Without this step, T1–T7 apply only to idealized surrogates rather than the published algorithms.
minor comments (2)
  1. [Empirical validation paragraph] The abstract reports only preliminary results on a single dataset and defers full baseline comparisons to a companion paper; the main text should include at least one additional dataset and direct numerical comparison with the original SHAP/IG/LIME implementations to support the claimed faithfulness gains.
  2. [Introduction / Theorem statements] Notation for the canonical triple (Q, w, Delta) and the precise definition of the measure mu should be introduced earlier and used consistently when stating the theorems.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We thank the referee for the careful review and for identifying this foundational point. The concern about explicit embeddings is well-taken and directly affects the scope of Theorems T1–T7. We address it below and will revise the manuscript accordingly.

read point-by-point responses
  1. Referee: [Abstract (statements of T1–T7) and algebraic appendix] The central claim that SHAP, IG, LIME, and linearized GradCAM belong to the class of continuous linear functionals on the entire L^2(Q, mu) is load-bearing for all seven theorems. Standard definitions of these methods are given only for a fixed model f and a specific input distribution; the manuscript must supply explicit embeddings or constructions showing that the resulting functionals remain linear and bounded (hence continuous) when extended to arbitrary square-integrable functions while exactly reproducing the original numerical values. Without this step, T1–T7 apply only to idealized surrogates rather than the published algorithms.

    Authors: We agree that the referee's observation is correct: the manuscript currently states that the listed methods belong to the class of continuous linear functionals on L^2(Q, mu) but does not supply the explicit embeddings or boundedness proofs needed to justify this for the standard published algorithms. In the revised manuscript we will add a dedicated subsection (placed after the definition of the GRALIS triple and before the statement of the theorems) that provides the required constructions. For each method we will (i) define the corresponding functional on the full L^2 space, (ii) verify linearity in the model output, (iii) prove boundedness with respect to the L^2 norm (using the background measure and the specific form of each method), and (iv) show that the functional coincides with the original numerical output on the domain where the method is conventionally defined. These additions will ensure that T1–T7 apply directly to the published algorithms rather than to idealized surrogates. We will also update the abstract, introduction, and algebraic appendix to reference the new material and will include a short table summarizing the embedding for each method. revision: yes

Circularity Check

0 steps flagged

No circularity; central claim applies external Riesz theorem to assumed linear functionals

full rationale

The derivation rests on the standard Riesz Representation Theorem (an external result from functional analysis) to obtain the unique canonical (Q, w, Delta) form for any additive linear continuous functional on L^2(Q, mu). The abstract states that the GRALIS-SIV correspondence is justified algebraically via the Möbius transform without circularity. No parameters are fitted on a data subset and then presented as predictions, no self-definitional loops exist (e.g., X defined in terms of Y and vice versa), and no load-bearing self-citations or uniqueness theorems imported from the authors' prior work appear. The claim that SHAP/IG/LIME/linearized GradCAM belong to the class is a substantive modeling assumption rather than a tautology; the paper is therefore self-contained against external mathematical benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 1 invented entities

The central claim rests on one standard theorem from functional analysis and the modeling assumption that attributions are linear functionals; no free parameters or new physical entities are introduced in the abstract.

axioms (1)
  • standard math Riesz Representation Theorem: every continuous linear functional on a Hilbert space has a unique representation via inner product with an element of the space
    Invoked to prove that every qualifying attribution admits a unique (Q, w, Delta) form.
invented entities (1)
  • GRALIS canonical triple (Q, w, Delta) no independent evidence
    purpose: Unified representation of linear attribution functionals
    Newly defined object that recovers existing methods as instances; no independent falsifiable prediction supplied in abstract.

pith-pipeline@v0.9.0 · 5630 in / 1516 out tokens · 78333 ms · 2026-05-08T17:00:06.245084+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

28 extracted references · 4 canonical work pages

  1. [1]

    Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., & Batra, D. (2017). Grad-CAM: Visual explanations from deep networks via gradient-based localization. ICCV, 618–626

  2. [2]

    Lundberg, S.M., & Lee, S.-I. (2017). A unified approach to interpreting model predic- tions.NeurIPS 30

  3. [3]

    Why should I trust you?

    Ribeiro, M.T., Singh, S., & Guestrin, C. (2016). “Why should I trust you?”: Explaining the predictions of any classifier.KDD, 1135–1144

  4. [4]

    Sundararajan, M., Taly, A., & Yan, Q. (2017). Axiomatic attribution for deep networks. ICML, 3319–3328

  5. [5]

    Ancona, M., Ceolini, E., Öztireli, C., & Gross, M. (2018). Towards better understanding of gradient-based attribution methods for deep neural networks.ICLR. 23

  6. [6]

    Montavon, G., Lapuschkin, S., Binder, A., Müller, K.-R., & Samek, W. (2017). Explain- ing nonlinear classification decisions with deep Taylor decomposition.Pattern Recogni- tion, 65, 211–222

  7. [7]

    Simonyan, K., Vedaldi, A., & Zisserman, A. (2013). Deep inside convolutional net- works: Visualising image classification models and saliency maps.arXiv preprint arXiv:1312.6034

  8. [8]

    Chattopadhyay, A., Sarkar, A., Howlader, P., & Balasubramanian, V.N. (2018). Grad- CAM++: Generalized gradient-based visual explanations for deep convolutional net- works.WACV, 839–847

  9. [9]

    Covert, I., & Lee, S.-I. (2021). Improving KernelSHAP: Practical Shapley value estima- tion using linear regression.AISTATS

  10. [10]

    Lundstrom, D., Jain, T., & Koyejo, S. (2022). A rigorous study of integrated gradi- ents method and extensions to internal neuron attributions.Transactions on Machine Learning Research (TMLR)

  11. [11]

    Kindermans, P.-J., Hooker, S., Adebayo, J., Alber, M., Schütt, K.T., Dähne, S., Erhan, D., & Kim, B. (2019). The (un)reliability of saliency methods. InExplainability of AI, Springer LNCS, pp. 267–280

  12. [12]

    Hooker, S., Erhan, D., Kindermans, P.-J., & Kim, B. (2019). A benchmark for inter- pretability methods in deep neural networks.NeurIPS 32

  13. [13]

    Wang, H., Wang, Z., Du, M., Yang, F., Zhang, Z., Ding, S., Mardziel, P., & Hu, X. (2020). Score-CAM: Score-weighted visual explanations for convolutional neural net- works.CVPR Workshops

  14. [14]

    Fu, R., Hu, Q., Dong, X., Guo, Y., Gao, Y., & Li, B. (2020). Axiom-based Grad-CAM: Towards accurate visualization and explanation of CNNs.BMVC

  15. [15]

    Draelos, R.L., & Carin, L. (2021). Use HiResCAM instead of Grad-CAM for faithful explanations of convolutional neural networks.arXiv preprint arXiv:2011.08891

  16. [16]

    Petsiuk, V., Das, A., & Saenko, K. (2018). RISE: Randomized input sampling for ex- planation of black-box models.BMVC

  17. [17]

    Rong, Y., Leemann, T., Nguyen, T.-N., Zeitler, L., Jyothiprakash, P., Bhatt, U., Kasneci, E., & Kasneci, G. (2022). Evaluating the faithfulness of saliency-based explanations via the ROAD benchmark.arXiv preprint arXiv:2202.00449

  18. [18]

    Bhatt, U., Weller, A., &Moura, J.M.F.(2020).Evaluatingandaggregatingfeature-based model explanations.IJCAI, 3016–3022

  19. [19]

    Grabisch, M., & Roubens, M. (1999). An axiomatic approach to the concept of interac- tion among players in cooperative games.International Journal of Game Theory, 28(4), 547–565

  20. [20]

    Hoeffding, W. (1948). A class of statistics with asymptotically normal distribution.An- nals of Mathematical Statistics, 19(3), 293–325

  21. [21]

    Efron, B., & Stein, C. (1981). The jackknife estimate of variance.Annals of Statistics, 9(3), 586–596

  22. [22]

    Sobol’, I.M. (1993). Sensitivity estimates for nonlinear mathematical models.Mathemat- ical Modelling and Computational Experiments, 1(4), 407–414. 24

  23. [23]

    Achanta, R., Shaji, A., Smith, K., Lucchi, A., Fua, P., & Süsstrunk, S. (2012). SLIC su- perpixels compared to state-of-the-art superpixel methods.IEEE TPAMI, 34(11), 2274– 2282

  24. [24]

    A., Oliveira, L

    Spanhol, F. A., Oliveira, L. S., Petitjean, C., & Heutte, L. (2016). A dataset for breast cancer histological image classification.IEEE Transactions on Biomedical Engineering, 63(7), 1455–1462

  25. [25]

    Riesz, F. (1909). Sur les opérations fonctionnelles linéaires.Comptes Rendus de l’Académie des Sciences, 149, 974–977

  26. [26]

    Fanale, R., Martini, G., Sciarrone, F., & Caldelli, R. (2026). Explainable ar- tificial intelligence for the analysis of histopathological images of breast cancer: Methods, interpretability and emerging directions.Frontiers in Signal Processing. doi:10.3389/frsip.2026.1795809

  27. [27]

    Fanale, R. et al. (2025). ExpiScore: A quantitative framework for evaluating XAI meth- ods in medical imaging. Manuscript under review.Transparency note: this work shares authorship with the present paper; results involving ExpiScore should be interpreted with this in mind

  28. [28]

    Fanale, R. (2026). GRALIS-LLM: Multimodal explainable AI for automated clinical report generation in breast cancer histology. Manuscript in preparation. 25