pith. machine review for the scientific record. sign in

arxiv: 2605.04497 · v1 · submitted 2026-05-06 · 💻 cs.LG

Recognition: unknown

Quadrature-TreeSHAP: Depth-Independent TreeSHAP and Shapley Interactions

Authors on Pith no claims yet

Pith reviewed 2026-05-08 16:43 UTC · model grok-4.3

classification 💻 cs.LG
keywords Shapley valuesTreeSHAPinteraction valuesquadratureexplainable AIgradient boostingnumerical stabilityBanzhaf values
0
0 comments X

The pith

Quadrature reformulation of TreeSHAP using eight fixed points delivers stable Shapley values and interactions independent of tree depth.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper shows that Path-Dependent SHAP for tree ensembles can be rewritten as the integral of a weighted-Banzhaf polynomial over the unit interval. By evaluating this integral with a fixed eight-point Gauss-Legendre quadrature rule, Shapley values and arbitrary-order interaction values are recovered. The resulting algorithm is numerically more stable than the original recursive TreeSHAP, runs faster on both CPU and GPU, and scales independently of tree depth. These properties make higher-order interaction explanations feasible for large models. The approach is implemented in XGBoost and demonstrated on twelve benchmarks with reported speedups ranging from 1.06x to over 1000x depending on the order of interactions.

Core claim

By expressing the Banzhaf interaction values as expectations under a participation probability p and recovering the corresponding Shapley values through quadrature integration of the resulting polynomial, one obtains a depth-independent computation of Shapley values and any-order interactions that attains machine precision with only eight fixed quadrature nodes and improves numerical stability over direct recursion.

What carries the argument

The weighted-Banzhaf interaction polynomial, which encodes the expectations needed for Banzhaf values under a variable participation probability and permits recovery of Shapley values by a single integral over that probability.

If this is right

  • Shapley values for tree models can be computed to high precision without recursion over tree depth.
  • Pairwise and higher-order Shapley interactions become computationally practical for gradient-boosted trees.
  • CPU and GPU implementations benefit from fixed-point arithmetic without depth-dependent branching.
  • The method integrates into existing libraries such as XGBoost without changing model training.
  • Speedups increase with interaction order, reaching three orders of magnitude for high-order terms.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • If the eight-point rule holds across all realistic trees, then production explainers can drop adaptive quadrature entirely.
  • The polynomial view may allow analytic solutions for trees with limited depth or specific split structures.
  • Similar quadrature reformulations could be applied to other attribution methods that currently rely on path enumeration.
  • Faster access to multi-way interactions supports routine auditing of feature dependencies in deployed models.

Load-bearing premise

The weighted-Banzhaf polynomial together with a fixed eight-point quadrature exactly reproduces the expectations required by the path-dependent Shapley definition for any tree structure and any feature distribution.

What would settle it

Run Quadrature-TreeSHAP with eight points and exact TreeSHAP on a deep tree with non-uniform feature values and observe whether the two sets of Shapley values differ by more than floating-point roundoff error.

Figures

Figures reproduced from arXiv: 2605.04497 by Peng Yu, Ron Wettenstein, Rory Mitchell.

Figure 1
Figure 1. Figure 1: First-order numerical error on Fashion￾MNIST-sparse vs. max_depth view at source ↗
read the original abstract

Shapley values are a standard tool for explaining predictions of tree ensembles, with Path-Dependent SHAP being the most widely used variant. Despite substantial progress, existing methods still exhibit trade-offs between depth-dependent runtime, numerical stability, and support for higher-order interactions. To address these challenges, we introduce Quadrature-TreeSHAP, a quadrature-based reformulation of Path-Dependent TreeSHAP that is numerically stable, naturally extends to any-order Shapley interaction values and is practically insensitive to tree depth. Our implementation supports both CPU and GPU and is integrated into XGBoost. Our method is based on a weighted-Banzhaf interaction polynomial, which expresses Banzhaf interaction values as expectations under a feature participation probability $p$. Shapley values and any-order interaction values are then recovered by integrating these polynomials over $p$ from 0 to 1. We evaluate these integrals using Gauss-Legendre quadrature, and show that, in practice, only 8 fixed quadrature points are sufficient to reach machine precision. In fact, Quadrature-TreeSHAP with 8 fixed points achieves greater numerical stability than TreeSHAP. This fixed-point formulation removes depth dependence from the inner computation and enables efficient SIMD execution. We confirm these advantages empirically. On 12 XGBoost benchmarks, Quadrature-TreeSHAP computes Shapley values 1.06x-10.59x faster than TreeSHAP on CPU and 1.84x-6.95x faster than GPUTreeSHAP on GPU. Shapley pairwise interactions are 3.80x-58.11x faster on CPU, with higher-order interactions achieving speedups of up to 1200x compared to TreeSHAP-IQ.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 0 minor

Summary. The manuscript introduces Quadrature-TreeSHAP, a quadrature-based reformulation of Path-Dependent TreeSHAP. It defines a weighted-Banzhaf interaction polynomial that expresses Banzhaf values as expectations under a participation probability p, recovers Shapley values and arbitrary-order interactions by integrating the polynomial over p ∈ [0,1] via fixed 8-point Gauss-Legendre quadrature, and claims that this yields machine-precision results, removes depth dependence from the inner loop, improves numerical stability over TreeSHAP, and delivers speedups of 1.06×–10.59× for Shapley values and up to 1200× for higher-order interactions on 12 XGBoost benchmarks, with both CPU and GPU implementations integrated into XGBoost.

Significance. If the quadrature approximation is shown to be accurate to machine precision for the tree depths and feature distributions encountered in practice, the work would provide a practically useful advance in post-hoc explanation of tree ensembles: a depth-independent algorithm that is faster, more stable, and naturally supports higher-order Shapley interactions. The reported empirical speed and stability numbers on 12 benchmarks constitute concrete evidence of practical benefit.

major comments (1)
  1. [Abstract and quadrature construction] The central claim that 8 fixed Gauss-Legendre points achieve machine precision rests on the integrand being a polynomial of degree ≤15. For any root-to-leaf path of length d the contribution to the integrand contains a term of exact degree d. Gauss-Legendre quadrature with 8 nodes is algebraically exact only up to degree 15; beyond that the error is nonzero. The manuscript asserts machine precision “in practice” on 12 benchmarks but supplies neither an a-priori error bound, an analysis of the leading coefficients, nor the maximum depths of the trees used in those benchmarks. This assumption underpins both the numerical-stability claim and the assertion of depth independence.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We thank the referee for their careful and constructive review. The major comment raises an important point about the theoretical basis for the quadrature precision claim. We address it directly below and will revise the manuscript to incorporate additional supporting information.

read point-by-point responses
  1. Referee: The central claim that 8 fixed Gauss-Legendre points achieve machine precision rests on the integrand being a polynomial of degree ≤15. For any root-to-leaf path of length d the contribution to the integrand contains a term of exact degree d. Gauss-Legendre quadrature with 8 nodes is algebraically exact only up to degree 15; beyond that the error is nonzero. The manuscript asserts machine precision “in practice” on 12 benchmarks but supplies neither an a-priori error bound, an analysis of the leading coefficients, nor the maximum depths of the trees used in those benchmarks. This assumption underpins both the numerical-stability claim and the assertion of depth independence.

    Authors: We agree that 8-point Gauss-Legendre quadrature is algebraically exact only for polynomials of degree ≤15 and that paths with d>15 introduce higher-degree terms for which the error is nonzero in principle. Our claim of machine precision is empirical, based on observed relative errors at the level of 1e-16 across the 12 XGBoost benchmarks. In the revised manuscript we will report the maximum tree depths present in those benchmarks and add a short analysis of the quadrature remainder. The leading coefficients of the weighted-Banzhaf polynomial are bounded (products of split probabilities ≤1), which keeps the 16th-derivative term small on the unit interval and explains why the observed error remains at machine epsilon even when d exceeds 15. The depth-independence claim does not rely on algebraic exactness; it follows from replacing the depth-dependent recursion with a fixed loop over the 8 quadrature nodes, each of which can be evaluated in constant time with respect to d. Numerical stability is likewise demonstrated empirically rather than asserted from exactness. We have revised the abstract, the quadrature-construction section, and the experimental section to include the tree-depth statistics and the brief error discussion. revision: yes

Circularity Check

0 steps flagged

No significant circularity detected in derivation chain

full rationale

The paper derives Quadrature-TreeSHAP by expressing Banzhaf interaction values via a weighted polynomial in participation probability p, then recovering Shapley and higher-order values through integration over p using standard Gauss-Legendre quadrature. This rests on the known equivalence between Banzhaf and Shapley values plus fixed-point numerical integration, with no self-definitional loops, no fitted parameters renamed as predictions, and no load-bearing self-citations or uniqueness theorems invoked from prior author work. The 8-point rule and machine-precision claim are presented as empirical observations on benchmarks rather than tautological reductions, leaving the central construction independent of its own outputs.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The approach relies on standard mathematical quadrature rules and the algebraic relation between Banzhaf and Shapley values; no new free parameters, ad-hoc axioms, or invented entities are introduced.

axioms (1)
  • standard math Gauss-Legendre quadrature with a fixed number of nodes accurately approximates the integral of the Banzhaf polynomial over p in [0,1]
    Invoked when the paper states that 8 fixed points reach machine precision.

pith-pipeline@v0.9.0 · 5622 in / 1261 out tokens · 21501 ms · 2026-05-08T16:43:17.613461+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

64 extracted references

  1. [1]

    Banzhaf values for facts in query answering, 2023

    Omer Abramovich, Daniel Deutch, Nave Frost, Ahmet Kara, and Dan Olteanu. Banzhaf values for facts in query answering, 2023

  2. [2]

    John F. Banzhaf. Weighted voting doesn’t work: A mathematical analysis.Rutgers Law Review, 19:317–343, 1965

  3. [3]

    Blackard and Denis J

    Jock A. Blackard and Denis J. Dean. Covertype data set.https://archive.ics.uci.edu/ml/ datasets/covertype, 1999

  4. [4]

    Random forests.Machine Learning, 45(1):5–32, Oct 2001

    Leo Breiman. Random forests.Machine Learning, 45(1):5–32, Oct 2001

  5. [5]

    Bussmann, P

    N. Bussmann, P. Giudici, D. Marinelli, and J. Papenbrock. Explainable machine learning for credit risk management.Computational Economics, 2021. 11

  6. [6]

    Xgboost: A scalable tree boosting system

    Tianqi Chen and Carlos Guestrin. Xgboost: A scalable tree boosting system. InProceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 785–794. Association for Computing Machinery, 2016

  7. [7]

    Academic Press, 2 edition, 1984

    Philip J Davis and Philip Rabinowitz.Methods of Numerical Integration. Academic Press, 2 edition, 1984

  8. [8]

    The shapley taylor interaction index

    Kedar Dhamdhere, Ashish Agarwal, and Mukund Sundararajan. The shapley taylor interaction index. InProceedings of the 37th International Conference on Machine Learning, ICML’20. JMLR.org, 2020

  9. [9]

    Fighting biases with dynamic boosting.CoRR, 2017

    Anna Veronika Dorogush, Andrey Gulin, Gleb Gusev, Nikita Kazeev, Liudmila Ostroumova Prokhorenkova, and Aleksandr Vorobev. Fighting biases with dynamic boosting.CoRR, 2017

  10. [10]

    Peiyu Du, Heju Huai, Xiaoyang Wu, Hongjia Wang, Wen Liu, and Xiumei Tang. Using xgboost-shap for understanding the ecosystem services trade-off effects and driving mechanisms in ecologically fragile areas.Frontiers in Plant Science, Volume 16 - 2025, 2025

  11. [11]

    On marginal feature attributions of tree-based models.Foundations of Data Science, 6(4):395–467, 2024

    Khashayar Filom, Alexey Miroshnikov, Konstandinos Kotsiopoulos, and Arjun Ravi Kannan. On marginal feature attributions of tree-based models.Foundations of Data Science, 6(4):395–467, 2024

  12. [12]

    Axiomatic characterizations of probabilistic and cardinal-probabilistic interaction indices.Games and Economic Behavior, 55(1):72–99, 2006

    Katsushige Fujimoto, Ivan Kojadinovic, and Jean-Luc Marichal. Axiomatic characterizations of probabilistic and cardinal-probabilistic interaction indices.Games and Economic Behavior, 55(1):72–99, 2006

  13. [13]

    Shap-iq: unified approximation of any-order shapley interactions

    Fabian Fumagalli, Maximilian Muschalik, Patrick Kolpaczki, Eyke Hüllermeier, and Barbara Hammer. Shap-iq: unified approximation of any-order shapley interactions. InProceedings of the 37th International Conference on Neural Information Processing Systems, NIPS ’23, Red Hook, NY, USA, 2023. Curran Associates Inc

  14. [14]

    What every computer scientist should know about floating-point arithmetic

    David Goldberg. What every computer scientist should know about floating-point arithmetic. ACM Comput. Surv., 23(1):5–48, March 1991

  15. [15]

    Calculation of gauss quadrature rules.Mathematics of Computation, 23(106):221–230, 1969

    Gene H Golub and John H Welsch. Calculation of gauss quadrature rules.Mathematics of Computation, 23(106):221–230, 1969

  16. [16]

    An axiomatic approach to the concept of interaction among players in cooperative games.International Journal of Game Theory, 28(4):547–565, Nov 1999

    Michel Grabisch and Marc Roubens. An axiomatic approach to the concept of interaction among players in cooperative games.International Journal of Game Theory, 28(4):547–565, Nov 1999

  17. [17]

    A. A. Hanani, T. B. Donmez, M. Kutlu, and M. Mansour. Predicting thyroid cancer recurrence using supervised catboost: A shap-based explainable ai approach.Medicine, 104(22):e42667, 2025

  18. [18]

    Higham.Accuracy and Stability of Numerical Algorithms

    Nicholas J. Higham.Accuracy and Stability of Numerical Algorithms. Society for Industrial and Applied Mathematics, 2 edition, 2002

  19. [19]

    Updates on the complexity of shap scores

    Xuanxiang Huang and Joao Marques-Silva. Updates on the complexity of shap scores. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, IJCAI ’24, 2024. 12

  20. [20]

    Single-instruction multiple-data execution.Synthesis Lectures on Computer Architecture, 10:1–121, 05 2015

    Christopher Hughes. Single-instruction multiple-data execution.Synthesis Lectures on Computer Architecture, 10:1–121, 05 2015

  21. [21]

    Improved feature importance computation for tree models based on the banzhaf value

    Adam Karczmarz, Tomasz Michalak, Anish Mukherjee, Piotr Sankowski, and Piotr Wygocki. Improved feature importance computation for tree models based on the banzhaf value. In James Cussens and Kun Zhang, editors,Proceedings of UAI 2022, volume 180 ofProceedings of Machine Learning Research, pages 969–979. PMLR, 01–05 Aug 2022

  22. [22]

    Understanding interventional treeshap : How and why it works, 2022

    Gabriel Laberge and Yann Pequignot. Understanding interventional treeshap : How and why it works, 2022

  23. [23]

    Interpreting machine learning models based on shap values in predicting suspended sediment concentration.International Journal of Sediment Research, 40(1):91–107, 2025

    Houda Lamane, Latifa Mouhir, Rachid Moussadek, Bouamar Baghdad, Ozgur Kisi, and Ali El Bilali. Interpreting machine learning models based on shap values in predicting suspended sediment concentration.International Journal of Sediment Research, 40(1):91–107, 2025

  24. [24]

    Loan default predictability with explainable machine learning

    Huan Li and Weixing Wu. Loan default predictability with explainable machine learning. Finance Research Letters, 60:104867, 2024

  25. [25]

    Treegrad-ranker: Feature ranking via o(l)-time gradients for decision trees, 2026

    Weida Li, Yaoliang Yu, and Bryan Kian Hsiang Low. Treegrad-ranker: Feature ranking via o(l)-time gradients for decision trees, 2026

  26. [26]

    Bin Liao, Tao Zhou, Yanping Liu, Min Li, and Tao Zhang. Tackling the wildfire prediction challenge: An explainable artificial intelligence (xai) model combining extreme gradient boosting (xgboost) with shapley additive explanations (shap) for enhanced interpretability and accuracy. Forests, 16(4), 2025

  27. [27]

    Rapid landslide extraction from high-resolution remote sensing images using shap-opt-xgboost.Remote

    Na Lin, Di Zhang, Shanshan Feng, Kai Ding, Libing Tan, Bin Wang, Tao Chen, Weile Li, Xiaoai Dai, Jianping Pan, and Feifei Tang. Rapid landslide extraction from high-resolution remote sensing images using shap-opt-xgboost.Remote. Sens., 15(15):3901, 2023

  28. [28]

    Wright, Niklas Pfister, and Munir Hiabu

    Jinyang Liu, Tessa Steensgaard, Marvin N. Wright, Niklas Pfister, and Munir Hiabu. Fast estimation of partial dependence functions using trees. In Aarti Singh, Maryam Fazel, Daniel Hsu, Simon Lacoste-Julien, Felix Berkenkamp, Tegan Maharaj, Kiri Wagstaff, and Jerry Zhu, editors,Proceedings of the 42nd International Conference on Machine Learning, volume 2...

  29. [29]

    Teal Witter, Flip Korn, Tarfah Alrashed, Dimitris Paparas, Christopher Musco, and Juliana Freire

    Yurong Liu, R. Teal Witter, Flip Korn, Tarfah Alrashed, Dimitris Paparas, Christopher Musco, and Juliana Freire. Kernel banzhaf: A fast and robust estimator for banzhaf values, 2025

  30. [30]

    Surrogateshap: Training- free contributor attribution for text-to-image (t2i) models, 2026

    Mingyu Lu, Soham Gadgil, Chris Lin, Chanwoo Kim, and Su-In Lee. Surrogateshap: Training- free contributor attribution for text-to-image (t2i) models, 2026

  31. [31]

    Exploiting cpu simd extensions to speed-up document scoring with tree ensembles

    Claudio Lucchese, Franco Maria Nardini, Salvatore Orlando, Raffaele Perego, Nicola Tonellotto, and Rossano Venturini. Exploiting cpu simd extensions to speed-up document scoring with tree ensembles. InProceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’16, page 833–836, New York, NY, USA,...

  32. [32]

    Lundberg, Gabriel Erion, Hugh Chen, Alex DeGrave, Jordan M

    Scott M. Lundberg, Gabriel Erion, Hugh Chen, Alex DeGrave, Jordan M. Prutkin, Bala Nair, Ronit Katz, Jonathan Himmelfarb, Nisha Bansal, and Su-In Lee. From local explanations to global understanding with explainable ai for trees.Nature Machine Intelligence, 2(1):56–67, 2020. 13

  33. [33]

    A unified approach to interpreting model predictions

    Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems, volume 30, 2017

  34. [34]

    Weighted banzhaf power and interaction indexes through weighted approximations of games.European Journal of Operational Research, 211(2):352–358, june 2011

    Jean-Luc Marichal and Pierre Mathonet. Weighted banzhaf power and interaction indexes through weighted approximations of games.European Journal of Operational Research, 211(2):352–358, june 2011

  35. [35]

    Np-completeness for calculating power indices of weighted majority games.Theoretical Computer Science, 263(1):305–310, 2001

    Yasuko Matsui and Tomomi Matsui. Np-completeness for calculating power indices of weighted majority games.Theoretical Computer Science, 263(1):305–310, 2001. Combinatorics and Computer Science

  36. [36]

    Model-agnostic interpretability with shapley values

    Andreas Messalas, Yiannis Kanellopoulos, and Christos Makris. Model-agnostic interpretability with shapley values. In2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA), pages 1–7, 2019

  37. [37]

    Gputreeshap: massively parallel exact calculation of shap scores for tree ensembles.PeerJ

    Rory Mitchell, Eibe Frank, and Geoffrey Holmes. Gputreeshap: massively parallel exact calculation of shap scores for tree ensembles.PeerJ. Computer science, 8:e880, 2022

  38. [38]

    Gore, and Bistra Dilkina

    Hannah Murray, Meredith L. Gore, and Bistra Dilkina. Encoding and decoding illegal wildlife trade networks reveals key airport characteristics and undetected hotspots.Communications Earth & Environment, 6(1):399, May 2025

  39. [39]

    shapiq: Shapley interactions for machine learning

    Maximilian Muschalik, Hubert Baniecki, Fabian Fumagalli, Patrick Kolpaczki, Barbara Hammer, and Eyke Hüllermeier. shapiq: Shapley interactions for machine learning. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Systems, volume 37, pages 130324–130357. Curran Associat...

  40. [40]

    Beyond treeshap: Efficient computation of any-order shapley interactions for tree ensembles.Proceedings of the AAAI Conference on Artificial Intelligence, 38(13):14388–14396, Mar

    Maximilian Muschalik, Fabian Fumagalli, Barbara Hammer, and Eyke Hüllermeier. Beyond treeshap: Efficient computation of any-order shapley interactions for tree ensembles.Proceedings of the AAAI Conference on Artificial Intelligence, 38(13):14388–14396, Mar. 2024

  41. [41]

    From decision trees to boolean logic: A fast and unified shap algorithm.Proceedings of the AAAI Conference on Artificial Intelligence, 40(29):24476– 24485, Mar

    Alexander Nadel and Ron Wettenstein. From decision trees to boolean logic: A fast and unified shap algorithm.Proceedings of the AAAI Conference on Artificial Intelligence, 40(29):24476– 24485, Mar. 2026

  42. [42]

    Nowak and Tadeusz Radzik

    Andrzej S. Nowak and Tadeusz Radzik. An alternative characterization of the weighted banzhaf value.International Journal of Game Theory, 29(1):127–132, Feb 2000

  43. [43]

    CUDA C++ Programming Guide

    NVIDIA. CUDA C++ Programming Guide. NVIDIA Documentation, 2026. See SIMT execution, warps, and divergence behavior

  44. [44]

    Multilinear extensions of games.Management Science, 18(5-part-2):64–79, 1972

    Guillermo Owen. Multilinear extensions of games.Management Science, 18(5-part-2):64–79, 1972

  45. [45]

    Kelley Pace and Ronald Barry

    R. Kelley Pace and Ronald Barry. Sparse spatial autoregressions.Statistics & Probability Letters, 33(3):291–297, 1997

  46. [46]

    Nowak, and Theo S

    Tadeusz Radzik, Andrzej S. Nowak, and Theo S. H. Driessen. Weighted banzhaf values. Mathematical Methods of Operations Research, 45(1):109–118, Feb 1997

  47. [47]

    Raykar, Arindam Jati, Sumanta Mukherjee, Nupur Aggarwal, Kanthi Sarpatwar, Giridhar Ganapavarapu, and Roman Vaculin

    Vikas C. Raykar, Arindam Jati, Sumanta Mukherjee, Nupur Aggarwal, Kanthi Sarpatwar, Giridhar Ganapavarapu, and Roman Vaculin. Tsshap: Robust model agnostic feature-based explainability for time series forecasting, 2023. 14

  48. [48]

    A value for n-person games.Contributions to the Theory of Games, pages 307–317, 1953

    Lloyd S Shapley. A value for n-person games.Contributions to the Theory of Games, pages 307–317, 1953

  49. [49]

    S. Suriya. Credit card fraud detection using explainable ai methods.Journal of Information Systems Engineering and Management, 10:415–428, 03 2025

  50. [50]

    Alviar, Larry A

    Constantine Tarabanis, Evangelos Kalampokis, Mahmoud Khalil, Carlos L. Alviar, Larry A. Chinitz, and Lior Jankelson. Explainable shap-xgboost models for in-hospital mortality after myocardial infarction.Cardiovascular Digital Health Journal, 4(4):126–132, 2023

  51. [51]

    Adult data set

    UCI Machine Learning Repository. Adult data set. https://archive.ics.uci.edu/ml/ datasets/adult, 1996

  52. [52]

    On the tractability of shap explanations.J

    Guy Van den Broeck, Anton Lykov, Maximilian Schleich, and Dan Suciu. On the tractability of shap explanations.J. Artif. Int. Res., 74, September 2022

  53. [53]

    Wang and Ruoxi Jia

    Jiachen T. Wang and Ruoxi Jia. Data banzhaf: A robust data valuation framework for machine learning. InProceedings of the 26th International Conference on Artificial Intelligence and Statistics (AISTATS), volume 206 ofProceedings of Machine Learning Research, pages 6388–6421, 2023

  54. [54]

    Hypershap: Shapley values and interactions for explaining hyperparameter optimization.Proceedings of the AAAI Conference on Artificial Intelligence, 40(32):26867–26875, Mar

    Marcel Wever, Maximilian Muschalik, Fabian Fumagalli, and Marius Lindauer. Hypershap: Shapley values and interactions for explaining hyperparameter optimization.Proceedings of the AAAI Conference on Artificial Intelligence, 40(32):26867–26875, Mar. 2026

  55. [55]

    Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms, 2017

    Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms, 2017

  56. [56]

    Fast treeshap: Accelerating shap value computation for trees, 2022

    Jilei Yang. Fast treeshap: Accelerating shap value computation for trees, 2022

  57. [57]

    Zou, Bin Gao, and Ruofei Zhang

    Ting Ye, Hucheng Zhou, Will Y. Zou, Bin Gao, and Ruofei Zhang. Rapidscorer: Fast tree ensemble evaluation by maximizing compactness in data level parallelization. InProceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD ’18, page 941–950, New York, NY, USA, 2018. Association for Computing Machinery

  58. [58]

    Linear treeshap

    Peng Yu, Albert Bifet, Jesse Read, and Chao Xu. Linear treeshap. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors,Advances in Neural Information Processing Systems, volume 35, pages 25818–25828. Curran Associates, Inc., 2022

  59. [59]

    Interpretable machine learning models for crime prediction.Computers, Environment and Urban Systems, 94:101789, 2022

    Xu Zhang, Lin Liu, Minxuan Lan, Guangwen Song, Luzi Xiao, and Jianguo Chen. Interpretable machine learning models for crime prediction.Computers, Environment and Urban Systems, 94:101789, 2022

  60. [60]

    Xgboost-shap based multi-scale evaluation and nonlinear response of river health in guangdong province

    Xingchi Zhou, Guodong Liu, Yongxiang Wu, Gaoxu Wang, Yi Xu, and Tian Liu. Xgboost-shap based multi-scale evaluation and nonlinear response of river health in guangdong province. Ecological Indicators, 178:114138, 2025. A Additional First-Order Python Comparison with TreeGrad To complement the XGBoost first-order benchmark in Sect. 8, Table 5 compares Pyth...

  61. [61]

    By the distributive property:(T∪L)∩M(v) = (T∩M(v))∪(L∩M(v))

  62. [62]

    SinceT, Sare disjoint:T∩M(v) =T∩(M(v)\S)

  63. [63]

    SinceT, Sare disjoint andL⊆S, it follows thatTandLare disjoint

  64. [64]

    fv(T∪L) =R v ∅ Y j∈(T∪L)∩M(v) qv j =R v ∅   Y j∈T∩(M(v)\S) qv j     Y j∈L∩M(v) qv j  

    SinceT, Lare disjoint,(T∩(M(v)\S))and(L∩M(v))are also disjoint. fv(T∪L) =R v ∅ Y j∈(T∪L)∩M(v) qv j =R v ∅   Y j∈T∩(M(v)\S) qv j     Y j∈L∩M(v) qv j   . Thus, (the setT∩(M(v)\S)is independent ofL): ∆Sfv(T) =R v ∅   Y j∈T∩(M(v)\S) qv j  X L⊆S (−1)|S|−|L| Y j∈L∩M(v) qv j . By Lemma 4 (see below): X L⊆S (−1)|S|−|L| Y j∈L∩M(v) qv j =    Y j∈S ...