Recognition: unknown
Quadrature-TreeSHAP: Depth-Independent TreeSHAP and Shapley Interactions
Pith reviewed 2026-05-08 16:43 UTC · model grok-4.3
The pith
Quadrature reformulation of TreeSHAP using eight fixed points delivers stable Shapley values and interactions independent of tree depth.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
By expressing the Banzhaf interaction values as expectations under a participation probability p and recovering the corresponding Shapley values through quadrature integration of the resulting polynomial, one obtains a depth-independent computation of Shapley values and any-order interactions that attains machine precision with only eight fixed quadrature nodes and improves numerical stability over direct recursion.
What carries the argument
The weighted-Banzhaf interaction polynomial, which encodes the expectations needed for Banzhaf values under a variable participation probability and permits recovery of Shapley values by a single integral over that probability.
If this is right
- Shapley values for tree models can be computed to high precision without recursion over tree depth.
- Pairwise and higher-order Shapley interactions become computationally practical for gradient-boosted trees.
- CPU and GPU implementations benefit from fixed-point arithmetic without depth-dependent branching.
- The method integrates into existing libraries such as XGBoost without changing model training.
- Speedups increase with interaction order, reaching three orders of magnitude for high-order terms.
Where Pith is reading between the lines
- If the eight-point rule holds across all realistic trees, then production explainers can drop adaptive quadrature entirely.
- The polynomial view may allow analytic solutions for trees with limited depth or specific split structures.
- Similar quadrature reformulations could be applied to other attribution methods that currently rely on path enumeration.
- Faster access to multi-way interactions supports routine auditing of feature dependencies in deployed models.
Load-bearing premise
The weighted-Banzhaf polynomial together with a fixed eight-point quadrature exactly reproduces the expectations required by the path-dependent Shapley definition for any tree structure and any feature distribution.
What would settle it
Run Quadrature-TreeSHAP with eight points and exact TreeSHAP on a deep tree with non-uniform feature values and observe whether the two sets of Shapley values differ by more than floating-point roundoff error.
Figures
read the original abstract
Shapley values are a standard tool for explaining predictions of tree ensembles, with Path-Dependent SHAP being the most widely used variant. Despite substantial progress, existing methods still exhibit trade-offs between depth-dependent runtime, numerical stability, and support for higher-order interactions. To address these challenges, we introduce Quadrature-TreeSHAP, a quadrature-based reformulation of Path-Dependent TreeSHAP that is numerically stable, naturally extends to any-order Shapley interaction values and is practically insensitive to tree depth. Our implementation supports both CPU and GPU and is integrated into XGBoost. Our method is based on a weighted-Banzhaf interaction polynomial, which expresses Banzhaf interaction values as expectations under a feature participation probability $p$. Shapley values and any-order interaction values are then recovered by integrating these polynomials over $p$ from 0 to 1. We evaluate these integrals using Gauss-Legendre quadrature, and show that, in practice, only 8 fixed quadrature points are sufficient to reach machine precision. In fact, Quadrature-TreeSHAP with 8 fixed points achieves greater numerical stability than TreeSHAP. This fixed-point formulation removes depth dependence from the inner computation and enables efficient SIMD execution. We confirm these advantages empirically. On 12 XGBoost benchmarks, Quadrature-TreeSHAP computes Shapley values 1.06x-10.59x faster than TreeSHAP on CPU and 1.84x-6.95x faster than GPUTreeSHAP on GPU. Shapley pairwise interactions are 3.80x-58.11x faster on CPU, with higher-order interactions achieving speedups of up to 1200x compared to TreeSHAP-IQ.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces Quadrature-TreeSHAP, a quadrature-based reformulation of Path-Dependent TreeSHAP. It defines a weighted-Banzhaf interaction polynomial that expresses Banzhaf values as expectations under a participation probability p, recovers Shapley values and arbitrary-order interactions by integrating the polynomial over p ∈ [0,1] via fixed 8-point Gauss-Legendre quadrature, and claims that this yields machine-precision results, removes depth dependence from the inner loop, improves numerical stability over TreeSHAP, and delivers speedups of 1.06×–10.59× for Shapley values and up to 1200× for higher-order interactions on 12 XGBoost benchmarks, with both CPU and GPU implementations integrated into XGBoost.
Significance. If the quadrature approximation is shown to be accurate to machine precision for the tree depths and feature distributions encountered in practice, the work would provide a practically useful advance in post-hoc explanation of tree ensembles: a depth-independent algorithm that is faster, more stable, and naturally supports higher-order Shapley interactions. The reported empirical speed and stability numbers on 12 benchmarks constitute concrete evidence of practical benefit.
major comments (1)
- [Abstract and quadrature construction] The central claim that 8 fixed Gauss-Legendre points achieve machine precision rests on the integrand being a polynomial of degree ≤15. For any root-to-leaf path of length d the contribution to the integrand contains a term of exact degree d. Gauss-Legendre quadrature with 8 nodes is algebraically exact only up to degree 15; beyond that the error is nonzero. The manuscript asserts machine precision “in practice” on 12 benchmarks but supplies neither an a-priori error bound, an analysis of the leading coefficients, nor the maximum depths of the trees used in those benchmarks. This assumption underpins both the numerical-stability claim and the assertion of depth independence.
Simulated Author's Rebuttal
We thank the referee for their careful and constructive review. The major comment raises an important point about the theoretical basis for the quadrature precision claim. We address it directly below and will revise the manuscript to incorporate additional supporting information.
read point-by-point responses
-
Referee: The central claim that 8 fixed Gauss-Legendre points achieve machine precision rests on the integrand being a polynomial of degree ≤15. For any root-to-leaf path of length d the contribution to the integrand contains a term of exact degree d. Gauss-Legendre quadrature with 8 nodes is algebraically exact only up to degree 15; beyond that the error is nonzero. The manuscript asserts machine precision “in practice” on 12 benchmarks but supplies neither an a-priori error bound, an analysis of the leading coefficients, nor the maximum depths of the trees used in those benchmarks. This assumption underpins both the numerical-stability claim and the assertion of depth independence.
Authors: We agree that 8-point Gauss-Legendre quadrature is algebraically exact only for polynomials of degree ≤15 and that paths with d>15 introduce higher-degree terms for which the error is nonzero in principle. Our claim of machine precision is empirical, based on observed relative errors at the level of 1e-16 across the 12 XGBoost benchmarks. In the revised manuscript we will report the maximum tree depths present in those benchmarks and add a short analysis of the quadrature remainder. The leading coefficients of the weighted-Banzhaf polynomial are bounded (products of split probabilities ≤1), which keeps the 16th-derivative term small on the unit interval and explains why the observed error remains at machine epsilon even when d exceeds 15. The depth-independence claim does not rely on algebraic exactness; it follows from replacing the depth-dependent recursion with a fixed loop over the 8 quadrature nodes, each of which can be evaluated in constant time with respect to d. Numerical stability is likewise demonstrated empirically rather than asserted from exactness. We have revised the abstract, the quadrature-construction section, and the experimental section to include the tree-depth statistics and the brief error discussion. revision: yes
Circularity Check
No significant circularity detected in derivation chain
full rationale
The paper derives Quadrature-TreeSHAP by expressing Banzhaf interaction values via a weighted polynomial in participation probability p, then recovering Shapley and higher-order values through integration over p using standard Gauss-Legendre quadrature. This rests on the known equivalence between Banzhaf and Shapley values plus fixed-point numerical integration, with no self-definitional loops, no fitted parameters renamed as predictions, and no load-bearing self-citations or uniqueness theorems invoked from prior author work. The 8-point rule and machine-precision claim are presented as empirical observations on benchmarks rather than tautological reductions, leaving the central construction independent of its own outputs.
Axiom & Free-Parameter Ledger
axioms (1)
- standard math Gauss-Legendre quadrature with a fixed number of nodes accurately approximates the integral of the Banzhaf polynomial over p in [0,1]
Reference graph
Works this paper leans on
-
[1]
Banzhaf values for facts in query answering, 2023
Omer Abramovich, Daniel Deutch, Nave Frost, Ahmet Kara, and Dan Olteanu. Banzhaf values for facts in query answering, 2023
2023
-
[2]
John F. Banzhaf. Weighted voting doesn’t work: A mathematical analysis.Rutgers Law Review, 19:317–343, 1965
1965
-
[3]
Blackard and Denis J
Jock A. Blackard and Denis J. Dean. Covertype data set.https://archive.ics.uci.edu/ml/ datasets/covertype, 1999
1999
-
[4]
Random forests.Machine Learning, 45(1):5–32, Oct 2001
Leo Breiman. Random forests.Machine Learning, 45(1):5–32, Oct 2001
2001
-
[5]
Bussmann, P
N. Bussmann, P. Giudici, D. Marinelli, and J. Papenbrock. Explainable machine learning for credit risk management.Computational Economics, 2021. 11
2021
-
[6]
Xgboost: A scalable tree boosting system
Tianqi Chen and Carlos Guestrin. Xgboost: A scalable tree boosting system. InProceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 785–794. Association for Computing Machinery, 2016
2016
-
[7]
Academic Press, 2 edition, 1984
Philip J Davis and Philip Rabinowitz.Methods of Numerical Integration. Academic Press, 2 edition, 1984
1984
-
[8]
The shapley taylor interaction index
Kedar Dhamdhere, Ashish Agarwal, and Mukund Sundararajan. The shapley taylor interaction index. InProceedings of the 37th International Conference on Machine Learning, ICML’20. JMLR.org, 2020
2020
-
[9]
Fighting biases with dynamic boosting.CoRR, 2017
Anna Veronika Dorogush, Andrey Gulin, Gleb Gusev, Nikita Kazeev, Liudmila Ostroumova Prokhorenkova, and Aleksandr Vorobev. Fighting biases with dynamic boosting.CoRR, 2017
2017
-
[10]
Peiyu Du, Heju Huai, Xiaoyang Wu, Hongjia Wang, Wen Liu, and Xiumei Tang. Using xgboost-shap for understanding the ecosystem services trade-off effects and driving mechanisms in ecologically fragile areas.Frontiers in Plant Science, Volume 16 - 2025, 2025
2025
-
[11]
On marginal feature attributions of tree-based models.Foundations of Data Science, 6(4):395–467, 2024
Khashayar Filom, Alexey Miroshnikov, Konstandinos Kotsiopoulos, and Arjun Ravi Kannan. On marginal feature attributions of tree-based models.Foundations of Data Science, 6(4):395–467, 2024
2024
-
[12]
Axiomatic characterizations of probabilistic and cardinal-probabilistic interaction indices.Games and Economic Behavior, 55(1):72–99, 2006
Katsushige Fujimoto, Ivan Kojadinovic, and Jean-Luc Marichal. Axiomatic characterizations of probabilistic and cardinal-probabilistic interaction indices.Games and Economic Behavior, 55(1):72–99, 2006
2006
-
[13]
Shap-iq: unified approximation of any-order shapley interactions
Fabian Fumagalli, Maximilian Muschalik, Patrick Kolpaczki, Eyke Hüllermeier, and Barbara Hammer. Shap-iq: unified approximation of any-order shapley interactions. InProceedings of the 37th International Conference on Neural Information Processing Systems, NIPS ’23, Red Hook, NY, USA, 2023. Curran Associates Inc
2023
-
[14]
What every computer scientist should know about floating-point arithmetic
David Goldberg. What every computer scientist should know about floating-point arithmetic. ACM Comput. Surv., 23(1):5–48, March 1991
1991
-
[15]
Calculation of gauss quadrature rules.Mathematics of Computation, 23(106):221–230, 1969
Gene H Golub and John H Welsch. Calculation of gauss quadrature rules.Mathematics of Computation, 23(106):221–230, 1969
1969
-
[16]
An axiomatic approach to the concept of interaction among players in cooperative games.International Journal of Game Theory, 28(4):547–565, Nov 1999
Michel Grabisch and Marc Roubens. An axiomatic approach to the concept of interaction among players in cooperative games.International Journal of Game Theory, 28(4):547–565, Nov 1999
1999
-
[17]
A. A. Hanani, T. B. Donmez, M. Kutlu, and M. Mansour. Predicting thyroid cancer recurrence using supervised catboost: A shap-based explainable ai approach.Medicine, 104(22):e42667, 2025
2025
-
[18]
Higham.Accuracy and Stability of Numerical Algorithms
Nicholas J. Higham.Accuracy and Stability of Numerical Algorithms. Society for Industrial and Applied Mathematics, 2 edition, 2002
2002
-
[19]
Updates on the complexity of shap scores
Xuanxiang Huang and Joao Marques-Silva. Updates on the complexity of shap scores. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, IJCAI ’24, 2024. 12
2024
-
[20]
Single-instruction multiple-data execution.Synthesis Lectures on Computer Architecture, 10:1–121, 05 2015
Christopher Hughes. Single-instruction multiple-data execution.Synthesis Lectures on Computer Architecture, 10:1–121, 05 2015
2015
-
[21]
Improved feature importance computation for tree models based on the banzhaf value
Adam Karczmarz, Tomasz Michalak, Anish Mukherjee, Piotr Sankowski, and Piotr Wygocki. Improved feature importance computation for tree models based on the banzhaf value. In James Cussens and Kun Zhang, editors,Proceedings of UAI 2022, volume 180 ofProceedings of Machine Learning Research, pages 969–979. PMLR, 01–05 Aug 2022
2022
-
[22]
Understanding interventional treeshap : How and why it works, 2022
Gabriel Laberge and Yann Pequignot. Understanding interventional treeshap : How and why it works, 2022
2022
-
[23]
Interpreting machine learning models based on shap values in predicting suspended sediment concentration.International Journal of Sediment Research, 40(1):91–107, 2025
Houda Lamane, Latifa Mouhir, Rachid Moussadek, Bouamar Baghdad, Ozgur Kisi, and Ali El Bilali. Interpreting machine learning models based on shap values in predicting suspended sediment concentration.International Journal of Sediment Research, 40(1):91–107, 2025
2025
-
[24]
Loan default predictability with explainable machine learning
Huan Li and Weixing Wu. Loan default predictability with explainable machine learning. Finance Research Letters, 60:104867, 2024
2024
-
[25]
Treegrad-ranker: Feature ranking via o(l)-time gradients for decision trees, 2026
Weida Li, Yaoliang Yu, and Bryan Kian Hsiang Low. Treegrad-ranker: Feature ranking via o(l)-time gradients for decision trees, 2026
2026
-
[26]
Bin Liao, Tao Zhou, Yanping Liu, Min Li, and Tao Zhang. Tackling the wildfire prediction challenge: An explainable artificial intelligence (xai) model combining extreme gradient boosting (xgboost) with shapley additive explanations (shap) for enhanced interpretability and accuracy. Forests, 16(4), 2025
2025
-
[27]
Rapid landslide extraction from high-resolution remote sensing images using shap-opt-xgboost.Remote
Na Lin, Di Zhang, Shanshan Feng, Kai Ding, Libing Tan, Bin Wang, Tao Chen, Weile Li, Xiaoai Dai, Jianping Pan, and Feifei Tang. Rapid landslide extraction from high-resolution remote sensing images using shap-opt-xgboost.Remote. Sens., 15(15):3901, 2023
2023
-
[28]
Wright, Niklas Pfister, and Munir Hiabu
Jinyang Liu, Tessa Steensgaard, Marvin N. Wright, Niklas Pfister, and Munir Hiabu. Fast estimation of partial dependence functions using trees. In Aarti Singh, Maryam Fazel, Daniel Hsu, Simon Lacoste-Julien, Felix Berkenkamp, Tegan Maharaj, Kiri Wagstaff, and Jerry Zhu, editors,Proceedings of the 42nd International Conference on Machine Learning, volume 2...
2025
-
[29]
Teal Witter, Flip Korn, Tarfah Alrashed, Dimitris Paparas, Christopher Musco, and Juliana Freire
Yurong Liu, R. Teal Witter, Flip Korn, Tarfah Alrashed, Dimitris Paparas, Christopher Musco, and Juliana Freire. Kernel banzhaf: A fast and robust estimator for banzhaf values, 2025
2025
-
[30]
Surrogateshap: Training- free contributor attribution for text-to-image (t2i) models, 2026
Mingyu Lu, Soham Gadgil, Chris Lin, Chanwoo Kim, and Su-In Lee. Surrogateshap: Training- free contributor attribution for text-to-image (t2i) models, 2026
2026
-
[31]
Exploiting cpu simd extensions to speed-up document scoring with tree ensembles
Claudio Lucchese, Franco Maria Nardini, Salvatore Orlando, Raffaele Perego, Nicola Tonellotto, and Rossano Venturini. Exploiting cpu simd extensions to speed-up document scoring with tree ensembles. InProceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’16, page 833–836, New York, NY, USA,...
2016
-
[32]
Lundberg, Gabriel Erion, Hugh Chen, Alex DeGrave, Jordan M
Scott M. Lundberg, Gabriel Erion, Hugh Chen, Alex DeGrave, Jordan M. Prutkin, Bala Nair, Ronit Katz, Jonathan Himmelfarb, Nisha Bansal, and Su-In Lee. From local explanations to global understanding with explainable ai for trees.Nature Machine Intelligence, 2(1):56–67, 2020. 13
2020
-
[33]
A unified approach to interpreting model predictions
Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions. In Advances in Neural Information Processing Systems, volume 30, 2017
2017
-
[34]
Weighted banzhaf power and interaction indexes through weighted approximations of games.European Journal of Operational Research, 211(2):352–358, june 2011
Jean-Luc Marichal and Pierre Mathonet. Weighted banzhaf power and interaction indexes through weighted approximations of games.European Journal of Operational Research, 211(2):352–358, june 2011
2011
-
[35]
Np-completeness for calculating power indices of weighted majority games.Theoretical Computer Science, 263(1):305–310, 2001
Yasuko Matsui and Tomomi Matsui. Np-completeness for calculating power indices of weighted majority games.Theoretical Computer Science, 263(1):305–310, 2001. Combinatorics and Computer Science
2001
-
[36]
Model-agnostic interpretability with shapley values
Andreas Messalas, Yiannis Kanellopoulos, and Christos Makris. Model-agnostic interpretability with shapley values. In2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA), pages 1–7, 2019
2019
-
[37]
Gputreeshap: massively parallel exact calculation of shap scores for tree ensembles.PeerJ
Rory Mitchell, Eibe Frank, and Geoffrey Holmes. Gputreeshap: massively parallel exact calculation of shap scores for tree ensembles.PeerJ. Computer science, 8:e880, 2022
2022
-
[38]
Gore, and Bistra Dilkina
Hannah Murray, Meredith L. Gore, and Bistra Dilkina. Encoding and decoding illegal wildlife trade networks reveals key airport characteristics and undetected hotspots.Communications Earth & Environment, 6(1):399, May 2025
2025
-
[39]
shapiq: Shapley interactions for machine learning
Maximilian Muschalik, Hubert Baniecki, Fabian Fumagalli, Patrick Kolpaczki, Barbara Hammer, and Eyke Hüllermeier. shapiq: Shapley interactions for machine learning. In A. Globerson, L. Mackey, D. Belgrave, A. Fan, U. Paquet, J. Tomczak, and C. Zhang, editors,Advances in Neural Information Processing Systems, volume 37, pages 130324–130357. Curran Associat...
2024
-
[40]
Beyond treeshap: Efficient computation of any-order shapley interactions for tree ensembles.Proceedings of the AAAI Conference on Artificial Intelligence, 38(13):14388–14396, Mar
Maximilian Muschalik, Fabian Fumagalli, Barbara Hammer, and Eyke Hüllermeier. Beyond treeshap: Efficient computation of any-order shapley interactions for tree ensembles.Proceedings of the AAAI Conference on Artificial Intelligence, 38(13):14388–14396, Mar. 2024
2024
-
[41]
From decision trees to boolean logic: A fast and unified shap algorithm.Proceedings of the AAAI Conference on Artificial Intelligence, 40(29):24476– 24485, Mar
Alexander Nadel and Ron Wettenstein. From decision trees to boolean logic: A fast and unified shap algorithm.Proceedings of the AAAI Conference on Artificial Intelligence, 40(29):24476– 24485, Mar. 2026
2026
-
[42]
Nowak and Tadeusz Radzik
Andrzej S. Nowak and Tadeusz Radzik. An alternative characterization of the weighted banzhaf value.International Journal of Game Theory, 29(1):127–132, Feb 2000
2000
-
[43]
CUDA C++ Programming Guide
NVIDIA. CUDA C++ Programming Guide. NVIDIA Documentation, 2026. See SIMT execution, warps, and divergence behavior
2026
-
[44]
Multilinear extensions of games.Management Science, 18(5-part-2):64–79, 1972
Guillermo Owen. Multilinear extensions of games.Management Science, 18(5-part-2):64–79, 1972
1972
-
[45]
Kelley Pace and Ronald Barry
R. Kelley Pace and Ronald Barry. Sparse spatial autoregressions.Statistics & Probability Letters, 33(3):291–297, 1997
1997
-
[46]
Nowak, and Theo S
Tadeusz Radzik, Andrzej S. Nowak, and Theo S. H. Driessen. Weighted banzhaf values. Mathematical Methods of Operations Research, 45(1):109–118, Feb 1997
1997
-
[47]
Raykar, Arindam Jati, Sumanta Mukherjee, Nupur Aggarwal, Kanthi Sarpatwar, Giridhar Ganapavarapu, and Roman Vaculin
Vikas C. Raykar, Arindam Jati, Sumanta Mukherjee, Nupur Aggarwal, Kanthi Sarpatwar, Giridhar Ganapavarapu, and Roman Vaculin. Tsshap: Robust model agnostic feature-based explainability for time series forecasting, 2023. 14
2023
-
[48]
A value for n-person games.Contributions to the Theory of Games, pages 307–317, 1953
Lloyd S Shapley. A value for n-person games.Contributions to the Theory of Games, pages 307–317, 1953
1953
-
[49]
S. Suriya. Credit card fraud detection using explainable ai methods.Journal of Information Systems Engineering and Management, 10:415–428, 03 2025
2025
-
[50]
Alviar, Larry A
Constantine Tarabanis, Evangelos Kalampokis, Mahmoud Khalil, Carlos L. Alviar, Larry A. Chinitz, and Lior Jankelson. Explainable shap-xgboost models for in-hospital mortality after myocardial infarction.Cardiovascular Digital Health Journal, 4(4):126–132, 2023
2023
-
[51]
Adult data set
UCI Machine Learning Repository. Adult data set. https://archive.ics.uci.edu/ml/ datasets/adult, 1996
1996
-
[52]
On the tractability of shap explanations.J
Guy Van den Broeck, Anton Lykov, Maximilian Schleich, and Dan Suciu. On the tractability of shap explanations.J. Artif. Int. Res., 74, September 2022
2022
-
[53]
Wang and Ruoxi Jia
Jiachen T. Wang and Ruoxi Jia. Data banzhaf: A robust data valuation framework for machine learning. InProceedings of the 26th International Conference on Artificial Intelligence and Statistics (AISTATS), volume 206 ofProceedings of Machine Learning Research, pages 6388–6421, 2023
2023
-
[54]
Hypershap: Shapley values and interactions for explaining hyperparameter optimization.Proceedings of the AAAI Conference on Artificial Intelligence, 40(32):26867–26875, Mar
Marcel Wever, Maximilian Muschalik, Fabian Fumagalli, and Marius Lindauer. Hypershap: Shapley values and interactions for explaining hyperparameter optimization.Proceedings of the AAAI Conference on Artificial Intelligence, 40(32):26867–26875, Mar. 2026
2026
-
[55]
Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms, 2017
Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms, 2017
2017
-
[56]
Fast treeshap: Accelerating shap value computation for trees, 2022
Jilei Yang. Fast treeshap: Accelerating shap value computation for trees, 2022
2022
-
[57]
Zou, Bin Gao, and Ruofei Zhang
Ting Ye, Hucheng Zhou, Will Y. Zou, Bin Gao, and Ruofei Zhang. Rapidscorer: Fast tree ensemble evaluation by maximizing compactness in data level parallelization. InProceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD ’18, page 941–950, New York, NY, USA, 2018. Association for Computing Machinery
2018
-
[58]
Linear treeshap
Peng Yu, Albert Bifet, Jesse Read, and Chao Xu. Linear treeshap. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors,Advances in Neural Information Processing Systems, volume 35, pages 25818–25828. Curran Associates, Inc., 2022
2022
-
[59]
Interpretable machine learning models for crime prediction.Computers, Environment and Urban Systems, 94:101789, 2022
Xu Zhang, Lin Liu, Minxuan Lan, Guangwen Song, Luzi Xiao, and Jianguo Chen. Interpretable machine learning models for crime prediction.Computers, Environment and Urban Systems, 94:101789, 2022
2022
-
[60]
Xgboost-shap based multi-scale evaluation and nonlinear response of river health in guangdong province
Xingchi Zhou, Guodong Liu, Yongxiang Wu, Gaoxu Wang, Yi Xu, and Tian Liu. Xgboost-shap based multi-scale evaluation and nonlinear response of river health in guangdong province. Ecological Indicators, 178:114138, 2025. A Additional First-Order Python Comparison with TreeGrad To complement the XGBoost first-order benchmark in Sect. 8, Table 5 compares Pyth...
2025
-
[61]
By the distributive property:(T∪L)∩M(v) = (T∩M(v))∪(L∩M(v))
-
[62]
SinceT, Sare disjoint:T∩M(v) =T∩(M(v)\S)
-
[63]
SinceT, Sare disjoint andL⊆S, it follows thatTandLare disjoint
-
[64]
fv(T∪L) =R v ∅ Y j∈(T∪L)∩M(v) qv j =R v ∅ Y j∈T∩(M(v)\S) qv j Y j∈L∩M(v) qv j
SinceT, Lare disjoint,(T∩(M(v)\S))and(L∩M(v))are also disjoint. fv(T∪L) =R v ∅ Y j∈(T∪L)∩M(v) qv j =R v ∅ Y j∈T∩(M(v)\S) qv j Y j∈L∩M(v) qv j . Thus, (the setT∩(M(v)\S)is independent ofL): ∆Sfv(T) =R v ∅ Y j∈T∩(M(v)\S) qv j X L⊆S (−1)|S|−|L| Y j∈L∩M(v) qv j . By Lemma 4 (see below): X L⊆S (−1)|S|−|L| Y j∈L∩M(v) qv j = Y j∈S ...
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.