Recognition: unknown
On the explainability of max-plus neural networks
Pith reviewed 2026-05-09 20:55 UTC · model grok-4.3
The pith
Max-plus neural networks determine their output value from a single most-activated neuron, which directly supports a pixel fragility measure for explanations.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
In max-plus neural networks a single most-activated neuron governs the output value. This property lets the authors define a pixel fragility measure that quantifies whether a change to one pixel can alter the classification decision. Experiments on PneumoniaMnist show the resulting explanations match or exceed the fidelity of SHAP and Integrated Gradients.
What carries the argument
The single most-activated neuron property, which fixes the network output and thereby allows direct computation of each pixel's influence on that output.
If this is right
- Explanations can be read directly from the network activations without training auxiliary models.
- The fragility score identifies individual pixels whose modification is likely to change the predicted class.
- The same traceability holds after training, not only at initialization.
- The method applies at least as well as standard attribution techniques on medical-image classification.
Where Pith is reading between the lines
- The single-neuron rule could be used to prune or regularize networks so that only the dominant path remains active, potentially improving both speed and interpretability.
- Because the property originates from the max-plus algebra, similar traceability may appear in other tropical or max-min architectures.
- If the fragility measure correlates with human-annotated lesion locations on chest X-rays, it could serve as an automated quality check for model decisions in clinical workflows.
Load-bearing premise
The decision process can always be reduced to one most-activated neuron whose value determines the output, and that this reduction yields a faithful pixel fragility score.
What would settle it
An observation that the classification output changes even though the most-activated neuron remains unaffected, or that perturbing the highest-fragility pixel leaves the output unchanged while a low-fragility pixel alters it, on the same PneumoniaMnist images.
Figures
read the original abstract
We investigate the explanability properties of the recently proposed linear-min-max neural networks. At initialization, they can be interpreted as k-medoids with the infinity norm as a distance. Then, they are trained using subgradient descent to better fit the data. The model has been shown to be a universal approximator. Yet, we can trace the decision process because a single most activated neuron is responsible for the value of the output. Using this property, we designed a pixel fragility measure that determines whether changes to a single pixel may be responsible to a change in the classification output. Experiments on the PneumoniaMnist dataset show that this explanation for the output of the neural network compares favorably to SHAP and Integrated Gradient.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper examines the explainability of linear-min-max neural networks, which at initialization act as k-medoids clustering under the infinity norm and are trained via subgradient descent while remaining universal approximators. The central claim is that the output is always determined by a single most-activated neuron, enabling a pixel fragility measure that attributes potential classification changes to individual input pixels; experiments on PneumoniaMnist show this measure compares favorably to SHAP and Integrated Gradients.
Significance. If the single-neuron responsibility property propagates faithfully through layers and the derived fragility measure yields causally accurate pixel attributions, the work would strengthen the case for architecture-specific interpretability in max-plus networks, combining universal approximation with an efficient, non-perturbation-based explanation tool that could outperform generic methods like SHAP in targeted domains such as medical imaging.
major comments (2)
- [Section on the pixel fragility measure and network architecture] The single most-activated neuron property is invoked to justify the pixel fragility measure, but in a multi-layer setting the final argmax depends on intermediate max selections whose pre-images are sets of activations rather than unique input pixels. This makes it unclear whether the fragility score correctly isolates responsibility to specific pixels after composition of max and linear-min operations (see the section defining the pixel fragility measure and the multi-layer architecture description).
- [Experiments section] The experimental claim of favorable comparison on PneumoniaMnist lacks the exact definition or formula for the pixel fragility measure, any statistical tests, sample sizes, hyperparameter controls, or baseline implementation details for SHAP and Integrated Gradients. Without these, the empirical support for the central explainability claim cannot be verified (see the Experiments section).
minor comments (2)
- [Abstract] The abstract contains the typographical error 'explanability' which should read 'explainability'.
- [Abstract] The abstract phrase 'responsible to a change in the classification output' should be revised to 'responsible for a change in the classification output' for grammatical precision.
Simulated Author's Rebuttal
We thank the referee for the constructive comments on our work regarding the explainability of max-plus neural networks. We address each major comment in detail below and will revise the manuscript accordingly to improve clarity and completeness.
read point-by-point responses
-
Referee: The single most-activated neuron property is invoked to justify the pixel fragility measure, but in a multi-layer setting the final argmax depends on intermediate max selections whose pre-images are sets of activations rather than unique input pixels. This makes it unclear whether the fragility score correctly isolates responsibility to specific pixels after composition of max and linear-min operations (see the section defining the pixel fragility measure and the multi-layer architecture description).
Authors: We appreciate this observation. The single most-activated neuron property is defined recursively across layers: at each max layer, one dominant activation is selected, and the linear-min operations map this back to previous layer activations. The pixel fragility measure is computed by tracing this dominant path from the output neuron back to the input pixels, assigning fragility scores based on the sensitivity along this path. While pre-images under max are indeed sets, the measure focuses on the selected dominant branch. To make this explicit, we will expand the definition section with a formal recursive definition and a worked example for multi-layer networks. revision: yes
-
Referee: The experimental claim of favorable comparison on PneumoniaMnist lacks the exact definition or formula for the pixel fragility measure, any statistical tests, sample sizes, hyperparameter controls, or baseline implementation details for SHAP and Integrated Gradients. Without these, the empirical support for the central explainability claim cannot be verified (see the Experiments section).
Authors: We agree that additional details are necessary for reproducibility and verification. In the revised manuscript, we will provide the precise mathematical formula for the pixel fragility measure, specify the number of samples used from PneumoniaMnist (e.g., the test set size or subset evaluated), include statistical significance tests comparing the measures, detail the hyperparameters for training the network and for the baseline methods, and describe the implementation of SHAP and Integrated Gradients (including any libraries or custom code used). revision: yes
Circularity Check
No circularity: pixel fragility measure follows from architectural max property without reduction to fits or self-citations
full rationale
The derivation begins from the max-plus network structure itself, where the output is defined as the maximum over neuron activations; the claim that a single most-activated neuron determines the output value is therefore a direct consequence of the max operation rather than a fitted parameter, renamed result, or self-citation. The pixel fragility measure is constructed by tracing responsibility through this architectural selection, and the PneumoniaMnist experiments serve only as external empirical comparison to SHAP and Integrated Gradients. No load-bearing step equates a prediction to its own input by construction, and the universal-approximator reference is external to the explainability argument.
Axiom & Free-Parameter Ledger
axioms (2)
- domain assumption Linear-min-max neural networks are universal approximators
- domain assumption At initialization the networks correspond to k-medoids with infinity norm
invented entities (1)
-
pixel fragility measure
no independent evidence
Reference graph
Works this paper leans on
-
[1]
Learning morphological representations of image transformations: In- fluence of initialization and layer differentiability
Mihaela Dimitrova, Samy Blusseau, and Santiago Velasco-Forero. Learning morphological representations of image transformations: In- fluence of initialization and layer differentiability. InInternational Conference on Discrete Geometry and Mathematical Morphology, pages 371–383, 2025
2025
-
[2]
Exploiting subgradient sparsity in max-plus neural networks.hal-05502128 preprint, 2026
Ikhlas Enaieh and Olivier Fercoq. Exploiting subgradient sparsity in max-plus neural networks.hal-05502128 preprint, 2026
2026
-
[3]
Methods and applications of (max,+) linear algebra
St ´ephane Gaubert and Max Plus. Methods and applications of (max,+) linear algebra. InAnnual symposium on theoretical aspects of computer science, pages 261–282. Springer, 1997
1997
-
[4]
A survey of methods for explain- ing black box models.ACM computing surveys (CSUR), 51(5):1–42, 2018
Riccardo Guidotti, Anna Monreale, Salvatore Ruggieri, Franco Turini, Fosca Giannotti, and Dino Pedreschi. A survey of methods for explain- ing black box models.ACM computing surveys (CSUR), 51(5):1–42, 2018
2018
-
[5]
On calibration of modern neural networks
Chuan Guo, Geoff Pleiss, Yu Sun, and Kilian Q Weinberger. On calibration of modern neural networks. InInternational conference on machine learning, pages 1321–1330. PMLR, 2017
2017
-
[6]
Adam: A method for stochastic optimization
Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. InProc. of ICLR, 2015
2015
-
[7]
A unified approach to interpreting model predictions.Advances in neural information processing systems, 30, 2017
Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions.Advances in neural information processing systems, 30, 2017
2017
-
[8]
Min-max-plus neural networks.preprint arXiv:2102.06358, 2021
Ye Luo and Shiqing Fan. Min-max-plus neural networks.preprint arXiv:2102.06358, 2021
-
[9]
Christoph Molnar.Interpretable machine learning. Lulu. com, 2020
2020
-
[10]
Morphological network: How far can we go with morphological neurons? InBritish Machine Vision Conference, 2019
Ranjan Mondal, Soumendu Sundar Mukherjee, Sanchayan Santra, and Bhabatosh Chanda. Morphological network: How far can we go with morphological neurons? InBritish Machine Vision Conference, 2019
2019
-
[11]
Clustering by means of medoids
LKPJ Rdusseeun and P Kaufman. Clustering by means of medoids. InProceedings of the statistical data analysis based on the L1 norm conference, neuchatel, switzerland, volume 31, page 28, 1987
1987
-
[12]
An introduction to morphological neural networks
Gerhard X Ritter and Peter Sussner. An introduction to morphological neural networks. InProceedings of 13th International Conference on Pattern Recognition, volume 4, pages 709–717. IEEE, 1996
1996
-
[13]
Grad-CAM: Why did you say that?
Ramprasaath R Selvaraju, Abhishek Das, Ramakrishna Vedantam, Michael Cogswell, Devi Parikh, and Dhruv Batra. Grad-cam: Why did you say that?arXiv preprint arXiv:1611.07450, 2016
work page Pith review arXiv 2016
-
[14]
A comparative evaluation of explainability techniques for image data
Mykyta Skliarov, Radwa El Shawi, Chedia Dhaoui, and Nada Ahmed. A comparative evaluation of explainability techniques for image data. Scientific Reports, 15(1):41898, 2025
2025
-
[15]
Axiomatic attribution for deep networks
Mukund Sundararajan, Ankur Taly, and Qiqi Yan. Axiomatic attribution for deep networks. InInternational conference on machine learning, pages 3319–3328. PMLR, 2017
2017
-
[16]
Medmnist v2-a large-scale lightweight benchmark for 2d and 3d biomedical image classification
Jiancheng Yang, Rui Shi, Donglai Wei, Zequan Liu, Lin Zhao, Bilian Ke, Hanspeter Pfister, and Bingbing Ni. Medmnist v2-a large-scale lightweight benchmark for 2d and 3d biomedical image classification. Scientific data, 10(1):41, 2023. Fig. 1. Examples of SHAP and IntGrad explanations for the multi-layer perceptron (lighter means more important) Fig. 2. Ex...
2023
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.