pith. machine review for the scientific record. sign in

arxiv: 2605.00889 · v1 · submitted 2026-04-27 · 💻 cs.CV · cs.LG

Recognition: unknown

On the explainability of max-plus neural networks

Authors on Pith no claims yet

Pith reviewed 2026-05-09 20:55 UTC · model grok-4.3

classification 💻 cs.CV cs.LG
keywords max-plus neural networksexplainabilityneural network interpretabilitypixel fragilityPneumoniaMnistSHAPintegrated gradients
0
0 comments X

The pith

Max-plus neural networks determine their output value from a single most-activated neuron, which directly supports a pixel fragility measure for explanations.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper shows that linear-min-max neural networks, interpretable at start as k-medoids clustering under the infinity norm and trained by subgradient descent, remain universal approximators while preserving traceability. Because the final output is always set by one dominant neuron, the authors construct a pixel fragility score that checks whether altering an individual pixel could flip the classification. On the PneumoniaMnist chest X-ray dataset this score produces explanations that compare favorably with SHAP and Integrated Gradients. The approach therefore supplies model-native attributions without post-hoc approximation, which matters for safety-critical vision tasks where users need to know exactly which image regions drive a decision.

Core claim

In max-plus neural networks a single most-activated neuron governs the output value. This property lets the authors define a pixel fragility measure that quantifies whether a change to one pixel can alter the classification decision. Experiments on PneumoniaMnist show the resulting explanations match or exceed the fidelity of SHAP and Integrated Gradients.

What carries the argument

The single most-activated neuron property, which fixes the network output and thereby allows direct computation of each pixel's influence on that output.

If this is right

  • Explanations can be read directly from the network activations without training auxiliary models.
  • The fragility score identifies individual pixels whose modification is likely to change the predicted class.
  • The same traceability holds after training, not only at initialization.
  • The method applies at least as well as standard attribution techniques on medical-image classification.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The single-neuron rule could be used to prune or regularize networks so that only the dominant path remains active, potentially improving both speed and interpretability.
  • Because the property originates from the max-plus algebra, similar traceability may appear in other tropical or max-min architectures.
  • If the fragility measure correlates with human-annotated lesion locations on chest X-rays, it could serve as an automated quality check for model decisions in clinical workflows.

Load-bearing premise

The decision process can always be reduced to one most-activated neuron whose value determines the output, and that this reduction yields a faithful pixel fragility score.

What would settle it

An observation that the classification output changes even though the most-activated neuron remains unaffected, or that perturbing the highest-fragility pixel leaves the output unchanged while a low-fragility pixel alters it, on the same PneumoniaMnist images.

Figures

Figures reproduced from arXiv: 2605.00889 by Garc\'ia \'Angel (DATSI, Ikhlas Enaieh (S2A, LTCI), Olivier Fercoq (S2A, UPM).

Figure 2
Figure 2. Figure 2: Examples of SHAP, IntGrad and Pixel Fragility explanations for the [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
read the original abstract

We investigate the explanability properties of the recently proposed linear-min-max neural networks. At initialization, they can be interpreted as k-medoids with the infinity norm as a distance. Then, they are trained using subgradient descent to better fit the data. The model has been shown to be a universal approximator. Yet, we can trace the decision process because a single most activated neuron is responsible for the value of the output. Using this property, we designed a pixel fragility measure that determines whether changes to a single pixel may be responsible to a change in the classification output. Experiments on the PneumoniaMnist dataset show that this explanation for the output of the neural network compares favorably to SHAP and Integrated Gradient.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 2 minor

Summary. The paper examines the explainability of linear-min-max neural networks, which at initialization act as k-medoids clustering under the infinity norm and are trained via subgradient descent while remaining universal approximators. The central claim is that the output is always determined by a single most-activated neuron, enabling a pixel fragility measure that attributes potential classification changes to individual input pixels; experiments on PneumoniaMnist show this measure compares favorably to SHAP and Integrated Gradients.

Significance. If the single-neuron responsibility property propagates faithfully through layers and the derived fragility measure yields causally accurate pixel attributions, the work would strengthen the case for architecture-specific interpretability in max-plus networks, combining universal approximation with an efficient, non-perturbation-based explanation tool that could outperform generic methods like SHAP in targeted domains such as medical imaging.

major comments (2)
  1. [Section on the pixel fragility measure and network architecture] The single most-activated neuron property is invoked to justify the pixel fragility measure, but in a multi-layer setting the final argmax depends on intermediate max selections whose pre-images are sets of activations rather than unique input pixels. This makes it unclear whether the fragility score correctly isolates responsibility to specific pixels after composition of max and linear-min operations (see the section defining the pixel fragility measure and the multi-layer architecture description).
  2. [Experiments section] The experimental claim of favorable comparison on PneumoniaMnist lacks the exact definition or formula for the pixel fragility measure, any statistical tests, sample sizes, hyperparameter controls, or baseline implementation details for SHAP and Integrated Gradients. Without these, the empirical support for the central explainability claim cannot be verified (see the Experiments section).
minor comments (2)
  1. [Abstract] The abstract contains the typographical error 'explanability' which should read 'explainability'.
  2. [Abstract] The abstract phrase 'responsible to a change in the classification output' should be revised to 'responsible for a change in the classification output' for grammatical precision.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for the constructive comments on our work regarding the explainability of max-plus neural networks. We address each major comment in detail below and will revise the manuscript accordingly to improve clarity and completeness.

read point-by-point responses
  1. Referee: The single most-activated neuron property is invoked to justify the pixel fragility measure, but in a multi-layer setting the final argmax depends on intermediate max selections whose pre-images are sets of activations rather than unique input pixels. This makes it unclear whether the fragility score correctly isolates responsibility to specific pixels after composition of max and linear-min operations (see the section defining the pixel fragility measure and the multi-layer architecture description).

    Authors: We appreciate this observation. The single most-activated neuron property is defined recursively across layers: at each max layer, one dominant activation is selected, and the linear-min operations map this back to previous layer activations. The pixel fragility measure is computed by tracing this dominant path from the output neuron back to the input pixels, assigning fragility scores based on the sensitivity along this path. While pre-images under max are indeed sets, the measure focuses on the selected dominant branch. To make this explicit, we will expand the definition section with a formal recursive definition and a worked example for multi-layer networks. revision: yes

  2. Referee: The experimental claim of favorable comparison on PneumoniaMnist lacks the exact definition or formula for the pixel fragility measure, any statistical tests, sample sizes, hyperparameter controls, or baseline implementation details for SHAP and Integrated Gradients. Without these, the empirical support for the central explainability claim cannot be verified (see the Experiments section).

    Authors: We agree that additional details are necessary for reproducibility and verification. In the revised manuscript, we will provide the precise mathematical formula for the pixel fragility measure, specify the number of samples used from PneumoniaMnist (e.g., the test set size or subset evaluated), include statistical significance tests comparing the measures, detail the hyperparameters for training the network and for the baseline methods, and describe the implementation of SHAP and Integrated Gradients (including any libraries or custom code used). revision: yes

Circularity Check

0 steps flagged

No circularity: pixel fragility measure follows from architectural max property without reduction to fits or self-citations

full rationale

The derivation begins from the max-plus network structure itself, where the output is defined as the maximum over neuron activations; the claim that a single most-activated neuron determines the output value is therefore a direct consequence of the max operation rather than a fitted parameter, renamed result, or self-citation. The pixel fragility measure is constructed by tracing responsibility through this architectural selection, and the PneumoniaMnist experiments serve only as external empirical comparison to SHAP and Integrated Gradients. No load-bearing step equates a prediction to its own input by construction, and the universal-approximator reference is external to the explainability argument.

Axiom & Free-Parameter Ledger

0 free parameters · 2 axioms · 1 invented entities

The work relies on the previously established universal approximation property of the networks and their initialization as k-medoids; the new contribution is the fragility measure derived from the activation property.

axioms (2)
  • domain assumption Linear-min-max neural networks are universal approximators
    Stated as previously shown; invoked to support the model's capability before introducing explainability.
  • domain assumption At initialization the networks correspond to k-medoids with infinity norm
    Used to ground the initial interpretability claim.
invented entities (1)
  • pixel fragility measure no independent evidence
    purpose: Quantify whether a single pixel change can alter the classification output via the dominant neuron
    Newly proposed in the paper to leverage the single-neuron responsibility property.

pith-pipeline@v0.9.0 · 5435 in / 1546 out tokens · 34101 ms · 2026-05-09T20:55:41.181338+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

16 extracted references · 2 canonical work pages

  1. [1]

    Learning morphological representations of image transformations: In- fluence of initialization and layer differentiability

    Mihaela Dimitrova, Samy Blusseau, and Santiago Velasco-Forero. Learning morphological representations of image transformations: In- fluence of initialization and layer differentiability. InInternational Conference on Discrete Geometry and Mathematical Morphology, pages 371–383, 2025

  2. [2]

    Exploiting subgradient sparsity in max-plus neural networks.hal-05502128 preprint, 2026

    Ikhlas Enaieh and Olivier Fercoq. Exploiting subgradient sparsity in max-plus neural networks.hal-05502128 preprint, 2026

  3. [3]

    Methods and applications of (max,+) linear algebra

    St ´ephane Gaubert and Max Plus. Methods and applications of (max,+) linear algebra. InAnnual symposium on theoretical aspects of computer science, pages 261–282. Springer, 1997

  4. [4]

    A survey of methods for explain- ing black box models.ACM computing surveys (CSUR), 51(5):1–42, 2018

    Riccardo Guidotti, Anna Monreale, Salvatore Ruggieri, Franco Turini, Fosca Giannotti, and Dino Pedreschi. A survey of methods for explain- ing black box models.ACM computing surveys (CSUR), 51(5):1–42, 2018

  5. [5]

    On calibration of modern neural networks

    Chuan Guo, Geoff Pleiss, Yu Sun, and Kilian Q Weinberger. On calibration of modern neural networks. InInternational conference on machine learning, pages 1321–1330. PMLR, 2017

  6. [6]

    Adam: A method for stochastic optimization

    Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. InProc. of ICLR, 2015

  7. [7]

    A unified approach to interpreting model predictions.Advances in neural information processing systems, 30, 2017

    Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions.Advances in neural information processing systems, 30, 2017

  8. [8]

    Min-max-plus neural networks.preprint arXiv:2102.06358, 2021

    Ye Luo and Shiqing Fan. Min-max-plus neural networks.preprint arXiv:2102.06358, 2021

  9. [9]

    Christoph Molnar.Interpretable machine learning. Lulu. com, 2020

  10. [10]

    Morphological network: How far can we go with morphological neurons? InBritish Machine Vision Conference, 2019

    Ranjan Mondal, Soumendu Sundar Mukherjee, Sanchayan Santra, and Bhabatosh Chanda. Morphological network: How far can we go with morphological neurons? InBritish Machine Vision Conference, 2019

  11. [11]

    Clustering by means of medoids

    LKPJ Rdusseeun and P Kaufman. Clustering by means of medoids. InProceedings of the statistical data analysis based on the L1 norm conference, neuchatel, switzerland, volume 31, page 28, 1987

  12. [12]

    An introduction to morphological neural networks

    Gerhard X Ritter and Peter Sussner. An introduction to morphological neural networks. InProceedings of 13th International Conference on Pattern Recognition, volume 4, pages 709–717. IEEE, 1996

  13. [13]

    Grad-CAM: Why did you say that?

    Ramprasaath R Selvaraju, Abhishek Das, Ramakrishna Vedantam, Michael Cogswell, Devi Parikh, and Dhruv Batra. Grad-cam: Why did you say that?arXiv preprint arXiv:1611.07450, 2016

  14. [14]

    A comparative evaluation of explainability techniques for image data

    Mykyta Skliarov, Radwa El Shawi, Chedia Dhaoui, and Nada Ahmed. A comparative evaluation of explainability techniques for image data. Scientific Reports, 15(1):41898, 2025

  15. [15]

    Axiomatic attribution for deep networks

    Mukund Sundararajan, Ankur Taly, and Qiqi Yan. Axiomatic attribution for deep networks. InInternational conference on machine learning, pages 3319–3328. PMLR, 2017

  16. [16]

    Medmnist v2-a large-scale lightweight benchmark for 2d and 3d biomedical image classification

    Jiancheng Yang, Rui Shi, Donglai Wei, Zequan Liu, Lin Zhao, Bilian Ke, Hanspeter Pfister, and Bingbing Ni. Medmnist v2-a large-scale lightweight benchmark for 2d and 3d biomedical image classification. Scientific data, 10(1):41, 2023. Fig. 1. Examples of SHAP and IntGrad explanations for the multi-layer perceptron (lighter means more important) Fig. 2. Ex...