Recognition: unknown
Incremental learning for audio classification with Hebbian Deep Neural Networks
Pith reviewed 2026-05-10 03:37 UTC · model grok-4.3
The pith
Hebbian learning with kernel plasticity enables stable incremental audio classification, reaching 76.3% accuracy over five steps on ESC-50.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The central claim is that a Hebbian Deep Neural Network equipped with kernel plasticity, which selectively modulates network kernels to learn new information on some and retain previous knowledge on others, supports effective incremental learning for sound classification. Using the ESC-50 dataset, this yields 76.3% overall accuracy over five incremental steps, outperforming the 68.7% baseline without kernel plasticity while maintaining significantly greater stability across tasks.
What carries the argument
Kernel plasticity: selective modulation of specific network kernels during each incremental learning step to acquire new classes without overwriting old ones.
Load-bearing premise
That selectively modulating network kernels during incremental steps will reliably balance acquisition of new classes with retention of prior ones.
What would settle it
A controlled test on ESC-50 where kernel plasticity is disabled but the rest of the Hebbian training is identical, and the accuracy falls to the baseline level of 68.7% with reduced stability.
read the original abstract
The ability of humans for lifelong learning is an inspiration for deep learning methods and in particular for continual learning. In this work, we apply Hebbian learning, a biologically inspired learning process, to sound classification. We propose a kernel plasticity approach that selectively modulates network kernels during incremental learning, acting on selected kernels to learn new information and on others to retain previous knowledge. Using the ESC-50 dataset, the proposed method achieves 76.3% overall accuracy over five incremental steps, outperforming a baseline without kernel plasticity (68.7%) and demonstrating significantly greater stability across tasks.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript proposes a Hebbian Deep Neural Network augmented with a kernel plasticity mechanism that selectively modulates kernels during incremental learning steps to acquire new audio classes while retaining prior knowledge. Evaluated on the ESC-50 dataset partitioned into five incremental tasks, the method reports 76.3% overall accuracy, outperforming a baseline without kernel plasticity (68.7%) and exhibiting greater stability as measured by average accuracy drop across tasks.
Significance. If the empirical results hold under scrutiny, the work contributes a biologically inspired approach to continual learning for audio classification that mitigates catastrophic forgetting via selective kernel modulation. The use of a public dataset, explicit five-task split protocol, and defined stability metric provides a reproducible empirical benchmark that could inform future lifelong learning systems in signal processing.
major comments (1)
- [Section 4 (Results)] Section 4 (Results): The headline accuracies of 76.3% and 68.7% are reported as single point estimates without error bars, standard deviations from repeated runs, or statistical significance tests. This directly affects the strength of the central claim that the kernel plasticity mechanism delivers both superior acquisition and retention on the ESC-50 splits.
minor comments (2)
- [Abstract] Abstract: Performance numbers are stated without any reference to architecture details, training protocol, or the precise Hebbian threshold rule, which are supplied only in the body and reduce immediate accessibility.
- [Section 3.2] Section 3.2: The kernel selection rule based on Hebbian activation thresholds is described in prose; adding a compact algorithm box or pseudocode would improve clarity and exact reproducibility of the reported gap.
Simulated Author's Rebuttal
We thank the referee for the detailed review and the constructive comment on the empirical reporting in Section 4. We address the concern below and will incorporate the suggested improvements in the revised manuscript.
read point-by-point responses
-
Referee: The headline accuracies of 76.3% and 68.7% are reported as single point estimates without error bars, standard deviations from repeated runs, or statistical significance tests. This directly affects the strength of the central claim that the kernel plasticity mechanism delivers both superior acquisition and retention on the ESC-50 splits.
Authors: We agree that single-point estimates limit the robustness of the central claims. The current results reflect a single training run per method on the fixed five-task ESC-50 split. In the revised manuscript we will rerun both the proposed kernel-plasticity model and the baseline for five independent trials using different random seeds, report mean accuracy and standard deviation, and add a paired statistical test (e.g., Wilcoxon signed-rank) between the two methods to quantify significance of the observed 7.6 percentage-point gap and the stability improvement. revision: yes
Circularity Check
No significant circularity
full rationale
The paper reports an empirical incremental learning method using Hebbian plasticity on ESC-50 audio classification, with headline results consisting of measured accuracies (76.3 % overall, 68.7 % baseline) across five task splits. No mathematical derivation chain exists that reduces a claimed prediction to a fitted parameter or self-citation by construction; the kernel-selection rule is stated as an explicit algorithmic procedure whose performance is then measured against an ablated baseline. The work is therefore self-contained against external benchmarks and receives the default non-circularity finding.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
-
[1]
INTRODUCTION Continual learning is the behavior of artificial intelligence models to incrementally acquire new information and learn new patterns, showing robustness and resistance in terms of data distribution shift- ing and task change [1]. By default, deep learning models suffer from Catastrophic Forgetting, defined as the abrupt forgetting of pre- vio...
work page internal anchor Pith review Pith/arXiv arXiv 2026
-
[2]
Hebbian learning uses only the correlation between the samples to learn new information, thus not needing feedback information
HEBBIAN LEARNING FOR AUDIO CLASSIFICATION Hebbian learning is a principle describing associative learning, in which neurons strengthen their synaptic connections when they are active simultaneously. Hebbian learning uses only the correlation between the samples to learn new information, thus not needing feedback information. The model used in this work is...
-
[3]
EXPERIMENTAL SETUP AND EV ALUATION 3.1. Model training procedure For the experiments in this work we use the ESC-50 dataset [21], a labeled collection of 2000 environmental audio recordings routinely used for benchmarking methods in environmental sound classifica- tion. The dataset consists of 5-second-long recordings organized into 50 semantical classes ...
2000
-
[4]
The incremen- tal learning process is controlled by a number of hyperparameters selected using the validation set
activation function and other pooling solutions. The incremen- tal learning process is controlled by a number of hyperparameters selected using the validation set. The fraction of top kernelstop k we protect from overwriting is 0.6; the learning rate modifiers for plastic vs important kernelsαandβare 0.15 and 0.9, respectively; the interval (in batches) f...
-
[5]
We compare the proposed method with a system that does not use kernel plasticity (KP) in the training, but uses the multi-head Hebbian learning setup
RESULTS AND DISCUSSION Table 1 shows the classification accuracy of different learning vari- ants after each incremental stage. We compare the proposed method with a system that does not use kernel plasticity (KP) in the training, but uses the multi-head Hebbian learning setup. We also provide a EWC-based [3] baseline system. We also compare with systems ...
-
[6]
FM is always non- negative, with lower values indicating better retention
Forgetting Measure (FM) quantifies how much knowledge is lost on previously learned tasks after new tasks are introduced, compar- ing peak task performance to its final accuracy. FM is always non- negative, with lower values indicating better retention. The Intransi- Fig. 2. Comparison of the task-wise accuracy between using or not using KP in the increme...
-
[7]
The proposed neuro-modulation selectively regulates kernel plasticity, enabling the model to preserve past knowledge while adapting to new tasks
CONCLUSIONS This work introduced a biologically inspired solution to catastrophic forgetting by integrating kernel plasticity with Hebbian deep neu- ral networks for incremental audio classification. The proposed neuro-modulation selectively regulates kernel plasticity, enabling the model to preserve past knowledge while adapting to new tasks. Experiments...
-
[8]
Continual Learning and Catastrophic Forgetting,
G. M. v. d. Ven, N. Soures, and D. Kudithipudi, “Continual Learning and Catastrophic Forgetting,” inLearning and Mem- ory: A Comprehensive Reference. Elsevier, 2025, pp. 153– 168
2025
-
[9]
Prioritized Experience Replay
T. Schaul, J. Quan, I. Antonoglou, and D. Silver, “Prioritized Experience Replay.” ICLR, Feb. 2016
2016
-
[10]
Overcoming catastrophic forgetting in neural networks,
J. Kirkpatrick, R. Pascanu, N. Rabinowitz, J. Veness, G. Des- jardins, A. A. Rusu, K. Milan, J. Quan, T. Ramalho, A. Grabska-Barwinska, D. Hassabis, C. Clopath, D. Kumaran, and R. Hadsell, “Overcoming catastrophic forgetting in neural networks,”Proceedings of the National Academy of Sciences, vol. 114, no. 13, pp. 3521–3526, Mar. 2017
2017
-
[11]
Continual Learning Through Synaptic Intelligence,
F. Zenke, B. Poole, and S. Ganguli, “Continual Learning Through Synaptic Intelligence,”ICLR, pp. 3987–3995, Jun. 2017
2017
-
[12]
Pro- gressive Neural Networks,
A. A. Rusu, N. C. Rabinowitz, G. Desjardins, H. Soyer, J. Kirk- patrick, K. Kavukcuoglu, R. Pascanu, and R. Hadsell, “Pro- gressive Neural Networks,” Oct. 2022
2022
-
[13]
PathNet: Evolution Channels Gradient Descent in Super Neural Networks
C. Fernando, D. Banarse, C. Blundell, Y . Zwols, D. Ha, A. A. Rusu, A. Pritzel, and D. Wierstra, “PathNet: Evolution Chan- nels Gradient Descent in Super Neural Networks,” Jan. 2017, arXiv:1701.08734 [cs]
work page Pith review arXiv 2017
-
[14]
Learning without Forgetting,
Z. Li and D. Hoiem, “Learning without Forgetting,”IEEE Computer Society, vol. 40, no. 12, Feb. 2017
2017
-
[15]
Alleviat- ing catastrophic forgetting using context-dependent gating and synaptic stabilization,
N. Y . Masse, G. D. Grant, and D. J. Freedman, “Alleviat- ing catastrophic forgetting using context-dependent gating and synaptic stabilization,”Proceedings of the National Academy of Sciences, vol. 115, no. 44, Oct. 2018
2018
-
[16]
Model-Agnostic Meta- Learning for Fast Adaptation of Deep Networks,
C. Finn, P. Abbeel, and S. Levine, “Model-Agnostic Meta- Learning for Fast Adaptation of Deep Networks,” inProceed- ings of the 34th International Conference on Machine Learn- ing. PMLR, Jul. 2017, pp. 1126–1135, iSSN: 2640-3498
2017
-
[17]
Meta-Learning Representations for Continual Learning,
K. Javed and M. White, “Meta-Learning Representations for Continual Learning,” inAdvances in Neural Information Pro- cessing Systems, vol. 32. Curran Associates, Inc., 2019
2019
-
[18]
A Closer Look at Class- Incremental Learning for Multi-Label Audio Classification,
M. Mulimani and A. Mesaros, “A Closer Look at Class- Incremental Learning for Multi-Label Audio Classification,” IEEE Transactions on Audio, Speech and Language Process- ing, vol. 33, pp. 1293–1306, 2025
2025
-
[19]
Continual Lifelong Learning with Neural Networks: A Re- view,
G. I. Parisi, R. Kemker, J. L. Part, C. Kanan, and S. Wermter, “Continual Lifelong Learning with Neural Networks: A Re- view,”Neural Networks, vol. 113, pp. 54–71, May 2019
2019
-
[20]
T. J. E. K. J. Schwartz,Principles of Neural Science, 3rd ed. Elsevier, 1991
1991
-
[21]
Hebbian learning and develop- ment,
Y . Munakata and J. Pfaffly, “Hebbian learning and develop- ment,”Developmental Science, vol. 7, no. 2, pp. 141–148, Apr. 2004
2004
-
[22]
Adversarial Continual Learning,
S. Ebrahimi, F. Meier, R. Calandra, T. Darrell, and M. Rohrbach, “Adversarial Continual Learning,” inComputer Vision – ECCV 2020, A. Vedaldi, H. Bischof, T. Brox, and J.- M. Frahm, Eds., vol. 12356. Cham: Springer International Publishing, 2020, pp. 386–402, series Title: Lecture Notes in Computer Science
2020
-
[23]
Hebbian Deep Learning Without Feedback
A. Journ ´e, H. G. Rodriguez, Q. Guo, and T. Moraitis, “Hebbian Deep Learning Without Feedback.” ICLR, 2023
2023
-
[24]
Learning Multiple Layers of Features from Tiny Images,
A. Krizhevsky, “Learning Multiple Layers of Features from Tiny Images,” Tech. Rep., 2009
2009
-
[25]
Three scenarios for continual learning
G. M. v. d. Ven and A. S. Tolias, “Three scenarios for continual learning,” Apr. 2019, arXiv:1904.07734 [cs]
work page Pith review arXiv 2019
-
[26]
A comprehensive survey of continual learning: Theory, method and application, 2024 a
L. Wang, X. Zhang, H. Su, and J. Zhu, “A Comprehensive Sur- vey of Continual Learning: Theory, Method and Application,” Feb. 2024, arXiv:2302.00487 [cs]
-
[27]
Neuromodulated dopamine plas- tic networks for heterogeneous transfer learning with hebbian principle,
A. Magotra and J. Kim, “Neuromodulated dopamine plas- tic networks for heterogeneous transfer learning with hebbian principle,”Symmetry, vol. 13, no. 8, Aug. 2021, publisher: MDPI AG
2021
-
[28]
ESC: Dataset for Environmental Sound Classifi- cation,
K. J. Piczak, “ESC: Dataset for Environmental Sound Classifi- cation,” inProceedings of the 23rd ACM international confer- ence on Multimedia. Brisbane Australia: ACM, Oct. 2015, pp. 1015–1018
2015
-
[29]
Activa- tion functions in deep learning: A comprehensive survey and benchmark,
S. R. Dubey, S. K. Singh, and B. B. Chaudhuri, “Activa- tion functions in deep learning: A comprehensive survey and benchmark,”Neurocomputing, vol. 503, pp. 92–108, Sep. 2022
2022
-
[30]
Gradient Episodic Memory for Continual Learning,
D. Lopez-Paz and M. A. Ranzato, “Gradient Episodic Memory for Continual Learning,” inAdvances in Neural Information Processing Systems, vol. 30. Curran Associates, Inc., 2017
2017
-
[31]
Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence,
A. Chaudhry, P. K. Dokania, T. Ajanthan, and P. H. S. Torr, “Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence,” inComputer Vision – ECCV 2018, V . Ferrari, M. Hebert, C. Sminchisescu, and Y . Weiss, Eds. Cham: Springer International Publishing, 2018, vol. 11215, pp. 556–572, series Title: Lecture Notes in Computer Science
2018
-
[32]
AST: Audio Spectrogram Transformer,
Y . Gong, Y .-A. Chung, and J. Glass, “AST: Audio Spectrogram Transformer,” inInterspeech 2021. ISCA, Aug. 2021, pp. 571–575
2021
-
[33]
A Dataset and Tax- onomy for Urban Sound Research,
J. Salamon, C. Jacoby, and J. P. Bello, “A Dataset and Tax- onomy for Urban Sound Research,” inProceedings of the 22nd ACM international conference on Multimedia. Orlando Florida USA: ACM, Nov. 2014, pp. 1041–1044
2014
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.