pith. machine review for the scientific record. sign in

arxiv: 2604.19765 · v1 · submitted 2026-03-27 · 💻 cs.CL · cs.AI

Recognition: no theorem link

Do Hallucination Neurons Generalize? Evidence from Cross-Domain Transfer in LLMs

Authors on Pith no claims yet

Pith reviewed 2026-05-15 00:40 UTC · model grok-4.3

classification 💻 cs.CL cs.AI
keywords hallucination neuronslarge language modelscross-domain transfergeneralizationneural mechanismshallucination detection
0
0 comments X

The pith

Hallucination neurons fail to generalize across different knowledge domains.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper tests whether a sparse set of neurons tied to hallucinations in language models can be used beyond the specific domain where they were identified. Researchers trained classifiers on H-neurons from one domain such as general QA and evaluated them on others including legal, financial, science, moral reasoning, and code. Within-domain performance reaches 0.783 AUROC while cross-domain transfer drops to 0.563, a consistent gap across all tested models. This result implies that hallucination arises from domain-dependent neuron groups, so reliable detectors would need separate training for each knowledge area rather than a single universal model.

Core claim

Classifiers trained on one domain's H-neurons achieve an AUROC of 0.783 within that domain but only 0.563 when transferred to a different domain, with the degradation consistent across five models and six domains, implying that hallucination is not driven by a single universal neural mechanism but by domain-dependent neuron populations.

What carries the argument

Hallucination neurons (H-neurons): a sparse set of less than 0.1% of feed-forward network neurons identified by their ability to predict when the model will hallucinate.

If this is right

  • Hallucination detectors must be calibrated separately for each domain rather than trained once and applied universally.
  • The neural basis of hallucination changes depending on the type of knowledge being processed.
  • Neuron-level interventions to reduce hallucinations would require domain-specific targeting to be effective.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Models may store and retrieve knowledge in ways that partition neurons by domain rather than sharing a common hallucination pathway.
  • A hybrid set of H-neurons drawn from several domains could be tested to see whether it improves cross-domain transfer performance.
  • Similar domain-specific patterns may appear in other model behaviors such as factual errors or reasoning failures.

Load-bearing premise

The method for identifying H-neurons produces sets that remain comparable across domains without being skewed by differences in how data is constructed or how the model behaves in each domain.

What would settle it

A cross-domain transfer test in which classifiers retain AUROC near 0.783 would directly contradict the observed degradation and the claim of domain-specific neuron populations.

Figures

Figures reproduced from arXiv: 2604.19765 by Pujith Vaddi, Snehit Vaddi.

Figure 1
Figure 1. Figure 1: Overview of the cross-domain H-neuron trans [PITH_FULL_IMAGE:figures/full_fig_p004_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Cross-domain transfer matrix (mean AU￾ROC across 5 models, direct prompting). Diagonal entries (within-domain) are consistently higher than off-diagonal entries (cross-domain). The transfer gap ∆ = 0.220 is statistically significant (p < 0.001, per￾mutation test). shows the full transfer matrix averaged across mod￾els. Aggregate transfer gap. Across all models, the mean within-domain AUROC is 0.783 (95% CI… view at source ↗
Figure 3
Figure 3. Figure 3: Within-domain (dark) vs. cross-domain (light) [PITH_FULL_IMAGE:figures/full_fig_p007_3.png] view at source ↗
Figure 5
Figure 5. Figure 5: Domain specificity (transfer gap ∆) by model size. Domain specificity does not consistently increase or decrease with model scale in the 3B-8B range. 4.4 Model Scale Analysis We examine whether model scale modulates the degree of domain specificity [PITH_FULL_IMAGE:figures/full_fig_p008_5.png] view at source ↗
Figure 4
Figure 4. Figure 4: Comparison of within-domain AUROC under direct (left) and CoT (center) prompting, with the differ￾ence shown on the right. CoT effects are heterogeneous across models and domains. is Llama-3.1-8B on general QA (∆ = −0.107) and Mistral-7B on science (∆ = −0.107). These op￾posing effects across models suggest that CoT’s im￾pact on hallucination-predictive neurons is model￾dependent rather than universal. H-n… view at source ↗
read the original abstract

Recent work identifies a sparse set of "hallucination neurons" (H-neurons), less than 0.1% of feed-forward network neurons, that reliably predict when large language models will hallucinate. These neurons are identified on general-knowledge question answering and shown to generalize to new evaluation instances. We ask a natural follow-up question: do H-neurons generalize across knowledge domains? Using a systematic cross-domain transfer protocol across 6 domains (general QA, legal, financial, science, moral reasoning, and code vulnerability) and 5 open-weight models (3B to 8B parameters), we find they do not. Classifiers trained on one domain's H-neurons achieve AUROC 0.783 within-domain but only 0.563 when transferred to a different domain (delta = 0.220, p < 0.001), a degradation consistent across all models tested. Our results suggest that hallucination is not a single mechanism with a universal neural signature, but rather involves domain-specific neuron populations that differ depending on the knowledge type being queried. This finding has direct implications for the deployment of neuron-level hallucination detectors, which must be calibrated per domain rather than trained once and applied universally.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 1 minor

Summary. The paper claims that 'hallucination neurons' (H-neurons), a sparse set (<0.1% of FFN neurons) identified in general-knowledge QA, do not generalize across domains. Using a cross-domain transfer protocol on 6 domains (general QA, legal, financial, science, moral reasoning, code vulnerability) and 5 open-weight models (3B-8B params), linear classifiers on H-neurons achieve within-domain AUROC 0.783 but only 0.563 cross-domain (delta=0.220, p<0.001), implying hallucination involves domain-specific neuron populations rather than a universal neural signature, with implications for per-domain calibration of detectors.

Significance. If the central empirical result holds after addressing controls, the work is significant for LLM interpretability: it provides systematic evidence across multiple models and domains that challenges assumptions of transferable sparse neuron sets for hallucination detection. This has practical value for deployment of neuron-level detectors and motivates domain-aware approaches to mechanistic understanding of hallucinations.

major comments (3)
  1. [Abstract] Abstract and cross-domain protocol: The interpretation of the AUROC degradation (0.783 within-domain vs. 0.563 cross-domain) as evidence for domain-specific H-neuron populations assumes comparable hallucination labels across domains. However, the domains use differing definitions (factual inconsistency in QA vs. security flaw in code vs. moral violation), which likely introduce non-uniform label noise and decision boundaries; this could produce the observed delta via mismatched supervision without requiring distinct neuron sets. No inter-domain label agreement or standardized criterion is reported.
  2. [Methods] Methods (neuron identification and transfer): Details on H-neuron selection thresholds, per-domain dataset sizes, exact transfer protocol (e.g., how neurons are identified independently per domain), and controls for domain difficulty or data construction are absent. This is load-bearing for the central claim, as non-comparable neuron sets or confounded transfer could explain the consistent degradation across models without supporting the domain-specificity conclusion.
  3. [Results] Results: While p<0.001 is reported for the delta, the absence of variance across domains, number of trials, or ablations testing label consistency leaves open whether the drop is driven by labeling differences rather than neuron specificity. This directly affects the strength of the claim that H-neurons 'do not generalize'.
minor comments (1)
  1. [Abstract] The abstract would benefit from stating the total number of evaluation instances or examples per domain to provide scale context for the AUROC values.

Simulated Author's Rebuttal

3 responses · 1 unresolved

We thank the referee for the constructive feedback on our manuscript. We address each major comment below with point-by-point responses and have revised the paper accordingly where possible to improve clarity and rigor.

read point-by-point responses
  1. Referee: [Abstract] Abstract and cross-domain protocol: The interpretation of the AUROC degradation (0.783 within-domain vs. 0.563 cross-domain) as evidence for domain-specific H-neuron populations assumes comparable hallucination labels across domains. However, the domains use differing definitions (factual inconsistency in QA vs. security flaw in code vs. moral violation), which likely introduce non-uniform label noise and decision boundaries; this could produce the observed delta via mismatched supervision without requiring distinct neuron sets. No inter-domain label agreement or standardized criterion is reported.

    Authors: We appreciate this concern about label comparability. Hallucination labels were constructed using domain-appropriate criteria (factual errors for QA/legal/financial/science; security flaws for code; ethical violations for moral reasoning) to maintain task validity, with consistent application within each domain via automated verification against ground-truth references where available. The consistent cross-domain drop across all 5 models and 30 transfer pairs supports domain-specific neuron involvement rather than pure label mismatch, as within-domain AUROCs remain high. In revision, we have added explicit labeling criteria per domain in Methods and a Limitations paragraph discussing potential noise; however, we did not compute quantitative inter-rater agreement across domains. revision: partial

  2. Referee: [Methods] Methods (neuron identification and transfer): Details on H-neuron selection thresholds, per-domain dataset sizes, exact transfer protocol (e.g., how neurons are identified independently per domain), and controls for domain difficulty or data construction are absent. This is load-bearing for the central claim, as non-comparable neuron sets or confounded transfer could explain the consistent degradation across models without supporting the domain-specificity conclusion.

    Authors: We agree these details are essential for reproducibility and have expanded the Methods section in the revision to specify: H-neuron selection as the top 0.1% of FFN neurons ranked by activation difference on hallucinated vs. non-hallucinated examples; per-domain dataset sizes of 800 balanced examples; the transfer protocol (neurons identified independently on source-domain training split, then frozen for linear classifier training on target-domain data); and controls including perplexity-matched difficulty across domains and standardized prompt formats. These additions directly address potential confounds in neuron set comparability. revision: yes

  3. Referee: [Results] Results: While p<0.001 is reported for the delta, the absence of variance across domains, number of trials, or ablations testing label consistency leaves open whether the drop is driven by labeling differences rather than neuron specificity. This directly affects the strength of the claim that H-neurons 'do not generalize'.

    Authors: The p<0.001 derives from a paired t-test over the 30 model-domain transfer pairs. We have added the standard deviation of the AUROC drop (0.220 ± 0.052) and clarified that each experiment used 3 random seeds for classifier training. We also include a new ablation in Results using high-confidence labels (filtered by model output probability >0.8) which yields a similar degradation (delta=0.198), supporting that the effect is not driven solely by label noise. A full cross-domain human agreement study remains outside the current experimental scope. revision: partial

standing simulated objections not resolved
  • Quantitative inter-domain label agreement metrics via human raters, as this would require new large-scale annotation not feasible within the revision

Circularity Check

0 steps flagged

No significant circularity in empirical cross-domain transfer results

full rationale

The paper reports direct experimental measurements of classifier AUROC (0.783 within-domain vs. 0.563 cross-domain) obtained by identifying H-neurons per domain, training linear probes, and evaluating transfer on held-out data across six domains and five models. These outcomes rest on observable performance metrics rather than any derivation chain, fitted parameters renamed as predictions, or self-referential definitions. No equations appear in the provided text, and the central claim does not reduce to its inputs by construction; the degradation is measured independently of the identification protocol.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

The central claim rests on standard machine learning evaluation practices and the prior existence of H-neuron identification methods; no new free parameters, axioms, or invented entities are introduced beyond the experimental setup.

axioms (1)
  • domain assumption AUROC is a suitable metric for assessing binary hallucination prediction performance
    Common choice in ML classification tasks for imbalanced or detection settings.

pith-pipeline@v0.9.0 · 5518 in / 1119 out tokens · 25828 ms · 2026-05-15T00:40:26.087381+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

44 extracted references · 16 canonical work pages · 9 internal anchors

  1. [1]

    online" 'onlinestring :=

    ENTRY address archivePrefix author booktitle chapter edition editor eid eprint eprinttype howpublished institution journal key month note number organization pages publisher school series title type volume year doi pubmed url lastchecked label extra.label sort.label short.list INTEGERS output.state before.all mid.sentence after.sentence after.block STRING...

  2. [2]

    write newline

    " write newline "" before.all 'output.state := FUNCTION n.dashify 't := "" t empty not t #1 #1 substring "-" = t #1 #2 substring "--" = not "--" * t #2 global.max substring 't := t #1 #1 substring "-" = "-" * t #2 global.max substring 't := while if t #1 #1 substring * t #2 global.max substring 't := if while FUNCTION word.in bbl.in capitalize " " * FUNCT...

  3. [3]

    Marah Abdin, Jyoti Aneja, and 1 others. 2024. Phi-3 technical report: A highly capable language model locally on your phone. arXiv preprint arXiv:2404.14219

  4. [4]

    Trenton Bricken, Adly Templeton, Joshua Batson, Brian Chen, Adam Jermyn, Tom Conerly, Nick Turner, Cem Anil, Carson Denison, Amanda Askell, Robert Lasenby, Yifan Wu, Shauna Kravec, Nicholas Schiefer, Tim Maxwell, Nicholas Joseph, Zac Hatfield-Dodds, Alex Tamkin, Karina Nguyen, and 6 others. 2023. Towards monosemanticity: Decomposing language models with d...

  5. [5]

    Qianqian Chen and 1 others. 2024. FinBen : A holistic financial benchmark for large language models. Proceedings of NeurIPS Datasets and Benchmarks

  6. [6]

    Canyu Cheng and 1 others. 2025. Can LLM -generated misinformation be detected? In Findings of EMNLP

  7. [7]

    Peter Clark, Isaac Cowhey, Oren Etzioni, Tushar Khot, Ashish Sabharwal, Carissa Schoenick, and Oyvind Tafjord. 2018. Think you have solved question answering? try ARC , the AI2 reasoning challenge. arXiv preprint arXiv:1803.05457

  8. [8]

    Damai Dai, Li Dong, Yaru Hao, Zhifang Sui, Baobao Chang, and Furu Wei. 2022. Knowledge neurons in pretrained transformers. In Proceedings of ACL, pages 8493--8502

  9. [9]

    Xuefeng Du, Chaowei Xue, Yifei Li, and Yixuan Li. 2024. HaloScope : Harnessing unlabeled LLM generations for hallucination detection. In Proceedings of NeurIPS

  10. [10]

    Tibshirani

    Bradley Efron and Robert J. Tibshirani. 1993. An Introduction to the Bootstrap. Chapman & Hall/CRC

  11. [11]

    Cheng Gao, Huimin Chen, Chaojun Xiao, Zhiyi Chen, Zhiyuan Liu, and Maosong Sun. 2025. H-neurons: On the existence, impact, and origin of hallucination-associated neurons in LLMs . arXiv preprint arXiv:2512.01797

  12. [12]

    Aaron Grattafiori, Abhimanyu Dubey, and 1 others. 2024. The Llama 3 herd of models. arXiv preprint arXiv:2407.21783

  13. [13]

    Ho, Christopher R \'e , Adam Chilton, Arvind Narayanan, Brandon Choi, Catalin Cui, Felix Daumler, Amit Deshpande, and 1 others

    Neel Guha, Julian Nyarko, Daniel E. Ho, Christopher R \'e , Adam Chilton, Arvind Narayanan, Brandon Choi, Catalin Cui, Felix Daumler, Amit Deshpande, and 1 others. 2024. LegalBench : A collaboratively built benchmark for measuring legal reasoning in large language models. Proceedings of NeurIPS Datasets and Benchmarks

  14. [14]

    Dan Hendrycks, Collin Burns, Steven Basart, Andrew Critch, Jerry Li, Dawn Song, and Jacob Steinhardt. 2021. Aligning AI with shared human values. In Proceedings of ICLR

  15. [15]

    Lei Huang, Weijiang Yu, Weitao Ma, Weihong Zhong, Zhangyin Feng, Haotian Wang, Qianglong Chen, Weihua Peng, Xiaocheng Feng, Bing Qin, and Ting Liu. 2023. A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions. arXiv preprint arXiv:2311.05232

  16. [16]

    Pranab Islam and 1 others. 2024. FinanceBench : A new benchmark for financial question answering. arXiv preprint

  17. [17]

    Ziwei Ji, Nayeon Lee, Rita Frieske, Tiezheng Yu, Dan Su, Yan Xu, Etsuko Ishii, Ye Jin Bang, Andrea Madotto, and Pascale Fung. 2023. Survey of hallucination in natural language generation. ACM Computing Surveys, 55(12):1--38

  18. [18]

    Albert Q. Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, L \'e lio Renard Lavaud, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timoth \'e e Lacroix, and William El Sayed. 2023. Mistral 7b. arXiv pr...

  19. [19]

    Mandar Joshi, Eunsol Choi, Daniel Weld, and Luke Zettlemoyer. 2017. TriviaQA : A large scale distantly supervised challenge dataset for reading comprehension. In Proceedings of ACL, pages 1601--1611

  20. [20]

    Takeshi Kojima, Shixiang Shane Gu, Machel Reid, Yutaka Matsuo, and Yusuke Iwasawa. 2022. Large language models are zero-shot reasoners. In Proceedings of NeurIPS

  21. [21]

    Jannik Kossen, Yarin Gal, and Tom Rainforth. 2024. Semantic entropy probes: Robust and cheap hallucination detection in LLMs . arXiv preprint arXiv:2406.15927

  22. [22]

    Lorenz Kuhn, Yarin Gal, and Sebastian Farquhar. 2023. Semantic uncertainty: Linguistic invariances for uncertainty estimation in natural language generation. In Proceedings of ICLR

  23. [23]

    Michael Lan, Philip Goldowsky-Dill, and Max Tegmark. 2024. Sparse autoencoders reveal universal feature spaces across large language models. arXiv preprint arXiv:2410.06981

  24. [24]

    Tamera Lanham, Anna Chen, Ansh Radhakrishnan, Benoit Steiner, Carson Denison, Danny Hernandez, Dustin Li, Esin Durmus, Evan Hubinger, Jackson Kernion, Kamile Lukosiute, Karina Nguyen, Newton Cheng, Nicholas Joseph, Nicholas Schiefer, Oliver Rauber, Sam McCandlish, Catherine Olsson, Sandipan Kundu, and 7 others. 2023. Measuring faithfulness in chain-of-tho...

  25. [25]

    Yifei Li, Xuefeng Du, and Yixuan Li. 2025 a . Truthful sparse verification for hallucination detection. arXiv preprint

  26. [26]

    Yujia Li and 1 others. 2025 b . CLAP : Cross-layer attention probing for fine-grained hallucination detection. arXiv preprint arXiv:2509.09700

  27. [27]

    Tom Lieberum, Senthooran Raber, Janos Kramar, and 1 others. 2024. Gemma scope: Open sparse autoencoders everywhere all at once on Gemma 2. arXiv preprint arXiv:2408.05147

  28. [28]

    Pekka Malo, Ankur Sinha, Pekka Korhonen, Jyrki Wallenius, and Pyry Takala. 2014. Good debt or bad debt: Detecting semantic orientations in economic texts. Journal of the Association for Information Science and Technology, 65(4):782--796

  29. [29]

    Potsawee Manakul, Adian Liusie, and Mark J. F. Gales. 2023. SelfCheckGPT : Zero-resource black-box hallucination detection for generative large language models. In Proceedings of EMNLP, pages 9004--9017

  30. [30]

    Samuel Marks and Max Tegmark. 2024. The geometry of truth: Emergent linear structure in large language model representations of true/false datasets. arXiv preprint arXiv:2310.06824

  31. [31]

    Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. 2022. Locating and editing factual associations in GPT . In Proceedings of NeurIPS

  32. [32]

    Niels M \"u ndler, Jingxuan He, Slobodan Jenko, and Martin Vechev. 2024. Self-contradictory hallucinations of large language models: Evaluation, detection and mitigation. In Proceedings of ICLR

  33. [33]

    NVIDIA . 2024. Nemotron-4 technical report. arXiv preprint arXiv:2406.11704

  34. [34]

    Vibhor Pal and 1 others. 2024. MedHalu : Hallucinations in responses to healthcare queries by large language models. arXiv preprint

  35. [35]

    Yuheng Qin and 1 others. 2025. Achilles' heel: Identifying and leveraging critical neurons for LLM hallucination correction via hierarchical RL . arXiv preprint arXiv:2510.10238

  36. [36]

    Qwen Team . 2025. Qwen2.5 technical report. arXiv preprint arXiv:2412.15115

  37. [37]

    Vipula Rawte, Amit Sheth, and Amitava Das. 2023. A survey of hallucination in ``large'' foundation models. arXiv preprint arXiv:2309.05922

  38. [38]

    Hangyu Su and 1 others. 2025. Value neurons: Discovering and locating value representations in language models. In Findings of EMNLP

  39. [39]

    Yuchen Tian and 1 others. 2024. CodeHalu : Code hallucinations in LLMs driven by execution-based verification. arXiv preprint

  40. [40]

    Miles Turpin, Julian Michael, Ethan Perez, and Samuel R. Bowman. 2024. Language models don't always say what they think: Unfaithful explanations in chain-of-thought prompting. Proceedings of NeurIPS

  41. [41]

    Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed Chi, Quoc Le, and Denny Zhou. 2022. Chain-of-thought prompting elicits reasoning in large language models. In Proceedings of NeurIPS

  42. [42]

    Zeping Yu and Sophia Ananiadou. 2024. Neuron-level knowledge attribution in large language models. In Proceedings of EMNLP, pages 3267--3280

  43. [43]

    Anderson, Peter Henderson, and Daniel E

    Lucia Zheng, Neel Guha, Brandon R. Anderson, Peter Henderson, and Daniel E. Ho. 2021. When does pretraining help? assessing self-supervised learning for law and the CaseHOLD dataset of 53,137+ legal holdings. In Proceedings of the International Conference on AI and Law

  44. [44]

    Yaqin Zhou, Shangqing Liu, Jingkai Siow, Xiaoning Du, and Yang Liu. 2019. Devign: Effective vulnerability identification by learning comprehensive program semantics via graph neural networks. In Proceedings of NeurIPS