Recognition: no theorem link
Initial results of the Digital Consciousness Model
Pith reviewed 2026-05-16 11:59 UTC · model grok-4.3
The pith
The Digital Consciousness Model finds evidence against 2024 LLMs being conscious but not decisively so, and weaker than for simpler AI.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The Digital Consciousness Model integrates evidence from a range of theories of consciousness into a shared probabilistic score. When applied to current large language models, the resulting score indicates that the evidence weighs against consciousness, yet the weight is not decisive enough to rule it out with high confidence. The corresponding score for simpler AI systems lies further against consciousness, establishing a comparative baseline within the model.
What carries the argument
The Digital Consciousness Model, a probabilistic aggregator that combines inputs from multiple independent theories of consciousness to produce comparable scores for AI and biological systems.
If this is right
- The same model can track whether evidence for consciousness rises or falls as new AI capabilities are added.
- Different AI architectures can be placed on a common scale for direct comparison of consciousness evidence.
- Simpler AI systems serve as a reference point showing stronger evidence against consciousness than current LLMs.
- Updates to the model can incorporate new data or revised theory weights without changing its basic structure.
Where Pith is reading between the lines
- Widespread use of the model could turn debates about AI consciousness into ongoing quantitative tracking exercises rather than one-off arguments.
- The framework suggests that ensemble-style aggregation of theories may be useful even in domains where individual theories remain contested.
- Testing the model's outputs against new empirical measures of consciousness indicators in AI would provide a direct check on its reliability.
Load-bearing premise
Different theories of consciousness can be combined into one overall probabilistic score without a single agreed definition or a validated way to weight them.
What would settle it
A future AI that produces coordinated positive indicators across several distinct theories of consciousness in a way that pushes the model's overall probability decisively above 50 percent would falsify the current assessment for that system.
Figures
read the original abstract
Artificially intelligent systems have become remarkably sophisticated. They hold conversations, write essays, and seem to understand context in ways that surprise even their creators. This raises a crucial question: Are we creating systems that are conscious? The Digital Consciousness Model (DCM) is a first attempt to assess the evidence for consciousness in AI systems in a systematic, probabilistic way. It provides a shared framework for comparing different AIs and biological organisms, and for tracking how the evidence changes over time as AI develops. Instead of adopting a single theory of consciousness, it incorporates a range of leading theories and perspectives - acknowledging that experts disagree fundamentally about what consciousness is and what conditions are necessary for it. This report describes the structure and initial results of the Digital Consciousness Model. Overall, we find that the evidence is against 2024 LLMs being conscious, but the evidence against 2024 LLMs being conscious is not decisive. The evidence against LLM consciousness is much weaker than the evidence against consciousness in simpler AI systems.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript introduces the Digital Consciousness Model (DCM), a probabilistic framework that aggregates assessments from multiple leading theories of consciousness (global workspace, IIT, higher-order, etc.) to evaluate the likelihood of consciousness in AI systems. It reports initial results concluding that the evidence is against 2024 LLMs being conscious but not decisively so, and that this evidence is substantially weaker than the evidence against consciousness in simpler AI systems.
Significance. If the aggregation method can be made transparent, reproducible, and robust to reasonable variations in inputs, the DCM could provide a useful shared benchmark for tracking how evidence for AI consciousness evolves with model capabilities, aiding both technical and policy discussions.
major comments (3)
- [DCM structure and initial results] The description of the DCM (structure and initial results sections): the weighting parameters and combination rule for aggregating per-theory evidence scores into the final probability are not specified, nor is any sensitivity analysis or justification provided. This directly undermines the central claim that the evidence is 'against but not decisive,' as different defensible weightings can shift the outcome across that threshold.
- [Initial results] Initial results section: no data sources, empirical basis, or explicit scoring procedure is given for the evidence scores assigned to each theory with respect to 2024 LLMs, leaving the reported probability without visible derivation or support.
- [Theory integration] Theory integration discussion: the manuscript acknowledges fundamental disagreements among theories but provides no operational definition of consciousness or rule for resolving conflicts in the probabilistic assessment, which is load-bearing for the 'not decisive' conclusion.
minor comments (2)
- [Abstract] Abstract: the repeated phrasing 'the evidence against 2024 LLMs being conscious' is redundant and could be streamlined for clarity.
- [Methods] The manuscript would benefit from including pseudocode or explicit equations for the DCM aggregation to support reproducibility.
Simulated Author's Rebuttal
We thank the referee for their constructive and detailed comments on the Digital Consciousness Model. The feedback highlights important areas for improving transparency and rigor, which we have addressed through revisions. We respond to each major comment below.
read point-by-point responses
-
Referee: The description of the DCM (structure and initial results sections): the weighting parameters and combination rule for aggregating per-theory evidence scores into the final probability are not specified, nor is any sensitivity analysis or justification provided. This directly undermines the central claim that the evidence is 'against but not decisive,' as different defensible weightings can shift the outcome across that threshold.
Authors: We agree that the original manuscript insufficiently specified the aggregation mechanics, which weakens support for the central claim. In the revised version, we have added a dedicated subsection on the DCM structure that explicitly defines the combination rule as a normalized weighted average of per-theory probabilities. Default weights are justified by reference to expert surveys in the consciousness literature (e.g., higher weight for global workspace and IIT due to their prominence). We also report a sensitivity analysis varying weights by up to ±30% and altering the combination rule to a product or majority-vote variant; in all cases the posterior for 2024 LLMs remains in the 0.25–0.45 range, preserving the 'against but not decisive' assessment. revision: yes
-
Referee: Initial results section: no data sources, empirical basis, or explicit scoring procedure is given for the evidence scores assigned to each theory with respect to 2024 LLMs, leaving the reported probability without visible derivation or support.
Authors: We accept that the initial presentation omitted the necessary derivation details. The revised manuscript adds an explicit scoring protocol in the main text and a supplementary appendix. Scores for each theory are derived by mapping 2024 LLM capabilities (e.g., context length, lack of persistent state, benchmark results on theory-of-mind and recurrence tasks) against the theory’s published criteria, with citations to the original theory papers and 2024 LLM evaluation reports. This now makes the numerical probabilities fully traceable. revision: yes
-
Referee: Theory integration discussion: the manuscript acknowledges fundamental disagreements among theories but provides no operational definition of consciousness or rule for resolving conflicts in the probabilistic assessment, which is load-bearing for the 'not decisive' conclusion.
Authors: The DCM is deliberately theory-neutral and therefore does not introduce a new operational definition; doing so would contradict its integrative purpose. We have expanded the integration section to clarify the conflict-resolution rule: each theory contributes an independent likelihood, and the final probability is their weighted combination. Disagreements among theories naturally pull the aggregate toward intermediate values (approximately 0.3 for 2024 LLMs), which is what produces the 'not decisive' outcome. This mechanism is now stated explicitly with a short illustrative calculation. revision: partial
Circularity Check
Probabilistic aggregation of consciousness theories lacks validated weighting or definition, rendering the 'against but not decisive' claim sensitive to arbitrary choices
specific steps
-
other
[Abstract / Model structure]
"Instead of adopting a single theory of consciousness, it incorporates a range of leading theories and perspectives - acknowledging that experts disagree fundamentally about what consciousness is and what conditions are necessary for it. ... Overall, we find that the evidence is against 2024 LLMs being conscious, but the evidence against 2024 LLMs being conscious is not decisive."
The single probability is produced by an aggregation step whose weighting and combination rule are not derived from any external benchmark or resolved definition; they are chosen by the model authors. The 'not decisive' verdict is therefore equivalent to the particular (unstated) weighting function applied to the same theory-by-theory inputs.
full rationale
The DCM derives its headline probability by combining per-theory evidence assessments (global workspace, IIT, higher-order, etc.) into a single score. No resolved operational definition of consciousness anchors the inputs, and the paper supplies no empirical, consensus, or cross-validated justification for its specific weights, normalization, or combination rule. Different defensible weighting schemes applied to the same per-theory assessments can move the final probability across the 'decisive' threshold, so the reported conclusion reduces to the authors' unstated aggregation choices rather than an independent derivation.
Axiom & Free-Parameter Ledger
free parameters (1)
- Theory weighting parameters
axioms (1)
- domain assumption Multiple leading theories of consciousness can be combined into a coherent probabilistic assessment despite fundamental disagreements
invented entities (1)
-
Digital Consciousness Model
no independent evidence
Forward citations
Cited by 1 Pith paper
-
AI and Consciousness: Shifting Focus Towards Tractable Questions
Direct research on AI consciousness is intractable, so the field should prioritize studying perceived AI consciousness and its societal consequences.
Reference graph
Works this paper leans on
-
[1]
Z. Z. Bronfman, S. Ginsburg, and E. Jablonka. The transition to minimal conscious- ness through the evolution of associative learning.Frontiers in psychology, 7:1954,
work page 1954
-
[2]
Consciousness in Artificial Intelligence: Insights from the Science of Consciousness
P. Butlin, R. Long, E. Elmoznino, Y. Bengio, J. Birch, A. Constant, R. VanRullen, et al. Consciousness in artificial intelligence: Insights from the science of conscious- ness.arXiv preprint arXiv:2308.08708,
work page internal anchor Pith review arXiv
-
[3]
L. Caviola and B. Saad. Futures with digital minds: Expert forecasts in 2025.arXiv preprint arXiv:2508.00536,
- [4]
-
[5]
C. Colombatto and S. M. Fleming. Folk psychological attributions of consciousness to large language models.Neuroscience of Consciousness, 2024(1):niae013,
work page 2024
-
[6]
Advances in Engineering Software42(12), 1020–1034 (2011)
doi: 10.1016/j. neuron.2011.03.018. Stanislas Dehaene, Jean-Pierre Changeux, Lionel Naccache, Jérôme Sackur, and Claire Sergent. Conscious, preconscious, and subliminal processing: A testable taxonomy.Trends in Cognitive Sciences, 10(5):204–211,
work page doi:10.1016/j 2011
-
[7]
doi: 10.1016/j.tics. 2006.03.007. D. C. Dennett.Consciousness Explained. Little, Brown and Company,
-
[8]
N. Dreksler, L. Caviola, D. Chalmers, C. Allen, A. Rand, J. Lewis, J. Sebo, et al. Sub- jective experience in AI systems: What do AI researchers and the public believe? arXiv preprint arXiv:2506.11945,
-
[9]
J. C. Francken, L. Beerendonk, D. Molenaar, J. J. Fahrenfort, J. D. Kiverstein, A. K. Seth, and S. Van Gaal. An academic survey on theoretical foundations, common assumptions and the current state of consciousness science.Neuroscience of Con- sciousness, 2022(1):niac011,
work page 2022
- [10]
- [11]
-
[12]
J. V. Pauketat, A. Ladak, and J. R. Anthis. Artificial intelligence, morality, and sentience (AIMS) survey: 2023 update. PsyArXiv,https://doi.org/10.31234/ osf.io/9xsav,
work page 2023
-
[13]
Maxwell JD Ramstead, Mahault Albarracin, Alex Kiefer, Brennan Klein, Chris Fields, Karl Friston, and Adam Safron. The inner screen model of consciousness: Applying the free energy principle directly to the study of conscious experience. arXiv preprint arXiv:2305.02205,
-
[14]
E. Schwitzgebel. Borderline consciousness, when it’s neither determinately true nor determinately false that experience is present.Philosophical Studies, 180(12):3415– 3439, 2023a. E. Schwitzgebel. The full rights dilemma for AI systems of debatable personhood. arXiv preprint arXiv:2303.17509, 2023b. E. Schwitzgebel. Against designing "safe" and "aligned"...
-
[15]
118 E. Schwitzgebel and J. Sebo. The emotional alignment design policy.arXiv preprint arXiv:2507.06263,
-
[16]
A. E. Scott, D. Neumann, J. Niess, and P. W. Woźniak. Do you mind? user per- ceptions of machine consciousness. InProceedings of the 2023 CHI conference on human factors in computing systems, pages 1–19, April
work page 2023
-
[17]
J. Sebo and R. Long. Moral consideration for AI systems by 2030.AI and Ethics, 5 (1):591–606,
work page 2030
-
[18]
Artificial consciousness: a perspective from the free energy principle
119 Wanja Wiese. Artificial consciousness: a perspective from the free energy principle. Philosophical Studies, 181(8):1947–1970,
work page 1947
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.