pith. machine review for the scientific record. sign in

arxiv: 2605.06540 · v1 · submitted 2026-05-07 · 💻 cs.AI · cs.GT

Recognition: unknown

Ex Ante Evaluation of AI-Induced Idea Diversity Collapse

Authors on Pith no claims yet

Pith reviewed 2026-05-08 09:46 UTC · model grok-4.3

classification 💻 cs.AI cs.GT
keywords AI evaluationidea diversitycreative AIdiversity collapseex ante evaluationcrowding coefficienthuman-relative ratioLLM performance
0
0 comments X

The pith

Three frontier LLMs fall below idea diversity parity across creative tasks.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Creative outputs lose value when many people produce similar ideas, creating a blind spot for evaluating AI that improves single outputs but may crowd the population. The paper provides a framework to benchmark this AI-induced diversity collapse in advance, without any data from people actually using AI. It compares the spread of ideas generated solely by the model to those from unaided humans, deriving measures of extra crowding. Applied to short stories, slogans, and alternative uses tasks, the three models show less diversity than humans. The approach also reveals that adjusting how ideas are generated can lower the crowding.

Core claim

We introduce a human-relative framework for benchmarking AI-induced human diversity collapse without requiring human-AI interaction data. By modeling ideas as congestible resources, source-level crowding is identifiable from within-distribution comparisons of model-only generations and matched unaided human baselines, yielding an excess-crowding coefficient Δ and a human-relative diversity ratio ρ with ρ≥1 as the no-excess-crowding parity condition. Across short stories, marketing slogans, and alternative-uses tasks, three frontier LLMs fall below parity across crowding kernels, with estimates stabilizing with feasible model-only sample sizes. Generation-protocol variants show crowding can

What carries the argument

The excess-crowding coefficient Δ and human-relative diversity ratio ρ, which quantify source-level crowding via within-distribution comparisons between model-only and human baseline idea generations.

Load-bearing premise

That within-distribution comparisons between model-only generations and matched unaided human baselines can reliably identify source-level crowding without human-AI interaction data.

What would settle it

A real-world study comparing the actual spread and adoption success of ideas in populations using the AI versus those not using it, to check if the predicted excess crowding matches observed redundancy.

Figures

Figures reproduced from arXiv: 2605.06540 by Nafis Saami Azad, Raiyan Abdul Baten.

Figure 1
Figure 1. Figure 1: Human-relative diversity under the primary semantic kernel. Points show task-family estimates of ρb for each model; bars show 95% bootstrap intervals. The dashed line marks ρ = 1, the no-excess-crowding condition from Proposition 1. Full numeric estimates are reported in Appendix E. Models and protocols. Model generations are collected from GPT-5.4, Claude Sonnet 4.5, and Gemini 2.5 Flash under matched tas… view at source ↗
Figure 2
Figure 2. Figure 2: Finite-sample stability of semantic crowding estimates. Curves show κb(n) as a function of sampled responses n, averaged across conditions within each task family. Shaded bands show 95% intervals from repeated rarefaction samples. confidence bound is also below one. Thus, under the primary semantic kernel, all evaluated neutral model-conditioned source distributions introduce positive excess crowding relat… view at source ↗
Figure 3
Figure 3. Figure 3: Persona-mixture prompting improves human-relative diversity. Bars compare the neutral main protocol at T = 1.0 with a persona-mixture protocol at T = 1.0. Error bars show 95% bootstrap intervals. The dashed line marks ρ = 1, the no-excess-crowding condition. distinctiveness value γk. At X = 10, the threshold ranges from 73.3% to 96.4% of γk; by X = 25, every model-task curve exceeds 96% (see Appendix G). I… view at source ↗
Figure 4
Figure 4. Figure 4: plots κb H against κb A at the task-family level. All model-task points lie above the diagonal, indicating that AI outputs are more semantically crowded than the matched human baseline in every main benchmark comparison. 0.60 0.65 0.70 0.75 0.80 0.85 0.90 0.95 Human crowding, 0.60 0.65 0.70 0.75 0.80 0.85 0.90 0.95 AI crowding, Claude Sonnet 4.5 / aut GPT-5.4 / aut Gemini 2.5 Flash / aut Claude Sonnet 4.5 … view at source ↗
Figure 5
Figure 5. Figure 5: plots the normalized critical-benefit curves implied by the main semantic excess-crowding estimates view at source ↗
Figure 6
Figure 6. Figure 6: Story crowding under full-text and plot-synopsis kernels. Points show task-family estimates of ρb for each model; bars show 95% bootstrap intervals. The dashed line marks ρ = 1, the no-excess-crowding condition. All models remain below parity under the plot-synopsis kernel, indicating that story crowding persists when similarity is computed over narrative content rather than full prose. similarity, Claude … view at source ↗
Figure 7
Figure 7. Figure 7: Rarefaction curve for story plot-synopsis crowding. Curves show mean plot-synopsis crowding κb(n) as a function of sampled responses n, aggregated across story prompts. The dashed vertical line marks n = 50. The curves assess whether the narrative-level crowding estimates are stable with the available model-only sample size. GPT-5.4 Claude Sonnet 4.5 Gemini 2.5 Flash 0.0 0.2 0.4 0.6 0.8 1.0 Human-relative … view at source ↗
Figure 8
Figure 8. Figure 8: Slogan crowding under semantic and lexical-template kernels. Points show ρb for each model; bars show 95% bootstrap intervals. The dashed line marks ρ = 1, the no-excess-crowding condition. All models remain below parity under semantic, word-overlap, and character-trigram kernels. [0.833, 0.894], Claude Sonnet 4.5 has ρb = 0.715 with 95% CI [0.665, 0.759], and Gemini 2.5 Flash has ρb = 0.938 with 95% CI [0… view at source ↗
Figure 9
Figure 9. Figure 9: Rarefaction curve for slogan non-stopword Jaccard crowding. Curves show mean lexical crowding κb(n) as a function of sampled slogans n. The dashed vertical line marks n = 50. 0 10 20 30 40 50 Sample size n 0.0 0.2 0.4 0.6 0.8 1.0 M e a n le xical cro w din g (n) Slogan lexical-template saturation: char_trigram_jaccard Source/model Human GPT-5.4 Claude Sonnet 4.5 Gemini 2.5 Flash view at source ↗
Figure 10
Figure 10. Figure 10: Rarefaction curve for slogan character-trigram crowding. Curves show mean lexical crowding κb(n) as a function of sampled slogans n. The dashed vertical line marks n = 50 view at source ↗
Figure 11
Figure 11. Figure 11: AUT crowding under semantic and concept-bucket kernels. Points show task-family estimates of ρb for each model; bars show 95% bootstrap intervals. The dashed line marks ρ = 1, the no-excess-crowding condition. All models remain below parity under the concept-bucket kernel, indicating excess concept reuse relative to the matched human baseline. 10 20 30 40 50 Sample size, n 0.0 0.1 0.2 0.3 0.4 0.5 0.6 C o … view at source ↗
Figure 12
Figure 12. Figure 12: Rarefaction curve for AUT concept-bucket crowding. Curves show mean concept￾bucket crowding κb(n) as a function of sampled responses n, aggregated across AUT objects. The dashed vertical line marks n = 50. The curves assess whether the concept-level crowding estimates are stable with the available model-only sample size view at source ↗
Figure 13
Figure 13. Figure 13: Persona-mixture prompting lowers the critical private benefit required for rational adoption. Solid curves show the neutral main protocol at T = 1.0; dashed curves show the persona￾mixture protocol at T = 1.0. Curves plot Bcrit(X)/γ = 1 − exp(−X∆) b . Lower curves indicate weaker congestion externalities. 0.4 0.6 0.8 1.0 1.2 Temperature 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 H u m a n-relativ e div ersity, S… view at source ↗
Figure 14
Figure 14. Figure 14: Temperature-grid effects on human-relative diversity. Points show task-level ρb under the primary semantic kernel across available temperatures. Error bars show 95% bootstrap intervals. Higher temperature increases ρb from the lowest to highest tested temperature in all nine model-task combinations, although strict monotonicity holds in six of nine. 0.4 0.6 0.8 1.0 1.2 Temperature 0.12 0.14 0.16 0.18 0.20… view at source ↗
Figure 15
Figure 15. Figure 15: Temperature-grid effects on excess crowding. Points show task-level ∆b under the primary semantic kernel across available temperatures. Error bars show 95% bootstrap intervals. Higher temperature decreases ∆b from the lowest to highest tested temperature in all nine model-task combinations. 24 view at source ↗
Figure 16
Figure 16. Figure 16: Critical benefit curves under neutral main and best observed protocols. For each model-task pair, the solid curve shows the neutral main protocol and the dashed curve shows the best observed protocol among the tested temperature and persona-mixture settings. Curves plot Bcrit(X)/γ = 1 − exp(−X∆) b . This figure is descriptive: it shows the best protocol found in the tested grid, not an optimized global pr… view at source ↗
read the original abstract

Creative AI systems are typically evaluated at the level of individual utility, yet creative outputs are consumed in populations: an idea loses value when many others produce similar ones. This creates an evaluation blind spot, as AI can improve individual outputs while increasing population-level crowding. We introduce a human-relative framework for benchmarking AI-induced human diversity collapse without requiring human-AI interaction data, providing an ex ante protocol to estimate crowding risk from model-only generations and matched unaided human baselines. By modeling ideas as congestible resources, we show that source-level crowding is identifiable from within-distribution comparisons, yielding an excess-crowding coefficient $\Delta$ and a human-relative diversity ratio $\rho$. We show that $\rho\ge1$ is the no-excess-crowding parity condition and connect $\Delta$ to an adoption game with exposure-dependent redundancy costs. Across short stories, marketing slogans, and alternative-uses tasks, three frontier LLMs fall below parity across crowding kernels. Estimates stabilize with feasible model-only sample sizes. Importantly, generation-protocol variants show that crowding can be reduced through targeted design, making diversity collapse an actionable, development-time evaluation target for population-aware creative AI.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper introduces a human-relative ex ante framework for assessing AI-induced idea diversity collapse. It models ideas as congestible resources and extracts an excess-crowding coefficient Δ and human-relative diversity ratio ρ from within-distribution comparisons of model-only LLM generations versus matched unaided human baselines, without requiring human-AI interaction data. The paper claims that ρ ≥ 1 is the no-excess-crowding parity condition, connects Δ to an adoption game with exposure-dependent redundancy costs, and reports that three frontier LLMs fall below parity across short stories, marketing slogans, and alternative-uses tasks. It further shows that estimates stabilize with feasible sample sizes and that targeted generation-protocol variants can reduce crowding.

Significance. If the identification from within-distribution comparisons holds, the framework supplies a practical, interaction-free protocol for population-level evaluation of creative AI, addressing the blind spot between individual utility and collective crowding. The empirical demonstration that crowding is measurable and mitigable at development time is a concrete strength, as is the stabilization of estimates with model-only samples; these elements make diversity collapse an actionable design target rather than a purely theoretical concern.

major comments (3)
  1. [Abstract / central construction] Abstract and central construction: the claim that source-level crowding is identifiable from within-distribution comparisons (yielding Δ and ρ) rests on an un-derived equivalence between the chosen crowding kernels and the redundancy-cost function in the adoption game. No steps are shown establishing that kernel statistics alone determine exposure-dependent payoffs once human choice, selection, or context-dependent valuation are admitted; this assumption is load-bearing for the ex ante protocol.
  2. [Abstract] Abstract: the assertion that ρ ≥ 1 constitutes the no-excess-crowding parity condition and that Δ maps directly to the adoption game is presented without the intermediate equations or kernel definitions needed to verify independence from the same within-distribution statistics used to compute ρ, raising a circularity risk for the redundancy-cost parameterization.
  3. [Empirical results] Empirical results (across tasks): the report that three LLMs fall below parity and that estimates stabilize with feasible model-only sample sizes is given without error bars, sample-size justification, or robustness checks against kernel choice, which weakens the claim that the protocol yields reliable ex ante signals.
minor comments (2)
  1. Notation for Δ and ρ is introduced in the abstract but would benefit from an explicit early section defining the crowding kernels and their application to the two distributions.
  2. The manuscript would be strengthened by a brief discussion of how the framework relates to existing measures of semantic diversity or novelty in computational creativity literature.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive report and for recognizing the practical value of an interaction-free ex ante protocol. We address each major comment below. Where the manuscript requires additional derivation or empirical safeguards, we have revised accordingly.

read point-by-point responses
  1. Referee: [Abstract / central construction] Abstract and central construction: the claim that source-level crowding is identifiable from within-distribution comparisons (yielding Δ and ρ) rests on an un-derived equivalence between the chosen crowding kernels and the redundancy-cost function in the adoption game. No steps are shown establishing that kernel statistics alone determine exposure-dependent payoffs once human choice, selection, or context-dependent valuation are admitted; this assumption is load-bearing for the ex ante protocol.

    Authors: We agree that the mapping from kernel statistics to exposure-dependent payoffs must be derived explicitly rather than asserted. The revised manuscript adds a dedicated subsection that starts from a standard random-utility adoption model in which an agent’s payoff declines linearly with the expected number of near-duplicates encountered. We then show that the first two moments of the within-distribution similarity kernel are sufficient statistics for this expectation under the maintained assumption that agents observe only pairwise similarities (not full context-dependent valuations). The derivation is now presented before the definitions of Δ and ρ, making the identifiability claim traceable to the kernel rather than circular. revision: yes

  2. Referee: [Abstract] Abstract: the assertion that ρ ≥ 1 constitutes the no-excess-crowding parity condition and that Δ maps directly to the adoption game is presented without the intermediate equations or kernel definitions needed to verify independence from the same within-distribution statistics used to compute ρ, raising a circularity risk for the redundancy-cost parameterization.

    Authors: The revised text inserts the missing intermediate equations immediately after the kernel definitions. ρ is defined solely as the ratio of average pairwise distances in the human versus model-only distributions; Δ is obtained by substituting the same kernel into the closed-form redundancy-cost term of the adoption game and solving for the excess-crowding multiplier that equates expected payoffs. Because the parity condition ρ ≥ 1 is obtained by setting Δ = 0 in the game, the two quantities are algebraically linked but computed from distinct operations on the kernel; we now display both the algebraic link and the separate computational paths to eliminate any appearance of circularity. revision: yes

  3. Referee: [Empirical results] Empirical results (across tasks): the report that three LLMs fall below parity and that estimates stabilize with feasible model-only sample sizes is given without error bars, sample-size justification, or robustness checks against kernel choice, which weakens the claim that the protocol yields reliable ex ante signals.

    Authors: We accept the criticism. The revised empirical section now reports bootstrap standard errors for both Δ and ρ, includes convergence plots that justify the chosen sample sizes (n = 200 per condition), and adds a robustness table repeating the main results under three alternative kernels (cosine on sentence embeddings, Jaccard on n-grams, and edit-distance on tokenized ideas). All three LLMs remain below parity and the stabilization result is unchanged, but the added diagnostics directly address the concern about reliability. revision: yes

Circularity Check

1 steps flagged

Central metrics Δ and ρ extracted by construction from within-distribution comparisons; mapping to adoption game assumed without independent derivation

specific steps
  1. self definitional [Abstract]
    "By modeling ideas as congestible resources, we show that source-level crowding is identifiable from within-distribution comparisons, yielding an excess-crowding coefficient Δ and a human-relative diversity ratio ρ. We show that ρ≥1 is the no-excess-crowding parity condition and connect Δ to an adoption game with exposure-dependent redundancy costs."

    Δ and ρ are defined directly as outputs of applying crowding kernels to the model-only and human baseline distributions; the 'identifiability' claim, the parity condition ρ≥1, and the connection of Δ to the adoption game's redundancy costs are then asserted as shown results. No separate equation or external mapping is supplied demonstrating that the kernel statistics determine the game's exposure-dependent costs independently of the same distributional inputs, rendering the central construction equivalent to its modeling assumptions by definition.

full rationale

The paper's core claim is that source-level crowding is identifiable from within-distribution comparisons of model-only generations versus human baselines, directly yielding Δ and ρ, with ρ ≥ 1 declared the parity condition and Δ connected to an adoption game. This identification and connection rest on the modeling choice of ideas as congestible resources and the selected kernels as sufficient statistics, without a separate derivation showing that distributional overlap alone determines exposure-dependent redundancy costs once human choice and context are admitted. The empirical results across tasks add non-circular content, but the load-bearing identification step reduces to the input definitions.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 0 invented entities

Abstract-only review; full derivations of ρ and Δ, kernel specifications, and any fitted parameters are not visible. Modeling ideas as congestible resources is treated as a foundational modeling choice.

axioms (1)
  • domain assumption Ideas function as congestible resources whose value decreases with population-level similarity
    Explicitly invoked to justify the crowding model and the identifiability of source-level crowding from within-distribution comparisons.

pith-pipeline@v0.9.0 · 5499 in / 1299 out tokens · 47999 ms · 2026-05-08T09:46:00.965470+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

54 extracted references · 13 canonical work pages · 1 internal anchor

  1. [1]

    Assessing associative distance among ideas elicited by tests of divergent thinking.Creativity Research Journal, 26(2):229–238, 2014

    Selcuk Acar and Mark A Runco. Assessing associative distance among ideas elicited by tests of divergent thinking.Creativity Research Journal, 26(2):229–238, 2014

  2. [2]

    From semantic memory to collective creativity: A generative cognitive foundation for social creativity models

    Mirza Nayeem Ahmed and Raiyan Abdul Baten. From semantic memory to collective creativity: A generative cognitive foundation for social creativity models. InProceedings of the Annual Meeting of the Cognitive Science Society, volume 48, 2026

  3. [3]

    Social psychology of creativity: A consensual assessment technique

    Teresa M Amabile. Social psychology of creativity: A consensual assessment technique. Journal of Personality and Social Psychology, 43(5):997, 1982

  4. [4]

    Homogenization effects of Large Language Models on human creative ideation

    Barrett R Anderson, Jash Hemant Shah, and Max Kreminski. Homogenization effects of Large Language Models on human creative ideation. InProceedings of the 16th Conference on Creativity & Cognition, pages 413–425, 2024

  5. [5]

    How AI ideas affect the creativity, diversity, and evolution of human ideas: evidence from a large, dynamic experiment

    Joshua Ashkinaze, Julia Mendelsohn, Li Qiwei, Ceren Budak, and Eric Gilbert. How AI ideas affect the creativity, diversity, and evolution of human ideas: evidence from a large, dynamic experiment. InProceedings of the ACM Collective Intelligence Conference, pages 198–213, 2025

  6. [6]

    MuseScorer: Idea originality scoring at scale

    Ali Sarosh Bangash, Krish Veera, Ishfat Abrar Islam, and Raiyan Abdul Baten. MuseScorer: Idea originality scoring at scale. InProceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 19947–19965, 2025

  7. [7]

    Cues to gender and racial identity reduce creativity in diverse social networks.Scientific Reports, 11(1):10261, 2021

    Raiyan Abdul Baten, Richard N Aslin, Gourab Ghoshal, and Ehsan Hoque. Cues to gender and racial identity reduce creativity in diverse social networks.Scientific Reports, 11(1):10261, 2021

  8. [8]

    Goldilocks

    Raiyan Abdul Baten, Richard N Aslin, Gourab Ghoshal, and Ehsan Hoque. Novel idea generation in social networks is optimized by exposure to a “Goldilocks” level of idea-variability. PNAS Nexus, 1(5):pgac255, 2022

  9. [9]

    Creativity in temporal social networks: How divergent thinking is impacted by one’s choice of peers.Journal of the Royal Society Interface, 17(171):20200667, 2020

    Raiyan Abdul Baten, Daryl Bagley, Ashely Tenesaca, Famous Clark, James P Bagrow, Gourab Ghoshal, and Ehsan Hoque. Creativity in temporal social networks: How divergent thinking is impacted by one’s choice of peers.Journal of the Royal Society Interface, 17(171):20200667, 2020

  10. [10]

    AI can enhance creativity in social networks.arXiv preprint arXiv:2410.15264, 2024

    Raiyan Abdul Baten, Ali Sarosh Bangash, Krish Veera, Gourab Ghoshal, and Ehsan Hoque. AI can enhance creativity in social networks.arXiv preprint arXiv:2410.15264, 2024

  11. [11]

    Automating creativity assessment with SemDis: An open platform for computing semantic distance.Behavior Research Methods, 53(2):757–780, 2021

    Roger E Beaty and Dan R Johnson. Automating creativity assessment with SemDis: An open platform for computing semantic distance.Behavior Research Methods, 53(2):757–780, 2021

  12. [12]

    Scoring divergent thinking tests by computer with a semantics-based algorithm.Europe’s Journal of Psychology, 12(2):210, 2016

    Kenes Beketayev and Mark A Runco. Scoring divergent thinking tests by computer with a semantics-based algorithm.Europe’s Journal of Psychology, 12(2):210, 2016

  13. [13]

    A semantic network approach to the Creativity Quotient (CQ).Creativity Research Journal, 21(1):64–71, 2009

    Terry Bossomaier, Mike Harré, Anthony Knittel, and Allan Snyder. A semantic network approach to the Creativity Quotient (CQ).Creativity Research Journal, 21(1):64–71, 2009

  14. [14]

    The machines take over: A comparison of various supervised learning approaches for automated scoring of divergent thinking tasks.The Journal of Creative Behavior, 57(1):17–36, 2023

    Philip Buczak, He Huang, Boris Forthmann, and Philipp Doebler. The machines take over: A comparison of various supervised learning approaches for automated scoring of divergent thinking tasks.The Journal of Creative Behavior, 57(1):17–36, 2023

  15. [15]

    Can good writing be generative? expert-level AI writing emerges through fine-tuning on high quality books

    Tuhin Chakrabarty and Paramveer S Dhillon. Can good writing be generative? expert-level AI writing emerges through fine-tuning on high quality books. InProceedings of the 2026 CHI Conference on Human Factors in Computing Systems, pages 1–27, 2026

  16. [16]

    Art or artifice? Large Language Models and the false promise of creativity

    Tuhin Chakrabarty, Philippe Laban, Divyansh Agarwal, Smaranda Muresan, and Chien-Sheng Wu. Art or artifice? Large Language Models and the false promise of creativity. InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems, pages 1–34, 2024

  17. [17]

    Cognitive abilities involved in insight problem solving: An individual differences model.Creativity Research Journal, 20(3):278–290, 2008

    Colin G DeYoung, Joseph L Flanders, and Jordan B Peterson. Cognitive abilities involved in insight problem solving: An individual differences model.Creativity Research Journal, 20(3):278–290, 2008. 26

  18. [18]

    Generative AI enhances individual creativity but reduces the collective diversity of novel content.Science Advances, 10(28):eadn5290, 2024

    Anil R Doshi and Oliver P Hauser. Generative AI enhances individual creativity but reduces the collective diversity of novel content.Science Advances, 10(28):eadn5290, 2024

  19. [19]

    Understanding fluency and originality: A latent variable perspective.Thinking Skills and Creativity, 14:56–67, 2014

    Denis Dumas and Kevin N Dunbar. Understanding fluency and originality: A latent variable perspective.Thinking Skills and Creativity, 14:56–67, 2014

  20. [20]

    Denis Dumas, Peter Organisciak, and Michael Doherty. Measuring divergent thinking originality with human raters and text-mining models: A psychometric comparison of methods.Psychology of Aesthetics, Creativity, and the Arts, 15(4):645, 2021

  21. [21]

    Hierarchical neural story generation

    Angela Fan, Mike Lewis, and Yann Dauphin. Hierarchical neural story generation. InProceed- ings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 889–898, 2018

  22. [22]

    Boris Forthmann, Sue Hyeon Paek, Denis Dumas, Baptiste Barbot, and Heinz Holling. Scru- tinizing the basis of originality in divergent thinking tests: On the measurement precision of response propensity estimates.British Journal of Educational Psychology, 90(3):683–699, 2020

  23. [23]

    McGraw-Hill, 1967

    Joy Paul Guilford.The Nature of Human Intelligence. McGraw-Hill, 1967

  24. [24]

    The Curious Case of Neural Text Degeneration

    Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. The curious case of neural text degeneration.arXiv preprint arXiv:1904.09751, 2019

  25. [25]

    CreativityPrism: A holistic benchmark for Large Language Model creativity.arXiv preprint arXiv:2510.20091, 2025

    Zhaoyi Joey Hou, Bowei Alvin Zhang, Yining Lu, Bhiman Kumar Baghel, Anneliese Brei, Ximing Lu, Meng Jiang, Faeze Brahman, Snigdha Chaturvedi, Haw-Shiuan Chang, Daniel Khashabi, and Xiang Lorraine Li. CreativityPrism: A holistic benchmark for Large Language Model creativity.arXiv preprint arXiv:2510.20091, 2025

  26. [26]

    arXiv preprint arXiv:2411.02316 , year =

    Mete Ismayilzada, Claire Stevenson, and Lonneke van der Plas. Evaluating creative short story generation in humans and Large Language Models.arXiv preprint arXiv:2411.02316, 2024

  27. [27]

    The innovation trade-off: how following superstars shapes academic novelty

    Sean Kelty, Raiyan Abdul Baten, Adiba Mahbub Proma, Ehsan Hoque, Johan Bollen, and Gourab Ghoshal. The innovation trade-off: how following superstars shapes academic novelty. Humanities and Social Sciences Communications, 12(1):1–13, 2025

  28. [28]

    Understanding the effects of rlhf on llm generalisation and diversity.arXiv preprint arXiv:2310.06452,

    Robert Kirk, Ishita Mediratta, Christoforos Nalmpantis, Jelena Luketina, Eric Hambro, Edward Grefenstette, and Roberta Raileanu. Understanding the effects of RLHF on LLM generalisation and diversity.arXiv preprint arXiv:2310.06452, 2023

  29. [29]

    Human creativity in the age of LLMs: Randomized experiments on divergent and convergent thinking

    Harsh Kumar, Jonathan Vincentius, Ewan Jordan, and Ashton Anderson. Human creativity in the age of LLMs: Randomized experiments on divergent and convergent thinking. InProceedings of the 2025 CHI Conference on Human Factors in Computing Systems, pages 1–18, 2025

  30. [30]

    Jointly reinforcing diversity and quality in language model generations.arXiv preprint arXiv:2509.02534,

    Tianjian Li, Yiming Zhang, Ping Yu, Swarnadeep Saha, Daniel Khashabi, Jason Weston, Jack Lanchantin, and Tianlu Wang. Jointly reinforcing diversity and quality in language model generations.arXiv preprint arXiv:2509.02534, 2025

  31. [31]

    LLM-Eval: Unified multi-dimensional automatic evaluation for open-domain conversations with Large Language Models

    Yen-Ting Lin and Yun-Nung Chen. LLM-Eval: Unified multi-dimensional automatic evaluation for open-domain conversations with Large Language Models. InProceedings of the 5th Workshop on NLP for Conversational AI (NLP4ConvAI 2023), pages 47–58, 2023

  32. [32]

    Creativity in LLM-based multi-agent systems: A survey

    Yi-Cheng Lin, Kang-Chieh Chen, Zhe-Yan Li, Tzu-Heng Wu, Tzu-Hsuan Wu, Kuan-Yu Chen, Hung-yi Lee, and Yun-Nung Chen. Creativity in LLM-based multi-agent systems: A survey. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 27572–27595, 2025

  33. [33]

    Evaluating text-to-visual generation with image-to-text generation

    Zhiqiu Lin, Deepak Pathak, Baiqi Li, Jiayao Li, Xide Xia, Graham Neubig, Pengchuan Zhang, and Deva Ramanan. Evaluating text-to-visual generation with image-to-text generation. In European Conference on Computer Vision, pages 366–384. Springer, 2024

  34. [34]

    AI as humanity’s Salieri: Quantifying linguistic creativity of language models via systematic attribution of machine text against web text.arXiv preprint arXiv:2410.04265, 2024

    Ximing Lu, Melanie Sclar, Skyler Hallinan, Niloofar Mireshghallah, Jiacheng Liu, Seungju Han, Allyson Ettinger, Liwei Jiang, Khyathi Chandu, Nouha Dziri, et al. AI as humanity’s Salieri: Quantifying linguistic creativity of language models via systematic attribution of machine text against web text.arXiv preprint arXiv:2410.04265, 2024. 27

  35. [35]

    Preprint, arXiv:2407.01082

    Minh Nhat Nguyen, Andrew Baker, Clement Neo, Allen Roush, Andreas Kirsch, and Ravid Shwartz-Ziv. Turning up the heat: Min-p sampling for creative and coherent LLM outputs. arXiv preprint arXiv:2407.01082, 2024

  36. [36]

    Naming unrelated words predicts creativity.Proceedings of the National Academy of Sciences, 118(25):e2022340118, 2021

    Jay A Olson, Johnny Nahas, Denis Chmoulevitch, Simon J Cropper, and Margaret E Webb. Naming unrelated words predicts creativity.Proceedings of the National Academy of Sciences, 118(25):e2022340118, 2021

  37. [37]

    Beyond semantic distance: Automated scoring of divergent thinking greatly improves with Large Language Models.Thinking Skills and Creativity, 49:101356, 2023

    Peter Organisciak, Selcuk Acar, Denis Dumas, and Kelly Berthiaume. Beyond semantic distance: Automated scoring of divergent thinking greatly improves with Large Language Models.Thinking Skills and Creativity, 49:101356, 2023

  38. [38]

    Open creativity scoring

    Peter Organisciak and Denis Dumas. Open creativity scoring. https://openscoring.du. edu, 2020. [Computer software]

  39. [39]

    Deep associations, high creativity: A simple yet effective metric for evaluating Large Language Models

    Ziliang Qiu and Renfen Hu. Deep associations, high creativity: A simple yet effective metric for evaluating Large Language Models. InProceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 10870–10883, 2025

  40. [40]

    Scoring divergent thinking tests: A review and systematic framework.Psychology of Aesthetics, Creativity, and the Arts, 13(2):144, 2019

    Roni Reiter-Palmon, Boris Forthmann, and Baptiste Barbot. Scoring divergent thinking tests: A review and systematic framework.Psychology of Aesthetics, Creativity, and the Arts, 13(2):144, 2019

  41. [41]

    The standard definition of creativity.Creativity Research Journal, 24(1):92–96, 2012

    Mark A Runco and Garrett J Jaeger. The standard definition of creativity.Creativity Research Journal, 24(1):92–96, 2012

  42. [42]

    Scoring divergent thinking tests using total ideational output and a creativity index.Educational and Psychological Measurement, 52(1):213–221, 1992

    Mark A Runco and Wayne Mraz. Scoring divergent thinking tests using total ideational output and a creativity index.Educational and Psychological Measurement, 52(1):213–221, 1992

  43. [43]

    Evaluating the diversity and quality of LLM generated content.CoRR, abs/2504.12522, 2025

    Alexander Shypula, Shuo Li, Botong Zhang, Vishakh Padmakumar, Kayo Yin, and Osbert Bastani. Evaluating the diversity and quality of LLM generated content.arXiv preprint arXiv:2504.12522, 2025

  44. [44]

    A comprehensive analysis of Large Language Model outputs: Similarity, diversity, and bias.arXiv preprint arXiv:2505.09056, 2025

    Brandon Smith, Mohamed Reda Bouadjenek, Tahsin Alamgir Kheya, Phillip Dawson, and Sunil Aryal. A comprehensive analysis of Large Language Model outputs: Similarity, diversity, and bias.arXiv preprint arXiv:2505.09056, 2025

  45. [45]

    The Creativity Quotient: An objective scoring of ideational fluency.Creativity Research Journal, 16(4):415–419, 2004

    Allan Snyder, John Mitchell, Terry Bossomaier, and Gerry Pallier. The Creativity Quotient: An objective scoring of ideational fluency.Creativity Research Journal, 16(4):415–419, 2004

  46. [46]

    Short and extra-short forms of the Big Five Inventory–2: The BFI-2-S and BFI-2-XS.Journal of Research in Personality, 68:69–81, 2017

    Christopher J Soto and Oliver P John. Short and extra-short forms of the Big Five Inventory–2: The BFI-2-S and BFI-2-XS.Journal of Research in Personality, 68:69–81, 2017

  47. [47]

    The homogenizing effect of Large Language Models on human expression and thought.Trends in Cognitive Sciences, 2026

    Zhivar Sourati, Alireza S Ziabari, and Morteza Dehghani. The homogenizing effect of Large Language Models on human expression and thought.Trends in Cognitive Sciences, 2026

  48. [48]

    Stevenson, I

    C. Stevenson, I. Smal, M. Baas, M. Dahrendorf, R. Grasman, C. Tanis, E. Scheurs, D. Sleiffer, and H. van der Maas. Automated AUT scoring using a big data variant of the consensual assessment technique. Report Final Technical Report, Modeling Creativity Project, Universiteit van Amsterdam, Amsterdam, July 2020. Faculty of Social and Behavioural Sciences (F...

  49. [49]

    Nurture of creative talents.Theory Into Practice, 5(4):167–173, 1966

    E Paul Torrance. Nurture of creative talents.Theory Into Practice, 5(4):167–173, 1966

  50. [50]

    Diverse AI personas can mitigate the homogenization effect in human-AI collaborative ideation.Computers in Human Behavior: Artificial Humans, page 100289, 2026

    Yun Wan and Yoram M Kalman. Diverse AI personas can mitigate the homogenization effect in human-AI collaborative ideation.Computers in Human Behavior: Artificial Humans, page 100289, 2026

  51. [51]

    We’re different, we’re the same: Creative homogeneity across llms.arXiv preprint arXiv:2501.19361,

    Emily Wenger and Yoed Kenett. We’re different, we’re the same: Creative homogeneity across LLMs.arXiv preprint arXiv:2501.19361, 2025

  52. [52]

    Echoes in AI: Quantifying lack of plot diversity in LLM outputs.Proceedings of the National Academy of Sciences, 122(35):e2504966122, 2025

    Weijia Xu, Nebojsa Jojic, Sudha Rao, Chris Brockett, and Bill Dolan. Echoes in AI: Quantifying lack of plot diversity in LLM outputs.Proceedings of the National Academy of Sciences, 122(35):e2504966122, 2025. 28

  53. [53]

    The price of format: Diversity collapse in llms, 2025

    Longfei Yun, Chenyang An, Zilong Wang, Letian Peng, and Jingbo Shang. The price of format: Diversity collapse in LLMs.arXiv preprint arXiv:2505.18949, 2025

  54. [54]

    Noveltybench: Evaluating language models for humanlike diversity.arXiv preprint arXiv:2504.05228,

    Yiming Zhang, Harshita Diddee, Susan Holm, Hanchen Liu, Xinyue Liu, Vinay Samuel, Barry Wang, and Daphne Ippolito. NoveltyBench: Evaluating language models for humanlike diversity. arXiv preprint arXiv:2504.05228, 2025. 29