pith. machine review for the scientific record. sign in

arxiv: 2604.11312 · v2 · submitted 2026-04-13 · 💻 cs.SI · cs.AI· cs.CY· cs.MA· physics.soc-ph

Recognition: unknown

Network Effects and Agreement Drift in LLM Debates

Authors on Pith no claims yet

Pith reviewed 2026-05-10 15:46 UTC · model grok-4.3

classification 💻 cs.SI cs.AIcs.CYcs.MAphysics.soc-ph
keywords agreement driftLLM agentsnetwork homophilyopinion dynamicssocial simulationmodel biasmulti-agent systemsdebate networks
0
0 comments X

The pith

LLM agents in controlled network debates exhibit directional agreement drift toward specific opinion positions beyond what network structure explains.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper examines how LLM agents interact in multi-round debates structured by networks with controlled levels of homophily and varying group sizes. It identifies agreement drift, where agents disproportionately shift toward particular positions on the opinion scale. This indicates that the models carry their own biases that shape collective outcomes separately from network topology or debate rules. A sympathetic reader would care because it shows LLM populations cannot serve as direct stand-ins for human groups until those biases are isolated from structural effects.

Core claim

Through simulations of multi-round debates on networks generated with tunable homophily and class imbalance, LLM agents display agreement drift by shifting toward specific positions on the opinion scale in a directional manner that exceeds what network structure alone predicts, revealing model biases that must be disentangled from structural influences before LLMs can be treated as behavioral proxies for human populations.

What carries the argument

A network generation model with controlled homophily and class sizes that organizes LLM agents for repeated debate rounds to track and compare opinion shifts against structural expectations.

If this is right

  • Network effects such as homophily can be isolated from model biases by using controlled generation of debate networks.
  • Minority groups in unbalanced networks experience amplified effects from the directional drift in LLM populations.
  • LLM-based social simulations require explicit correction for biases before they can model human agreement processes accurately.
  • Direct substitution of LLM agents for human groups in opinion dynamics studies risks misattributing model artifacts to social mechanisms.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Similar directional biases may appear in other LLM multi-agent setups such as negotiations or collaborative problem solving.
  • Varying prompt strategies or model architectures could quantify how much of the drift depends on implementation details.
  • Design of AI systems for social modeling or content moderation might need safeguards against unintended convergence on specific viewpoints.

Load-bearing premise

The directional shifts in agent opinions arise from inherent biases in the LLMs rather than from the specific debate protocol, opinion scale, network generation choices, or prompt designs.

What would settle it

Repeating the exact simulation protocol with human participants or a different family of LLMs and observing no consistent directional preference in opinion shifts would show the drift is not a general property of the models.

Figures

Figures reproduced from arXiv: 2604.11312 by Andrea Failla, Erica Cau, Giulio Rossetti.

Figure 1
Figure 1. Figure 1: Graphical schema of the LLM-OD framework. The LLM agents population is initialized as nodes in a network; each agent is an LLM instance with an initial opinion in the range [0, 6] (a). At each iteration, two neighbor agents are chosen and prompted to act as Opponent and Discussant (b). The Discussant is prompted to listen to the opinion of the Opponent around the discussion statement and may then accept, r… view at source ↗
Figure 3
Figure 3. Figure 3: Debate Require: Initial opinions: oi, oj , Statement: S, Maximum number of interaction rounds: max_rounds Ensure: Final opinion delta of the Discussant: o ∆+ i 1: Init Discussant’s Opinion Delta o ∆+ i ← 0 2: Discussant states its opinion oi on S 3: for round = 1 to max_rounds = 3 do 4: Opponent’s Turn: 5: Opponent evaluates Discussant’s argument 6: Opponent decides and justifies: ACCEPT, REJECT, IGNORE 7:… view at source ↗
Figure 4
Figure 4. Figure 4: Opinion trends for the base scenario (strongly-agreeing majority) across different homophily levels for minority fraction min=0.5 (a), min = 0.3 (b) and min = 0.1 (c) Each panel shows the proportion of agents holding different opinion states over 100 iterations. All points of the opinion scale are mapped to the colors in the legend, from Strongly Disagree (dark red) to Strongly Agree (dark green). 7 [PITH… view at source ↗
Figure 5
Figure 5. Figure 5: Opinion trends for the reversed scenario (strongly-disagreeing majority) across dif￾ferent homophily levels for minority fraction min=0.5 (a), min = 0.3 (b), and min = 0.1 (c) Each panel shows the proportion of agents holding different opinion states over 100 iterations with different levels of homophily. All points of the opinion scale are mapped to the colors in the legend, from Strongly Disagree (dark r… view at source ↗
Figure 6
Figure 6. Figure 6: Opinion trends for the base scenario (strongly-agreeing majority) under neighborhood opinion awareness across different homophily levels for minority fraction min=0.5 (a), min = 0.3 (b), and min = 0.1 (c) Each panel shows the proportion of agents holding different opinion states over 100 iterations. All points of the opinion scale are mapped to the colors in the legend, from Strongly Disagree (dark red) to… view at source ↗
Figure 7
Figure 7. Figure 7: Opinion trends for the reversed scenario (strongly-disagreeing majority) under neigh￾borhood awareness across different homophily levels for minority fraction min=0.5 (a), min = 0.3 (b), and min = 0.1 (c) Each panel shows the proportion of agents holding different opinion states over 100 iterations. All points of the opinion scale are mapped to the colors in the legend, from Strongly Disagree (dark red) to… view at source ↗
Figure 8
Figure 8. Figure 8: Opinion trends under neighborhood awareness across homophily levels. Panels (a–c) show the base scenario (strongly agreeing majority) for minority fractions m = 0.5, m = 0.3, and m = 0.1, respectively; panels (d–e) show the reversed scenario (strongly disagreeing majority) for m = 0.3 and m = 0.1. Each panel reports the proportion of agents in each opinion state over 100 iterations. colors map the 7-point … view at source ↗
Figure 9
Figure 9. Figure 9: Conditional probabilities of persuasion in pairwise interactions between LLM agents. Each matrix [PITH_FULL_IMAGE:figures/full_fig_p012_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Facet plots of statistically significant opinion shifts for Llama agents under different neighborhood [PITH_FULL_IMAGE:figures/full_fig_p013_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: Facet plots of statistically significant opinion shifts for Llama agents under different neighborhood [PITH_FULL_IMAGE:figures/full_fig_p014_11.png] view at source ↗
read the original abstract

Large Language Models (LLMs) have demonstrated an unprecedented ability to simulate human-like social behaviors, making them useful tools for simulating complex social systems. However, it remains unclear to what extent these simulations can be trusted to accurately capture key social mechanisms, particularly in highly unbalanced contexts involving minority groups. This paper uses a network generation model with controlled homophily and class sizes to examine how LLM agents behave collectively in multi-round debates. Moreover, our findings highlight a particular directional susceptibility that we term \textit{agreement drift}, in which agents are more likely to shift toward specific positions on the opinion scale. Overall, our findings highlight the need to disentangle structural effects from model biases before treating LLM populations as behavioral proxies for human groups.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

2 major / 0 minor

Summary. The manuscript employs a network generation model with controlled homophily and class sizes to simulate multi-round debates among LLM agents. It reports the emergence of a directional susceptibility termed 'agreement drift,' in which agents tend to shift toward specific positions on the opinion scale, and concludes that structural network effects must be disentangled from model biases before LLM populations can serve as reliable behavioral proxies for human groups.

Significance. If the reported agreement drift can be shown through appropriate controls to reflect intrinsic LLM properties rather than artifacts of the simulation protocol, the work would be significant for computational social science. It would provide a concrete cautionary example against treating LLM-based opinion dynamics as direct stand-ins for human behavior, particularly in unbalanced minority contexts, and would motivate more rigorous validation standards in this domain.

major comments (2)
  1. [Abstract] Abstract: The central claim of agreement drift and the call to disentangle structural effects from model biases are asserted without any reference to experimental details, statistical tests, sample sizes, or controls. This absence renders the empirical observation unevaluable and prevents assessment of whether the directional susceptibility is robust.
  2. [Methods/Results] Methods/Results (inferred from abstract description): The attribution of shifts to LLM model biases lacks demonstrated isolation from the chosen opinion scale, prompt instructions, network generation procedure, or debate protocol. No ablations (e.g., symmetric vs. asymmetric scales, randomized-update baselines, or neutral-prompt variants) are described, leaving open the possibility that the observed drift is an artifact of discretization or instruction design rather than an intrinsic bias.

Simulated Author's Rebuttal

2 responses · 0 unresolved

We thank the referee for their constructive and detailed feedback, which has helped us strengthen the presentation and rigor of our work on agreement drift in LLM debates. We address each major comment below, clarifying our experimental design and indicating the specific revisions made to the manuscript.

read point-by-point responses
  1. Referee: [Abstract] Abstract: The central claim of agreement drift and the call to disentangle structural effects from model biases are asserted without any reference to experimental details, statistical tests, sample sizes, or controls. This absence renders the empirical observation unevaluable and prevents assessment of whether the directional susceptibility is robust.

    Authors: We agree that the original abstract was too concise and omitted key details needed for evaluation. In the revised manuscript, we have expanded the abstract to explicitly reference the network generation procedure (controlled homophily and class sizes), the multi-round debate protocol, the opinion scale discretization, the number of simulation runs and LLM instances used, and the statistical tests applied to detect directional shifts (including regression-based analysis of opinion changes with significance thresholds). These additions make the empirical basis for agreement drift transparent and allow readers to assess its robustness directly from the abstract. revision: yes

  2. Referee: [Methods/Results] Methods/Results (inferred from abstract description): The attribution of shifts to LLM model biases lacks demonstrated isolation from the chosen opinion scale, prompt instructions, network generation procedure, or debate protocol. No ablations (e.g., symmetric vs. asymmetric scales, randomized-update baselines, or neutral-prompt variants) are described, leaving open the possibility that the observed drift is an artifact of discretization or instruction design rather than an intrinsic bias.

    Authors: We thank the referee for highlighting this important point. Our core contribution centers on separating network structural effects (via systematic variation of homophily and minority class sizes) from the observed directional drift, and the full methods section details the fixed prompt templates and debate rules used across conditions. However, we acknowledge that the original submission did not include the full set of ablations suggested. In the revision, we have added new experiments using symmetric opinion scales, randomized update order baselines, and neutral-prompt variants. These controls show that the agreement drift remains consistent in direction and magnitude, providing evidence that it is not solely an artifact of the original discretization or instructions. We have also expanded the methods to describe the prompt engineering and protocol more explicitly, while noting that exhaustive isolation from every conceivable prompt variant remains an open challenge for the field. revision: yes

Circularity Check

0 steps flagged

No derivation chain or self-referential reductions present

full rationale

The paper is an empirical simulation study of LLM agents in multi-round debates on networks with controlled homophily. It introduces the descriptive label 'agreement drift' for observed directional opinion shifts and calls for disentangling structural effects from model biases. No equations, first-principles derivations, fitted parameters, or predictions appear in the abstract or described content. No self-citations are invoked as load-bearing uniqueness theorems, ansatzes, or imported results. The central observations do not reduce to their own inputs by construction; they are presented as simulation outputs requiring external validation. This is a standard non-circular empirical report.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 1 invented entities

Only abstract available; no explicit free parameters, axioms, or invented entities are detailed beyond the introduction of the term 'agreement drift' to label an observed directional susceptibility.

invented entities (1)
  • agreement drift no independent evidence
    purpose: Label for the observed directional susceptibility in opinion shifts during LLM debates
    Term coined in the paper to describe agents being more likely to shift toward specific positions on the opinion scale.

pith-pipeline@v0.9.0 · 5426 in / 1196 out tokens · 38484 ms · 2026-05-10T15:46:42.602703+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

73 extracted references · 29 canonical work pages · 7 internal anchors

  1. [1]

    OpenAI,JoshAchiam, StevenAdler, SandhiniAgarwal, LamaAhmad, IlgeAkkaya, FlorenciaLeoniAle- man, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, Red Avila, Igor Babuschkin, Suchir Balaji, Valerie Balcom, Paul Baltescu, Haiming Bao, Mohammad Bavarian, Jeff Belgum, Irwan Bello, Jake Berdine, Gabriel Bernadett-Shapiro, Christopher Berner, Lenn...

  2. [2]

    Albert Q. Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, Lélio Renard Lavaud, Marie-Anne Lachaux, Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, and William El Sayed. Mistral 7b, 2023. URLhttps://arxiv...

  3. [3]

    Llama 3 model card.N.A., 2024

    AI@Meta. Llama 3 model card.N.A., 2024. URLhttps://github.com/meta-llama/llama3/blob/ main/MODEL_CARD.md

  4. [4]

    DeepSeek-V3 Technical Report

    DeepSeek-AI, Aixin Liu, Bei Feng, et al. Deepseek-v3 technical report, 2025. URLhttps://arxiv. org/abs/2412.19437

  5. [5]

    All that’s’ human’is not gold: Evaluating human evaluation of generated text.arXiv preprint arXiv:2107.00061, 2021

    Elizabeth Clark, Tal August, Sofia Serrano, Nikita Haduong, Suchin Gururangan, and Noah A Smith. All that’s’ human’is not gold: Evaluating human evaluation of generated text.arXiv preprint arXiv:2107.00061, 2021

  6. [6]

    Sentiment Analysis in the Era of Large Language Models: A Reality Check.CoRR abs/2305.15005, 2023

    Wenxuan Zhang, Yue Deng, Bing Liu, Sinno Jialin Pan, and Lidong Bing. Sentiment analysis in the era of large language models: A reality check.arXiv preprint arXiv:2305.15005, 2023

  7. [7]

    From text to emotion: Unveiling the emotion annotation capabilities of llms.arXiv preprint arXiv:2408.17026, 2024

    Minxue Niu, Mimansa Jaiswal, and Emily Mower Provost. From text to emotion: Unveiling the emotion annotation capabilities of llms.arXiv preprint arXiv:2408.17026, 2024

  8. [8]

    Llm-based topic modeling for dark web q&a forums: A comparative analysis with traditional methods.IEEE Access, 2025

    Luis De-Marcos and Adrián Domínguez-Díaz. Llm-based topic modeling for dark web q&a forums: A comparative analysis with traditional methods.IEEE Access, 2025

  9. [9]

    Large Language Models: A Survey

    ShervinMinaee, TomasMikolov, NarjesNikzad, MeysamChenaghlu, RichardSocher, XavierAmatriain, and Jianfeng Gao. Large language models: A survey, 2025. URLhttps://arxiv.org/abs/2402.06196

  10. [10]

    Evaluating the ability of large language models to emulate personality.Scientific reports, 15(1):519, 2025

    Yilei Wang, Jiabao Zhao, Deniz S Ones, Liang He, and Xin Xu. Evaluating the ability of large language models to emulate personality.Scientific reports, 15(1):519, 2025

  11. [11]

    Edoardo Sebastiano De Duro, Riccardo Improta, and Massimo Stella. Introducing counsellme: A dataset of simulated mental health dialogues for comparing llms like haiku, llamantino and chatgpt against humans.Emerging Trends in Drugs, Addictions, and Health, page 100170, 2025

  12. [12]

    Modeling social opinion evolution with llm agents: Integrating personality traits with embedded opinion dynamics

    Fan Yang, Jie Bai, Linjing Li, and Daniel Zeng. Modeling social opinion evolution with llm agents: Integrating personality traits with embedded opinion dynamics. In2025 International Joint Conference on Neural Networks (IJCNN), pages 1–8. IEEE, 2025

  13. [13]

    Generative agents: Interactive simulacra of human behavior

    Joon Sung Park, Joseph O’Brien, Carrie Jun Cai, Meredith Ringel Morris, Percy Liang, and Michael S Bernstein. Generative agents: Interactive simulacra of human behavior. InProceedings of the 36th annual acm symposium on user interface software and technology, pages 1–22, 2023

  14. [14]

    LLM Agents Grounded in Self-Reports Enable General-Purpose Simulation of Individuals

    Joon Sung Park, Carolyn Q. Zou, Aaron Shaw, Benjamin Mako Hill, Carrie Cai, Meredith Ringel Morris, Robb Willer, Percy Liang, and Michael S. Bernstein. Generative agent simulations of 1,000 people, 2024. URLhttps://arxiv.org/abs/2411.10109

  15. [15]

    Systematic biases in llm simulations of debates.arXiv preprint arXiv:2402.04049, 2024

    Amir Taubenfeld, Yaniv Dover, Roi Reichart, and Ariel Goldstein. Systematic biases in llm simulations of debates.arXiv preprint arXiv:2402.04049, 2024. 18

  16. [16]

    Hidden persuaders: Llms’ political leaning and their influence on voters

    Yujin Potter, Shiyang Lai, Junsol Kim, James Evans, and Dawn Song. Hidden persuaders: Llms’ political leaning and their influence on voters. InProceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 4244–4275, 2024

  17. [17]

    A foundation model to predict and capture human cognition.Nature, pages 1–8, 2025

    Marcel Binz, Elif Akata, Matthias Bethge, Franziska Brändle, Fred Callaway, Julian Coda-Forno, Peter Dayan, Can Demircan, Maria K Eckstein, Noémi Éltető, et al. A foundation model to predict and capture human cognition.Nature, pages 1–8, 2025

  18. [18]

    Ysocial: an ai-powered social media virtual twin.Big Data & Society, 2026

    Giulio Rossetti, Massimo Stella, Rémy Cazabet, Katherine Abramski, Erica Cau, Salvatore Citraro, Andrea Failla, Virginia Morini, and Valentina Pansanella. Ysocial: an ai-powered social media virtual twin.Big Data & Society, 2026. doi: 10.1177/20539517261431576

  19. [19]

    Emergent social conventions and col- lective bias in llm populations.Science Advances, 11(20):eadu9368, 2025

    Ariel Flint Ashery, Luca Maria Aiello, and Andrea Baronchelli. Emergent social conventions and col- lective bias in llm populations.Science Advances, 11(20):eadu9368, 2025

  20. [20]

    On the unexpected abilities of large language models.Adaptive Behavior, 32(6):493–502, 2024

    Stefano Nolfi. On the unexpected abilities of large language models.Adaptive Behavior, 32(6):493–502, 2024

  21. [21]

    Behavioral and Brain Sciences , author=

    David Premack and Guy Woodruff. Does the chimpanzee have a theory of mind?Behavioral and Brain Sciences, 1(4):515–526, December 1978. ISSN 1469-1825. doi: 10.1017/s0140525x00076512. URL http://dx.doi.org/10.1017/S0140525X00076512

  22. [22]

    Llm-based social simulations require a bound- ary.arXiv preprint arXiv:2506.19806, 2025

    Zengqing Wu, Run Peng, Takayuki Ito, and Chuan Xiao. Llm-based social simulations require a bound- ary.arXiv preprint arXiv:2506.19806, 2025

  23. [23]

    arXiv preprint arXiv:2310.10701 , year=

    Huao Li, Yu Quan Chong, Simon Stepputtis, Joseph Campbell, Dana Hughes, Michael Lewis, and Katia Sycara. Theory of mind for multi-agent collaboration via large language models.arXiv preprint arXiv:2310.10701, 2023

  24. [24]

    Neural theory-of-mind? on the limits of social intelligence in large lms.arXiv preprint arXiv:2210.13312, 2022

    Maarten Sap, Ronan LeBras, Daniel Fried, and Yejin Choi. Neural theory-of-mind? on the limits of social intelligence in large lms.arXiv preprint arXiv:2210.13312, 2022

  25. [25]

    Llms achieve adult human performance on higher-order theory of mind tasks.arXiv preprint arXiv:2405.18870, 2024

    Winnie Street, John Oliver Siy, Geoff Keeling, Adrien Baranes, Benjamin Barnett, Michael McKibben, Tatenda Kanyere, Alison Lentz, Robin IM Dunbar, et al. Llms achieve adult human performance on higher-order theory of mind tasks.arXiv preprint arXiv:2405.18870, 2024

  26. [26]

    Testing theory of mind in large language models and humans.Nature Human Behaviour, 8(7):1285–1295, 2024

    James WA Strachan, Dalila Albergo, Giulia Borghini, Oriana Pansardi, Eugenio Scaliti, Saurabh Gupta, Krati Saxena, Alessandro Rufo, Stefano Panzeri, Guido Manzi, et al. Testing theory of mind in large language models and humans.Nature Human Behaviour, 8(7):1285–1295, 2024

  27. [27]

    Large language models fail on trivial alterations to theory-of-mind tasks, 2023

    Tomer Ullman. Large language models fail on trivial alterations to theory-of-mind tasks.arXiv preprint arXiv:2302.08399, 2023

  28. [28]

    Evaluating large language models in theory of mind tasks.Proceedings of the National Academy of Sciences, 121(45):e2405460121, 2024

    Michal Kosinski. Evaluating large language models in theory of mind tasks.Proceedings of the National Academy of Sciences, 121(45):e2405460121, 2024

  29. [29]

    The persuasive power of large language models

    Simon Martin Breum, Daniel Væ dele Egdal, Victor Gram Mortensen, Anders Giovanni Mø ller, and Luca Maria Aiello. The persuasive power of large language models. InProceedings of the International AAAI Conference on Web and Social Media, volume 18, pages 152–163, 2024

  30. [30]

    Birds of a feather: Homophily in social networks.Annual review of sociology, 27(1):415–444, 2001

    Miller McPherson, Lynn Smith-Lovin, and James M Cook. Birds of a feather: Homophily in social networks.Annual review of sociology, 27(1):415–444, 2001

  31. [31]

    InACM CO- DASPY

    Xinnong Zhang, Jiayu Lin, Xinyi Mou, Shiyue Yang, Xiawei Liu, Libo Sun, Hanjia Lyu, Yihang Yang, Weihong Qi, Yue Chen, et al. Socioverse: A world model for social simulation powered by llm agents and a pool of 10 million real-world users.arXiv preprint arXiv:2504.10157, 2025

  32. [32]

    Understanding online polarization through human-agent interaction in a synthetic llm-based social network.N.A., 2025

    Tim Donkers and Jürgen Ziegler. Understanding online polarization through human-agent interaction in a synthetic llm-based social network.N.A., 2025. 19

  33. [33]

    Using large language models in psychology.Nature Reviews Psychology, 2(11):688–701, 2023

    Dorottya Demszky, Diyi Yang, David S Yeager, Christopher J Bryan, Margarett Clapper, Susannah Chandhok, Johannes C Eichstaedt, Cameron Hecht, Jeremy Jamieson, Meghann Johnson, et al. Using large language models in psychology.Nature Reviews Psychology, 2(11):688–701, 2023

  34. [34]

    The problems of llm-generated data in social science research.Sociologica: International Journal for Sociological Debate, 18(2):145–168, 2024

    Luca Rossi, Katherine Harrison, and Irina Shklovski. The problems of llm-generated data in social science research.Sociologica: International Journal for Sociological Debate, 18(2):145–168, 2024

  35. [35]

    Homophily-induced emergence of biased structures in llm- based multi-agent ai systems.Social Network Analysis and Mining, 15(1):1–25, 2025

    Aliakbar Mehdizadeh and Martin Hilbert. Homophily-induced emergence of biased structures in llm- based multi-agent ai systems.Social Network Analysis and Mining, 15(1):1–25, 2025

  36. [36]

    Homophily influences ranking of minorities in social networks.Scientific reports, 8(1):11077, 2018

    Fariba Karimi, Mathieu Génois, Claudia Wagner, Philipp Singer, and Markus Strohmaier. Homophily influences ranking of minorities in social networks.Scientific reports, 8(1):11077, 2018

  37. [37]

    Towards understanding sycophancy in language models

    Mrinank Sharma, Meg Tong, Tomasz Korbak, David Duvenaud, Amanda Askell, Samuel R Bowman, Esin DURMUS, Zac Hatfield-Dodds, Scott R Johnston, Shauna M Kravec, et al. Towards understanding sycophancy in language models. InThe Twelfth International Conference on Learning Representations, pages 383–390, 2025

  38. [38]

    Selective agreement, not syco- phancy: investigating opinion dynamics in llm interactions.EPJ Data Science, 14(1):59, 2025

    Erica Cau, Valentina Pansanella, Dino Pedreschi, and Giulio Rossetti. Selective agreement, not syco- phancy: investigating opinion dynamics in llm interactions.EPJ Data Science, 14(1):59, 2025

  39. [39]

    Statisticalphysicsofsocialdynamics.Reviews of modern physics, 81(2):591, 2009

    ClaudioCastellano, SantoFortunato, andVittorioLoreto. Statisticalphysicsofsocialdynamics.Reviews of modern physics, 81(2):591, 2009

  40. [40]

    Bounded confidence opinion dynamics in virtual networks and real networks

    Xia-Meng Si and Chen Li. Bounded confidence opinion dynamics in virtual networks and real networks. Journal of Computers, 29(3):220–228, 2018

  41. [41]

    Ergodic theorems for weakly interacting infinite systems and the voter model.The annals of probability, pages 643–663, 1975

    Richard A Holley and Thomas M Liggett. Ergodic theorems for weakly interacting infinite systems and the voter model.The annals of probability, pages 643–663, 1975

  42. [42]

    Mixing beliefs among interacting agents.Advances in Complex Systems, 3(01n04):87–98, 2000

    Guillaume Deffuant, David Neau, Frederic Amblard, and Gérard Weisbuch. Mixing beliefs among interacting agents.Advances in Complex Systems, 3(01n04):87–98, 2000

  43. [43]

    Dynamics of majority rule in two-state interacting spin systems

    Paul L Krapivsky and Sidney Redner. Dynamics of majority rule in two-state interacting spin systems. Physical Review Letters, 90(23):238701, 2003

  44. [44]

    Routledge, 2013

    Irving Crespi.The public opinion process: How the people speak. Routledge, 2013

  45. [45]

    Recent advances in opinion propagation dynamics: A 2020 survey.The European Physical Journal Plus, 135:1–20, 2020

    Hossein Noorazar. Recent advances in opinion propagation dynamics: A 2020 survey.The European Physical Journal Plus, 135:1–20, 2020

  46. [46]

    Opinion dynamics: models, ex- tensions and external effects.Participatory sensing, opinions and collective awareness, pages 363–401, 2017

    Alina Sîrbu, Vittorio Loreto, Vito DP Servedio, and Francesca Tria. Opinion dynamics: models, ex- tensions and external effects.Participatory sensing, opinions and collective awareness, pages 363–401, 2017

  47. [47]

    A model for spatial conflict.Biometrika, 60(3):581–588,

    PETER CLIFFORD and AIDAN SUDBURY. A model for spatial conflict.Biometrika, 60(3):581–588,

  48. [48]

    doi: 10.1093/biomet/60.3.581

    ISSN 1464-3510. doi: 10.1093/biomet/60.3.581. URLhttp://dx.doi.org/10.1093/biomet/60. 3.581

  49. [49]

    Morris H. Degroot. Reaching a consensus.Journal of the American Statistical Association, 69(345): 118–121, March 1974. ISSN 1537-274X. doi: 10.1080/01621459.1974.10480137. URLhttp://dx.doi. org/10.1080/01621459.1974.10480137

  50. [50]

    De Marzo, L

    Giordano De Marzo, Luciano Pietronero, and David Garcia. Emergence of scale-free networks in social interactions among large language models.arXiv preprint arXiv:2312.06619, 2023

  51. [51]

    S3: Social-network simulation system with large language model-empowered agents

    Chen Gao, Xiaochong Lan, Zhihong Lu, Jinzhu Mao, Jinghua Piao, Huandong Wang, Depeng Jin, and Yong Li. S3: Social-network simulation system with large language model-empowered agents.arXiv preprint arXiv:2307.14984, 2023. 20

  52. [52]

    arXiv preprint arXiv:2402.04559 , year=

    Chengxing Xie, Canyu Chen, Feiran Jia, Ziyu Ye, Kai Shu, Adel Bibi, Ziniu Hu, Philip Torr, Bernard Ghanem, and Guohao Li. Can large language model agents simulate human trust behaviors?arXiv preprint arXiv:2402.04559, 2024

  53. [53]

    Bernstein

    Joon Sung Park, Lindsay Popowski, Carrie Cai, Meredith Ringel Morris, Percy Liang, and Michael S. Bernstein. Social simulacra: Creating populated prototypes for social computing systems. InProceedings of the 35th Annual ACM Symposium on User Interface Software and Technology, UIST ’22, pages 383– 390, New York, NY, USA, 2022. Association for Computing Mac...

  54. [54]

    Can llm agents maintain a persona in discourse? InProceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 29201–29217, 2025

    Pranav Bhandari, Nicolas Fay, Michael J Wise, Amitava Datta, Stephanie Meek, Usman Naseem, and Mehwish Nasim. Can llm agents maintain a persona in discourse? InProceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 29201–29217, 2025

  55. [56]

    Agent-based social simulation of the covid-19 pandemic: A systematic review.Journal of Artificial Societies and Social Simulation, 24(3), 2021

    Fabian Lorig, Emil Johansson, and Paul Davidsson. Agent-based social simulation of the covid-19 pandemic: A systematic review.Journal of Artificial Societies and Social Simulation, 24(3), 2021

  56. [57]

    T¨ ornberg, D

    PetterTörnberg, DiliaraValeeva, JustusUitermark, andChristopherBail. Simulatingsocialmediausing large language models to evaluate alternative news feed algorithms.arXiv preprint arXiv:2310.05984, 2023

  57. [58]

    Yun-Shiuan Chuang, Agam Goyal, Nikunj Harlalka, Siddharth Suresh, Robert Hawkins, Sijia Yang, Dhavan Shah, Junjie Hu, and Timothy T. Rogers. Simulating opinion dynamics with networks of llm-based agents, 2024. URLhttps://arxiv.org/abs/2311.09618

  58. [59]

    The language of opinion change on social media under the lens of communicative action.Scientific Reports, 12(1):17920, 2022

    Corrado Monti, Luca Maria Aiello, Gianmarco De Francisci Morales, and Francesco Bonchi. The language of opinion change on social media under the lens of communicative action.Scientific Reports, 12(1):17920, 2022

  59. [60]

    arXiv preprint arXiv:2406.15492 , year=

    Pedro Cisneros-Velarde. On the principles behind opinion dynamics in multi-agent systems of large language models, 2024. URLhttps://arxiv.org/abs/2406.15492

  60. [61]

    The political preferences of llms.PLOS ONE, 19, 2024

    David Rozado. The political preferences of llms.PLOS ONE, 19, 2024. URLhttps://api. semanticscholar.org/CorpusID:267412830

  61. [62]

    Chatgpt in society: emerging issues.Frontiers in Artificial Intelli- gence, 6, 2023

    Mirko Farina and Andrea Lavazza. Chatgpt in society: emerging issues.Frontiers in Artificial Intelli- gence, 6, 2023. URLhttps://api.semanticscholar.org/CorpusID:259168081

  62. [63]

    Botsofafeather: mixingbiasesinllms’opiniondynamics

    EricaCau, AndreaFailla, andGiulioRossetti. Botsofafeather: mixingbiasesinllms’opiniondynamics. InInternational conference on complex networks and their applications, pages 166–176. Springer, 2024

  63. [64]

    When corrections fail: The persistence of political misperceptions

    Brendan Nyhan and Jason Reifler. When corrections fail: The persistence of political misperceptions. Political Behavior, 32(2):303–330, 2010

  64. [65]

    Dryden and A.H

    J. Dryden and A.H. Clough.Parallel Lives. e-artnow, 2018. ISBN 9788027244577. URLhttps: //books.google.it/books?id=-amSDwAAQBAJ

  65. [66]

    Gemma 3 Technical Report

    Gemma Team, Aishwarya Kamath, Johan Ferret, Shreya Pathak, Nino Vieillard, Ramona Merhej, Sarah Perrin, Tatiana Matejovicova, Alexandre Ramé, Morgane Rivière, et al. Gemma 3 technical report.arXiv preprint arXiv:2503.19786, 2025

  66. [67]

    Social influence and the collective dynamics of opinion formation.PloS one, 8(11):e78433, 2013

    Mehdi Moussaïd, Juliane E Kämmer, Pantelis P Analytis, and Hansjörg Neth. Social influence and the collective dynamics of opinion formation.PloS one, 8(11):e78433, 2013

  67. [68]

    Mehdizadeh and M

    Aliakbar Mehdizadeh and Martin Hilbert. When your ai agent succumbs to peer-pressure: Studying opinion-change dynamics of llms.arXiv preprint arXiv:2510.19107, 2025. 21

  68. [69]

    ELEPHANT: Measuring and understanding social sycophancy in LLMs

    Myra Cheng, Sunny Yu, Cinoo Lee, Pranav Khadpe, Lujain Ibrahim, and Dan Jurafsky. Elephant: Measuring and understanding social sycophancy in llms.arXiv preprint arXiv:2505.13995, 2025

  69. [70]

    Syceval: Evaluating llm sycophancy

    Aaron Fanous, Jacob Goldberg, Ank Agarwal, Joanna Lin, Anson Zhou, Sonnet Xu, Vasiliki Bikia, Roxana Daneshjou, and Sanmi Koyejo. Syceval: Evaluating llm sycophancy. InProceedings of the AAAI/ACM Conference on AI, Ethics, and Society, volume 8, pages 893–900, 2025

  70. [71]

    Mitigating voltage fluctuations in grid-tied pv systems: power qual- ity enhancement

    Sheena Tahira Khan, Vilas Ramrao Joshi, Deepak Bhardwaj, Surender Kumar Sharma, Shaikh Abdul Waheed, and Deepak K Sharma. Mitigating voltage fluctuations in grid-tied pv systems: power qual- ity enhancement. In2024 3rd International Conference on Computational Modelling, Simulation and Optimization (ICCMSO), pages 407–416. IEEE, 2024

  71. [72]

    Algorithmic bias amplifies opinion fragmentation and polarization: A bounded confidence model.PloS one, 14(3):e0213246, 2019

    Alina Sîrbu, Dino Pedreschi, Fosca Giannotti, and János Kertész. Algorithmic bias amplifies opinion fragmentation and polarization: A bounded confidence model.PloS one, 14(3):e0213246, 2019

  72. [73]

    From mean-field to complex topologies: net- work effects on the algorithmic bias model

    Valentina Pansanella, Giulio Rossetti, and Letizia Milli. From mean-field to complex topologies: net- work effects on the algorithmic bias model. InComplex Networks & Their Applications X: Volume 2, Proceedings of the Tenth International Conference on Complex Networks and Their Applications COMPLEX NETWORKS 2021 10, pages 329–340. Springer, 2022

  73. [74]

    Emergence of scaling in random networks.science, 286(5439): 509–512, 1999

    Albert-László Barabási and Réka Albert. Emergence of scaling in random networks.science, 286(5439): 509–512, 1999. 22