pith. machine review for the scientific record. sign in

arxiv: 2604.20166 · v1 · submitted 2026-04-22 · 💻 cs.CL · cs.HC

Recognition: unknown

Aligning Human-AI-Interaction Trust for Mental Health Support: Survey and Position for Multi-Stakeholders

Authors on Pith no claims yet

Pith reviewed 2026-05-10 00:41 UTC · model grok-4.3

classification 💻 cs.CL cs.HC
keywords trustworthy AImental healthhuman-AI interactiontrust frameworkAI evaluationmulti-stakeholdertherapeutic AINLP metrics
0
0 comments X

The pith

A three-layer trust framework organizes human, AI, and interaction perspectives to align stakeholders on trustworthy AI for mental health support.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

Different groups define trust in AI mental health tools inconsistently: technical researchers emphasize robustness and safety while practitioners focus on empathy, appropriateness, and user outcomes. The paper proposes a three-layer framework of human-oriented trust, AI-oriented trust, and interaction-oriented trust to integrate these viewpoints from practitioners, researchers, and regulators. It applies the framework to review existing AI-driven mental health research and current evaluation practices, from automatic metrics to clinically validated methods. This review identifies gaps between what NLP systems currently measure and what real-world therapeutic contexts require. The result is a research agenda aimed at creating socio-technically aligned AI that meets genuine mental health needs.

Core claim

We propose a three-layer trust framework covering human-oriented, AI-oriented, and interaction-oriented trust that integrates the viewpoints of key stakeholders including practitioners, researchers, and regulators. Using this framework we systematically review AI-driven mental health research and examine evaluation practices for trustworthiness, highlighting critical gaps between NLP metrics and real-world mental health requirements while outlining a research agenda for building aligned and genuinely trustworthy AI.

What carries the argument

The three-layer trust framework that structures trust into human-oriented, AI-oriented, and interaction-oriented dimensions to reconcile differing stakeholder definitions and enable reorganization of existing literature.

If this is right

  • Existing AI mental health studies can be reorganized under the three layers to clarify where technical and therapeutic priorities diverge.
  • Evaluation of AI trustworthiness must combine automatic metrics with clinically validated approaches to address identified gaps.
  • A concrete research agenda follows for developing AI systems that satisfy both technical criteria and therapeutic outcomes.
  • Multi-stakeholder alignment becomes feasible by using the shared framework as a common reference point for design and regulation.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • Regulators could require new AI mental health tools to demonstrate measurable performance across all three trust layers before deployment.
  • Interface designers might prioritize features that strengthen interaction-oriented trust, such as sustained conversation quality over time.
  • Longitudinal studies could test whether tools built under this framework produce better retention and outcome rates than current systems.

Load-bearing premise

The three-layer structure comprehensively captures all relevant dimensions of trust without significant overlaps or missing factors, and existing literature can be meaningfully reorganized under it.

What would settle it

A documented dimension of trust in AI mental health systems that cannot be placed into any of the three layers, or an empirical study showing that applying the framework fails to reduce inconsistencies in how stakeholders evaluate the same AI tools.

Figures

Figures reproduced from arXiv: 2604.20166 by Charlotte Gerritsen, Jiahuan Pei, Koen Hindriks, Mengyuan Zhang, Qingyu Meng, Saku Sugawara, Sander L. Koole, Xin Sun, Yifan Mo, Yue Su, Yuxuan Li.

Figure 1
Figure 1. Figure 1: Discipline network of 1,706 surveyed papers [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: Based on a cross-disciplinary literature re [PITH_FULL_IMAGE:figures/full_fig_p002_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: The proposed Three-Layer Trust Framework. Each layer reflects priorities across stakeholders and sum [PITH_FULL_IMAGE:figures/full_fig_p004_3.png] view at source ↗
read the original abstract

Building trustworthy AI systems for mental health support is a shared priority across stakeholders from multiple disciplines. However, "trustworthy" remains loosely defined and inconsistently operationalized. AI research often focuses on technical criteria (e.g., robustness, explainability, and safety), while therapeutic practitioners emphasize therapeutic fidelity (e.g., appropriateness, empathy, and long-term user outcomes). To bridge the fragmented landscape, we propose a three-layer trust framework, covering human-oriented, AI-oriented, and interaction-oriented trust, integrating the viewpoints of key stakeholders (e.g., practitioners, researchers, regulators). Using this framework, we systematically review existing AI-driven research in mental health domain and examine evaluation practices for ``trustworthy'' ranging from automatic metrics to clinically validated approaches. We highlight critical gaps between what NLP currently measures and what real-world mental health contexts require, and outline a research agenda for building socio-technically aligned and genuinely trustworthy AI for mental health support.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

1 major / 2 minor

Summary. The manuscript surveys AI research in mental health support and proposes a three-layer trust framework (human-oriented, AI-oriented, and interaction-oriented trust) to integrate multi-stakeholder perspectives from practitioners, researchers, regulators, and others. It reviews evaluation practices ranging from automatic NLP metrics to clinically validated approaches, identifies gaps between current technical measures and real-world clinical needs, and outlines a research agenda for socio-technically aligned trustworthy AI.

Significance. If adopted, the framework could help bridge disciplinary divides by providing a shared organizing structure for trust in mental health AI, surfacing actionable discrepancies between NLP evaluation practices and therapeutic requirements. The position paper's value lies in its constructive synthesis rather than new data or formal derivations; credit is due for explicitly framing the work as a starting point for multi-stakeholder alignment rather than a complete model.

major comments (1)
  1. [Abstract] Abstract and review description: the claim of a 'systematic review' of AI-driven mental health research is central to identifying gaps, yet no details are provided on search strategy, databases, inclusion/exclusion criteria, or the process used to map literature onto the three-layer framework. This limits assessment of whether the highlighted gaps are comprehensive or selective.
minor comments (2)
  1. [Framework proposal] The three layers are introduced as a synthesis of stakeholder viewpoints, but the manuscript would benefit from a brief explicit discussion of boundary cases or potential overlaps (e.g., where interaction-oriented trust intersects with human- or AI-oriented dimensions) to clarify the framework's structure.
  2. [Research agenda] As a position paper without new empirical validation, the research agenda section could more clearly distinguish between recommendations that follow directly from the reviewed literature and those that are forward-looking proposals.

Simulated Author's Rebuttal

1 responses · 0 unresolved

We thank the referee for their constructive review and positive recommendation for minor revision. We address the single major comment below and will update the manuscript to improve transparency around our literature review process.

read point-by-point responses
  1. Referee: [Abstract] Abstract and review description: the claim of a 'systematic review' of AI-driven mental health research is central to identifying gaps, yet no details are provided on search strategy, databases, inclusion/exclusion criteria, or the process used to map literature onto the three-layer framework. This limits assessment of whether the highlighted gaps are comprehensive or selective.

    Authors: We agree that the phrasing 'we systematically review' in the abstract and introduction implies a formal systematic review with explicit methodology, which is not provided. Our review is a structured, framework-guided synthesis of representative literature rather than a comprehensive PRISMA-style systematic review. To address this, we will (1) revise the abstract and relevant sections to state 'we review' or 'we conduct a structured review', (2) add a brief methods subsection describing the search strategy (databases such as ACL Anthology, PubMed, arXiv, Google Scholar; keywords combining 'AI', 'mental health', 'trust', 'evaluation'; time frame 2015-2024), inclusion criteria (peer-reviewed or preprint works on AI for mental health support), and the mapping process to the three trust layers, and (3) note the review's illustrative rather than exhaustive scope. These changes will make the process transparent while preserving the position-paper focus. revision: yes

Circularity Check

0 steps flagged

No significant circularity in proposed framework

full rationale

The paper is a survey and position piece whose central contribution is the proposal of a three-layer trust framework (human-oriented, AI-oriented, and interaction-oriented) as an organizing synthesis drawn from stakeholder viewpoints and literature gaps. No derivation chain, equations, fitted parameters, or formal predictions are presented; the framework is explicitly constructive rather than derived from first principles or self-referential definitions. No load-bearing step reduces to its own inputs by construction, and the paper does not invoke self-citation chains, uniqueness theorems, or ansatzes from prior author work to justify its structure. The assumption that existing literature can be reorganized under the layers serves only to surface gaps between NLP metrics and clinical needs, leaving the paper self-contained against external benchmarks.

Axiom & Free-Parameter Ledger

0 free parameters · 1 axioms · 3 invented entities

The paper's primary contribution is the conceptual framework itself, which rests on the domain assumption that trust is multi-dimensional and requires integration across stakeholder groups; no numerical free parameters are fitted and no new physical entities are postulated.

axioms (1)
  • domain assumption Trust in AI for mental health support is best understood through multiple layers (human-oriented, AI-oriented, interaction-oriented) and the viewpoints of multiple stakeholders.
    This premise directly motivates the proposal of the framework and the reorganization of existing research.
invented entities (3)
  • human-oriented trust layer no independent evidence
    purpose: To capture user perceptions, empathy, and therapeutic fidelity aspects
    Introduced as one of the three organizing layers in the new framework.
  • AI-oriented trust layer no independent evidence
    purpose: To capture technical properties such as robustness, explainability, and safety
    Introduced as one of the three organizing layers in the new framework.
  • interaction-oriented trust layer no independent evidence
    purpose: To capture dynamic, long-term aspects of human-AI conversations and outcomes
    Introduced as one of the three organizing layers in the new framework.

pith-pipeline@v0.9.0 · 5503 in / 1544 out tokens · 35007 ms · 2026-05-10T00:41:45.815780+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

130 extracted references · 64 canonical work pages · 1 internal anchor

  1. [2]

    and Masoud, Reem and Alnuhait, Deema and Alomairi, Afnan Y

    Alghamdi, Emad A. and Masoud, Reem and Alnuhait, Deema and Alomairi, Afnan Y. and Ashraf, Ahmed and Zaytoon, Mohamed. A ra T rust: An Evaluation of Trustworthiness for LLM s in A rabic. Proceedings of the 31st International Conference on Computational Linguistics. 2025

  2. [8]

    PAL to Lend a Helping Hand: Towards Building an Emotion Adaptive Polite and Empathetic Counseling Conversational Agent

    Mishra, Kshitij and Priya, Priyanshu and Ekbal, Asif. PAL to Lend a Helping Hand: Towards Building an Emotion Adaptive Polite and Empathetic Counseling Conversational Agent. Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2023. doi:10.18653/v1/2023.acl-long.685

  3. [11]

    Enhancing Emotional Support Conversations: A Framework for Dynamic Knowledge Filtering and Persona Extraction

    Hao, Jiawang and Kong, Fang. Enhancing Emotional Support Conversations: A Framework for Dynamic Knowledge Filtering and Persona Extraction. Proceedings of the 31st International Conference on Computational Linguistics. 2025

  4. [12]

    and Krahmer, Emiel

    Basar, Erkan and Sun, Xin and Hendrickx, Iris and de Wit, Jan and Bosse, Tibor and De Bruijn, Gert-Jan and Bosch, Jos A. and Krahmer, Emiel. How Well Can Large Language Models Reflect? A Human Evaluation of LLM -generated Reflections for Motivational Interviewing Dialogues. Proceedings of the 31st International Conference on Computational Linguistics. 2025

  5. [13]

    Can Large Language Models be Good Emotional Supporter? Mitigating Preference Bias on Emotional Support Conversation

    Kang, Dongjin and Kim, Sunghwan and Kwon, Taeyoon and Moon, Seungjun and Cho, Hyunsouk and Yu, Youngjae and Lee, Dongha and Yeo, Jinyoung. Can Large Language Models be Good Emotional Supporter? Mitigating Preference Bias on Emotional Support Conversation. Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Lo...

  6. [15]

    Huang, Yue and Sun, Lichao and Wang, Haoran and Wu, Siyuan and Zhang, Qihui and Li, Yuan and Gao, Chujie and Huang, Yixin and Lyu, Wenhan and Zhang, Yixuan and Li, Xiner and Sun, Hanchi and Liu, Zhengliang and Liu, Yixin and Wang, Yijue and Zhang, Zhikun and Vidgen, Bertie and Kailkhura, Bhavya and Xiong, Caiming and Xiao, Chaowei and Li, Chunyuan and Xin...

  7. [23]

    and Arora, Simran and Mazeika, Manias and Hendrycks, Dan and Lin, Zinan and Cheng, Yu and Koyejo, Sanmi and Song, Dawn and Li, Bo , title =

    Wang, Boxin and Chen, Weixin and Pei, Hengzhi and Xie, Chulin and Kang, Mintong and Zhang, Chenhui and Xu, Chejian and Xiong, Zidi and Dutta, Ritik and Schaeffer, Rylan and Truong, Sang T. and Arora, Simran and Mazeika, Manias and Hendrycks, Dan and Lin, Zinan and Cheng, Yu and Koyejo, Sanmi and Song, Dawn and Li, Bo , title =. Proceedings of the 37th Int...

  8. [24]

    Trustworthy

    Yang Liu and Yuanshun Yao and Jean-Francois Ton and Xiaoying Zhang and Ruocheng Guo and Hao Cheng and Yegor Klochkov and Muhammad Faaiz Taufiq and Hang Li , booktitle=. Trustworthy. 2023 , url=

  9. [25]

    Proceedings of the 34th USENIX Conference on Security Symposium , articleno =

    Kwesi, Jabari and Cao, Jiaxun and Manchanda, Riya and Emami-Naeini, Pardis , title =. Proceedings of the 34th USENIX Conference on Security Symposium , articleno =. 2025 , isbn =

  10. [26]

    Position: Beyond Assistance

    Abeer Badawi and Md Tahmid Rahman Laskar and Jimmy Huang and Shaina Raza and Elham Dolatabadi , booktitle=. Position: Beyond Assistance. 2025 , url=

  11. [30]

    Artificial Intelligence Act , url =

    EPEUCO , month =. Artificial Intelligence Act , url =

  12. [31]

    Augmented Intelligence Development, Deployment, and Use in Health Care , url =

    AMA , journal =. Augmented Intelligence Development, Deployment, and Use in Health Care , url =

  13. [32]

    Ethical Decision-Making Guidelines for Mental Health Clinicians in the Artificial Intelligence (AI) Era , volume =

    Yegan Pillay , doi =. Ethical Decision-Making Guidelines for Mental Health Clinicians in the Artificial Intelligence (AI) Era , volume =. Healthcare 2025, Vol. 13, , keywords =

  14. [33]

    Ethical Guidance for AI in the Professional Practice of Health Service Psychology , year =

    APA , month =. Ethical Guidance for AI in the Professional Practice of Health Service Psychology , year =

  15. [34]

    Ethics Guidelines for Trustworthy

    AIHLEG , month =. Ethics Guidelines for Trustworthy

  16. [35]

    Designing human-centered

    Thieme, Anja and Hanratty, Maryann and Lyons, Maria and Palacios, Jorge and Marques, Rita Faia and Morrison, Cecily and Doherty, Gavin , journal=. Designing human-centered. 2023 , publisher=

  17. [36]

    Uncovering contradictions in human-

    Namvarpour, Mohammad and Razi, Afsaneh , booktitle=. Uncovering contradictions in human-

  18. [37]

    Understanding Attitudes and Trust of Generative

    Wang, Yimeng and Wang, Yinzhou and Crace, Kelly and Zhang, Yixuan , booktitle=. Understanding Attitudes and Trust of Generative

  19. [38]

    Understanding Adolescents' Perceptions of Benefits and Risks in Health

    Lee, Jamie and Jung, Kyuha and Newman, Erin Gregg and Chow, Emilie and Chen, Yunan , booktitle=. Understanding Adolescents' Perceptions of Benefits and Risks in Health

  20. [39]

    Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems , pages=

    Private Yet Social: How LLM Chatbots Support and Challenge Eating Disorder Recovery , author=. Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems , pages=

  21. [40]

    Can I have my friend attending with me?

    " Can I have my friend attending with me?": Design Implications for Using Virtual Supporters in Remote Psychotherapy , author=. Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems , pages=

  22. [41]

    Extended abstracts of the 2023 CHI conference on human factors in computing systems , pages=

    Trust and perceived control in burnout support chatbots , author=. Extended abstracts of the 2023 CHI conference on human factors in computing systems , pages=

  23. [42]

    Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems , pages=

    Evaluating the experience of LGBTQ+ people using large language model based chatbots for mental health support , author=. Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems , pages=

  24. [43]

    Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems , pages=

    Tailored Virtual Agent Guidance for Stress Management: Comparing Directive and Non-Directive Approaches by Alexithymia Status , author=. Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems , pages=

  25. [44]

    Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems , pages=

    Customizing emotional support: How do individuals construct and interact with LLM-powered chatbots , author=. Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems , pages=

  26. [48]

    2022 , issue_date =

    Haque, Md Romael and Rubya, Sabirat , title =. 2022 , issue_date =. doi:10.1145/3555146 , journal =

  27. [50]

    and Moray, Neville , title =

    Lee, John D. and Moray, Neville , title =. Ergonomics , year =

  28. [51]

    Hurley and John Herrington and Eric Alan Storch and Casey J

    Ansh Rai and Meghan E. Hurley and John Herrington and Eric Alan Storch and Casey J. Zampella and Julia Parish-Morris and Anika Sonig and Gabriel Lázaro-Muñoz and Kristin Kostick-Quenet , doi =. Stakeholder Criteria for Trust in Artificial Intelligence–Based Computer Perception Tools in Health Care: Qualitative Interview Study , volume =. Journal of Medica...

  29. [52]

    In A.I. we trust?

    Seounmi Youn and S. Venus Jin , doi =. “In A.I. we trust?” The effects of parasocial interaction and technopian versus luddite ideological views on chatbot-based customer relationship management in the emerging “feeling economy” , volume =. Computers in Human Behavior , keywords =

  30. [53]

    Journal of Cognitive Engineering and Decision Making , year =

    Manzey, Dietrich and Reichenbach, Julian and Onnasch, Linda , title =. Journal of Cognitive Engineering and Decision Making , year =

  31. [69]

    Peter , doi =

    Christine Jacob and Noé Brasier and Emanuele Laurenzi and Sabina Heuss and Stavroula Georgia Mougiakakou and Arzu Cöltekin and Marc K. Peter , doi =. AI for IMPACTS Framework for Evaluating the Long-Term Real-World Impacts of AI-Powered Clinician Tools: Systematic Review and Narrative Synthesis , volume =. Journal of Medical Internet Research , keywords =

  32. [71]

    Proceedings of the ACM SIGIR Asia-Pacific Conference on Information Retrieval , year=

    Investigating Variability in Large Language Model-Based Personalized Conversational Information Retrieval , author=. Proceedings of the ACM SIGIR Asia-Pacific Conference on Information Retrieval , year=

  33. [72]

    Proceedings of the ACM International Conference on Web Search and Data Mining (WSDM) , year=

    Writing Style Matters: An Examination of Bias and Fairness in Text Embedding Models , author=. Proceedings of the ACM International Conference on Web Search and Data Mining (WSDM) , year=

  34. [73]

    Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (Tutorial) , year=

    Robust Information Retrieval , author=. Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval (Tutorial) , year=

  35. [74]

    Proceedings of the ACM SIGIR International Conference on the Theory of Information Retrieval (ICTIR) , year=

    Correctness is Not Faithfulness in Retrieval-Augmented Generation Attributions , author=. Proceedings of the ACM SIGIR International Conference on the Theory of Information Retrieval (ICTIR) , year=

  36. [75]

    Towards Trustworthy

    Amatya, Saurav , booktitle=. Towards Trustworthy. 2025 , publisher=

  37. [76]

    Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval , year=

    Rankers, Judges, and Assistants: On the Interaction Between Retrieval Models and Large Language Model-Based Evaluation , author=. Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval , year=

  38. [77]

    Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval (Short/Abstract Track) , year=

    Towards Explainable and Safe Systems for Health Data , author=. Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval (Short/Abstract Track) , year=

  39. [78]

    Proceedings of the ACM SIGIR Asia-Pacific Conference on Information Retrieval (Tutorial) , year=

    Trustworthy Information Retrieval in the Era of Large Language Models , author=. Proceedings of the ACM SIGIR Asia-Pacific Conference on Information Retrieval (Tutorial) , year=

  40. [79]

    Mayer and James H

    Roger C. Mayer and James H. Davis and F. David Schoorman , journal =. An Integrative Model of Organizational Trust , urldate =

  41. [81]

    The PRISMA 2020 statement: an updated guideline for reporting systematic reviews , volume =

    Page, Matthew J and McKenzie, Joanne E and Bossuyt, Patrick M and Boutron, Isabelle and Hoffmann, Tammy C and Mulrow, Cynthia D and Shamseer, Larissa and Tetzlaff, Jennifer M and Akl, Elie A and Brennan, Sue E and Chou, Roger and Glanville, Julie and Grimshaw, Jeremy M and Hr. The PRISMA 2020 statement: an updated guideline for reporting systematic review...

  42. [91]

    Eeyore: Realistic Depression Simulation via Expert-in-the-Loop Supervised and Preference Optimization

    Liu, Siyang and Brie, Bianca and Li, Wenda and Biester, Laura and Lee, Andrew and Pennebaker, James and Mihalcea, Rada. Eeyore: Realistic Depression Simulation via Expert-in-the-Loop Supervised and Preference Optimization. Findings of the Association for Computational Linguistics: ACL 2025. 2025. doi:10.18653/v1/2025.findings-acl.707

  43. [94]

    P sycho GAT : A Novel Psychological Measurement Paradigm through Interactive Fiction Games with LLM Agents

    Yang, Qisen and Wang, Zekun and Chen, Honghui and Wang, Shenzhi and Pu, Yifan and Gao, Xin and Huang, Wenhao and Song, Shiji and Huang, Gao. P sycho GAT : A Novel Psychological Measurement Paradigm through Interactive Fiction Games with LLM Agents. Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Pape...

  44. [95]

    Cactus: Towards Psychological Counseling Conversations using Cognitive Behavioral Theory

    Lee, Suyeon and Kim, Sunghwan and Kim, Minju and Kang, Dongjin and Yang, Dongil and Kim, Harim and Kang, Minseok and Jung, Dayi and Kim, Min Hee and Lee, Seungbeen and Chung, Kyong-Mee and Yu, Youngjae and Lee, Dongha and Yeo, Jinyoung. Cactus: Towards Psychological Counseling Conversations using Cognitive Behavioral Theory. Findings of the Association fo...

  45. [97]

    Depression Detection on Social Media with Large Language Models

    Lan, Xiaochong and Han, Zhiguang and Cheng, Yiming and Sheng, Li and Feng, Jie and Gao, Chen and Li, Yong. Depression Detection on Social Media with Large Language Models. Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track. 2025. doi:10.18653/v1/2025.emnlp-industry.151

  46. [98]

    AIHLEG. 2019. https://digital-strategy.ec.europa.eu/en/library/ethics-guidelines-trustworthy-ai Ethics guidelines for trustworthy AI . Technical report

  47. [99]

    Mehmet Emin Aktan, Zeynep Turhan, and İlknur Dolu. 2022. https://doi.org/10.1016/j.chb.2022.107273 Attitudes and perspectives towards the preferences for artificial intelligence in psychotherapy . Computers in Human Behavior, 133:107273

  48. [100]

    Alghamdi, Reem Masoud, Deema Alnuhait, Afnan Y

    Emad A. Alghamdi, Reem Masoud, Deema Alnuhait, Afnan Y. Alomairi, Ahmed Ashraf, and Mohamed Zaytoon. 2025. https://aclanthology.org/2025.coling-main.579/ A ra T rust: An evaluation of trustworthiness for LLM s in A rabic . In Proceedings of the 31st International Conference on Computational Linguistics, pages 8664--8679, Abu Dhabi, UAE. Association for Co...

  49. [101]

    AMA. 2024. https://www.ama-assn.org/system/files/ama-ai-principles.pdf Augmented intelligence development, deployment, and use in health care . Technical report

  50. [102]

    humans are still in the decision loop

    Naomi Aoki. 2021. https://doi.org/10.1016/j.chb.2020.106572 The importance of the assurance that “humans are still in the decision loop” for public trust in artificial intelligence: Evidence from an online experiment . Computers in Human Behavior, 114:106572

  51. [103]

    APA. 2025. Ethical guidance for ai in the professional practice of health service psychology

  52. [104]

    Abeer Badawi, Md Tahmid Rahman Laskar, Jimmy Huang, Shaina Raza, and Elham Dolatabadi. 2025. https://openreview.net/forum?id=j3totqf8xW Position: Beyond assistance reimagining LLM s as ethical and adaptive co-creators in mental health care . In Forty-second International Conference on Machine Learning Position Paper Track

  53. [105]

    Miguel Baidal, Erik Derner, and Nuria Oliver. 2025. https://doi.org/10.18653/v1/2025.nlp4pi-1.2 Guardians of trust: Risks and opportunities for LLM s in mental health . In Proceedings of the Fourth Workshop on NLP for Positive Impact (NLP4PI), pages 11--22, Vienna, Austria. Association for Computational Linguistics

  54. [106]

    Krisztian Balog, Donald Metzler, and Tao Qin. 2025. Rankers, judges, and assistants: On the interaction between retrieval models and large language model-based evaluation. In Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM

  55. [107]

    Guanqun Bi, Zhuang Chen, Zhoufu Liu, Hongkai Wang, Xiyao Xiao, Yuqiang Xie, Wen Zhang, Yongkang Huang, Yuxuan Chen, Libiao Peng, and Minlie Huang. 2025. https://doi.org/10.18653/v1/2025.findings-acl.1278 MAGI : Multi-agent guided interview for psychiatric assessment . In Findings of the Association for Computational Linguistics: ACL 2025, pages 24898--249...

  56. [108]

    Sabine Brunswicker, Yifan Zhang, Christopher Rashidian, and Daniel W. Linna. 2025. https://doi.org/10.1016/j.chb.2024.108516 Trust through words: The systemize-empathize-effect of language in task-oriented conversational agents . Computers in Human Behavior, 165:108516

  57. [109]

    Daniel Cabrera Lozoya, Eloy Hernandez Lua, Juan Alberto Barajas Perches, Mike Conway, and Simon D ' Alfonso. 2025. https://doi.org/10.18653/v1/2025.clpsych-1.13 Synthetic empathy: Generating and evaluating artificial psychotherapy dialogues to detect empathy in counseling sessions . In Proceedings of the 10th Workshop on Computational Linguistics and Clin...

  58. [110]

    can i have my friend attending with me?

    Jiashuo Cao, Yun Suen Pai, Chen Li, Simon Hoermann, and Mark Billinghurst. 2025. " can i have my friend attending with me?": Design implications for using virtual supporters in remote psychotherapy. In Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, pages 1--7

  59. [111]

    Yang Cao and et al. 2025. Writing style matters: An examination of bias and fairness in text embedding models. In Proceedings of the ACM International Conference on Web Search and Data Mining (WSDM). ACM

  60. [112]

    Yirong Chen, Xiaofen Xing, Jingkai Lin, Huimin Zheng, Zhenyu Wang, Qi Liu, and Xiangmin Xu. 2023. https://doi.org/10.18653/v1/2023.findings-emnlp.83 S oul C hat: Improving LLM s' empathy, listening, and comfort abilities through fine-tuning with multi-turn empathy conversations . In Findings of the Association for Computational Linguistics: EMNLP 2023, pa...

  61. [113]

    Oscar Hengxuan Chi, Shizhen Jia, Yafang Li, and Dogan Gursoy. 2021. https://doi.org/10.1016/j.chb.2021.106700 Developing a formative scale to measure consumers’ trust toward interaction with artificially intelligent ( AI ) social robots in service delivery . Computers in Human Behavior, 118:106700

  62. [114]

    Young Min Cho, Sunny Rai, Lyle Ungar, Jo \ a o Sedoc, and Sharath Guntuku. 2023. https://doi.org/10.18653/v1/2023.emnlp-main.698 An integrative survey on mental health conversational agents to bridge computer science and medical perspectives . In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 11346--11369, Si...

  63. [115]

    Ryuhaerang Choi, Taehan Kim, Subin Park, Jennifer G Kim, and Sung-Ju Lee. 2025. Private yet social: How llm chatbots support and challenge eating disorder recovery. In Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems, pages 1--19

  64. [116]

    Shehzaad Dhuliawala, Vil \'e m Zouhar, Mennatallah El-Assady, and Mrinmaya Sachan. 2023. https://doi.org/10.18653/v1/2023.emnlp-main.339 A diachronic perspective on user trust in AI under uncertainty . In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 5567--5580, Singapore. Association for Computational Linguistics

  65. [117]

    EPEUCO. 2024. https://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:32024R1689 Artificial intelligence act

  66. [118]

    FDA. 2021. www.fda.gov Artificial intelligence/machine learning (ai/ml)-based software as a medical device (samd) action plan

  67. [119]

    Saadia Gabriel, Isha Puri, Xuhai Xu, Matteo Malgaroli, and Marzyeh Ghassemi. 2024. https://doi.org/10.18653/v1/2024.findings-emnlp.120 Can AI relate: Testing large language model response for mental health support . In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 2206--2221, Miami, Florida, USA. Association for Computationa...

  68. [120]

    Felix Gille, Laura Maaß, Benjamin Ho, and Divya Srivastava. 2025. https://doi.org/10.2196/59111 From theory to practice: Viewpoint on economic indicators for trust in digital health . Journal of Medical Internet Research, 27

  69. [121]

    Melanie Goisauf, M \'o nica Cano Abad \'i a, Kaya Aky \"u z, Maciej Bobowicz, Alena Buyx, Ilaria Colussi, Marie-Christine Fritzsche, Karim Lekadir, Pekka Marttinen, Michaela Th Mayrhofer, and Janos Meszaros. 2025. https://doi.org/10.2196/71236 Trust, trustworthiness, and the future of medical ai: Outcomes of an interdisciplinary expert workshop

  70. [122]

    Sujatha Gollapalli, Beng Ang, and See-Kiong Ng. 2023. https://doi.org/10.18653/v1/2023.findings-emnlp.792 Identifying Early Maladaptive Schemas from mental health question texts . In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 11832--11843, Singapore. Association for Computational Linguistics

  71. [123]

    Wenyue Hua, Xianjun Yang, Mingyu Jin, Zelong Li, Wei Cheng, Ruixiang Tang, and Yongfeng Zhang. 2024. https://doi.org/10.18653/v1/2024.findings-emnlp.585 T rust A gent: Towards safe and trustworthy LLM -based agents . In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 10000--10016, Miami, Florida, USA. Association for Computati...

  72. [124]

    Yue Huang, Lichao Sun, Haoran Wang, Siyuan Wu, Qihui Zhang, Yuan Li, Chujie Gao, Yixin Huang, Wenhan Lyu, Yixuan Zhang, Xiner Li, Hanchi Sun, Zhengliang Liu, Yixin Liu, Yijue Wang, Zhikun Zhang, Bertie Vidgen, Bhavya Kailkhura, Caiming Xiong, and 52 others. 2024. Position: Trustllm: trustworthiness in large language models. In Proceedings of the 41st Inte...

  73. [125]

    Weiwei Huo, Guanghui Zheng, Jiaqi Yan, Le Sun, and Liuyi Han. 2022. https://doi.org/10.1016/j.chb.2022.107253 Interacting with medical artificial intelligence: Integrating self-responsibility attribution, human–computer trust, and personality . Computers in Human Behavior, 132:107253

  74. [126]

    Migyeong Kang, Goun Choi, Hyolim Jeon, Ji Hyun An, Daejin Choi, and Jinyoung Han. 2024. https://doi.org/10.18653/v1/2024.emnlp-main.994 CURE : Context- and uncertainty-aware mental disorder detection . In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 17924--17940, Miami, Florida, USA. Association for Computa...

  75. [127]

    Janne Kauttonen, Rebekah Rousi, and Ari Alamäki. 2025. https://doi.org/10.2196/65567 Trust and acceptance challenges in the adoption of AI applications in health care: Quantitative survey analysis . Journal of Medical Internet Research, 27

  76. [128]

    Jabari Kwesi, Jiaxun Cao, Riya Manchanda, and Pardis Emami-Naeini. 2025. Exploring user security and privacy attitudes and concerns toward the use of general-purpose llm chatbots for mental health. In Proceedings of the 34th USENIX Conference on Security Symposium, SEC '25, USA. USENIX Association

  77. [129]

    Jamie Lee, Kyuha Jung, Erin Gregg Newman, Emilie Chow, and Yunan Chen. 2025. Understanding adolescents' perceptions of benefits and risks in health AI technologies through design fiction. In Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems, pages 1--20

  78. [130]

    Lee and Neville Moray

    John D. Lee and Neville Moray. 1992. https://doi.org/10.1080/00140139208967392 Trust, control strategies and allocation of function in human--machine systems . Ergonomics, 35(10):1243--1270

  79. [131]

    Benedikt Leichtmann, Christina Humer, Andreas Hinterreiter, Marc Streit, and Martina Mara. 2023. https://doi.org/10.1016/j.chb.2022.107539 Effects of explainable artificial intelligence on trust and human behavior in a high-risk decision task . Computers in Human Behavior, 139:107539

  80. [132]

    Anqi Li, Yu Lu, Nirui Song, Shuai Zhang, Lizhi Ma, and Zhenzhong Lan. 2024 a . https://doi.org/10.18653/v1/2024.findings-emnlp.69 Understanding the therapeutic relationship between counselors and clients in online text-based counseling using LLM s . In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 1280--1303, Miami, Florida,...

Showing first 80 references.