Published on in Vol 27 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/84918, first published .
Factors Influencing Adoption of Large Language Models in Health Care: Multicenter Cross-Sectional Mixed Methods Observational Study

Factors Influencing Adoption of Large Language Models in Health Care: Multicenter Cross-Sectional Mixed Methods Observational Study

Factors Influencing Adoption of Large Language Models in Health Care: Multicenter Cross-Sectional Mixed Methods Observational Study

1Department of Thoracic Surgery, Guizhou Provincial People's Hospital, No. 83, Zhongshan East Road, Guiyang, Guizhou, China

2NHC Key Laboratory of Pulmonary Immunological Diseases, Guizhou Provincial People's Hospital, Guiyang, Guizhou, China

3Department of Cardio-Thoracic Surgery, Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, Guangdong, China

4Department of Anesthesiology, Third Xiangya Hospital, Central South University, Changsha, Hunan, China

5Department of Thoracic Surgery, Jiangxi Provincial Cancer Hospital, Nanchang, Jiangxi, China

6Department of Dermatology, University of Hong Kong-Shenzhen Hospital, Shenzhen, Guangdong, China

7Department of Respiratory Medicine, Guizhou Provincial People's Hospital, Guiyang, Guizhou, China

8The Second Clinical Medical College, Guangdong Medical University, Dongguan, Guangdong, China

*these authors contributed equally

Corresponding Author:

Xiongwen Yang, MD, PhD


Background: Large language models (LLMs) such as ChatGPT are transforming how health information is accessed, communicated, and applied. However, their adoption in health care remains limited by uncertainties surrounding trust, privacy, and digital readiness, particularly in low- and middle-income contexts.

Objective: This study aimed to examine how trust, information behavior, and sociotechnical readiness influence the willingness of health care professionals (HCPs) and patients or caregivers (PCs) in China to adopt LLMs for medical information and decision support.

Methods: We conducted a multicenter, cross-sectional mixed methods observational study across five tertiary hospitals, combining surveys of 240 HCPs and 480 PCs with semistructured interviews (n=30). Quantitative analyses included logistic regression (LR), random forest (RF), and extreme gradient boosting models with Shapley additive explanations-based interpretability. Qualitative data were thematically analyzed to capture role-specific concerns and expectations.

Results:: Among HCPs, mean age 39.9 (SD 6.5 years; 159/240, 66.2% physicians), 69.2% (166/240) were aware of LLMs and 36.7% (88/240) had previous experience. Among PCs (mean age 50.1, SD 12.6 years; 242/480, 50.4% male), only 26% (125/480) had previous exposure. Trust, perceived usefulness, and digital readiness were the strongest facilitators of adoption. Multivariable models identified trust as the dominant predictor for both groups (HCPs: odds ratio [OR] 3.78, 95% CI 2.15‐6.63; PCs: OR 36.34, 95% CI 18.41‐71.74; P<.001). For HCPs, previous use (OR 5.61, 95% CI 3.02‐10.44; P<.001) and legal clarity (OR 1.56, 95% CI 1.07‐2.27; P=.02) increased willingness, while privacy concerns reduced it (OR 0.72, 95% CI 0.53‐0.97; P=.03). Among PCs, perceived usefulness (OR 2.01, 95% CI 1.52‐2.67; P<.001), education, and digital tool use were positive predictors. Model performance was high (area under the receiver operating characteristic curve [AUC] 0.83‐0.85 for HCPs and 0.93‐0.96 for PCs). Qualitative findings identified 11 themes: HCPs stressed workflow integration and accountability, while PCs emphasized comprehensibility, reassurance, and equitable access; trust consistently linked technical credibility with social legitimacy.

Conclusions: Adoption of LLMs in health care depends less on algorithmic performance than on the management of trust, literacy, and institutional readiness. Trust functions as a multidimensional construct rooted in transparency, reliability, and contextual validation. Theoretically, this study extends technology adoption frameworks by embedding ethical trust, digital literacy, and institutional support within a unified sociotechnical readiness model, advancing information management theory beyond performance-centric paradigms. Empirically, trust and perceived usefulness outweighed demographic or structural factors, with predictive accuracy exceeding 0.9 across user groups. Practically, these findings offer actionable guidance for the design and governance of artificial intelligence systems, emphasizing role-sensitive interfaces, plain-language communication, and transparent accountability mechanisms to promote equitable and trustworthy adoption.

J Med Internet Res 2025;27:e84918

doi:10.2196/84918

Keywords



The rapid advancement of large language models (LLMs), such as ChatGPT (OpenAI) and other generative artificial intelligence (AI) tools, has opened new opportunities in the management and use of information across multiple sectors, including health care [1-4]. These models, trained on vast datasets of biomedical literature and clinical texts, are increasingly capable of performing tasks that fundamentally reshape the way information is accessed and applied, such as explaining medical concepts, summarizing patient records, supporting decision-making, and facilitating communication between patients and professionals [5-10]. Their accessibility, versatility, and potential for integration into organizational workflows highlight LLMs as a significant force in transforming information practices, with implications for efficiency, equity, and engagement [11].

In the field of information systems, the integration of LLMs into enterprise and knowledge management systems is shown to optimize data processing and automate business workflows, marking a new era of intelligent data capabilities [12]. Research in knowledge management has documented how generative AI enhances knowledge transfer, decision-making efficiency, and operational effectiveness through improved content creation and knowledge dissemination [13,14]. In the realm of information systems research, generative AI has been identified as a transformative technology with profound societal and organizational implications, calling for re-evaluation of research agendas and considering its evolving influence on business and social systems [15]. Additionally, LLMs are being deployed to support knowledge workers, streamlining complex tasks through natural language interfaces and enabling new kinds of workflow integration [16,17].

However, the integration of LLMs into sensitive domains such as health care raises critical questions for information management theory and practice. Concerns about accuracy, data privacy, transparency, embedded biases, and accountability are by no means resolved, reflecting broader challenges in governing AI-augmented information systems. Recent studies have shown that the trustworthiness, explainability, and ethical transparency of generative AI remain the decisive barriers for clinical and public adoption, especially when users lack institutional guidance or previous exposure to AI tools [18-21]. Information management research emphasizes that the governance of information artifacts in high-stakes environments must address these sociotechnical risks at multiple levels, from technology design to organizational oversight [22,23]. Importantly, successful adoption of LLMs depends not only on technical performance but also on stakeholder perceptions and acceptance.

Emerging literature now highlights trust, fairness, and data governance as critical enablers of responsible AI implementation in health and public services [24-27]. Understanding how health care professionals (HCPs) and patients or caregivers (PCs) assess usefulness, trustworthiness, and risk is essential, not only to ensure that LLMs enhance information use and knowledge exchange but also to prevent them from reinforcing existing inequalities or mistrust [28].

Previous studies in high- income contexts have begun to explore how clinicians interact with AI-generated information. However, limited evidence exists from low and middle-income countries, where factors like digital literacy, information trust, and institutional readiness critically shape technology adoption. A systematic scoping review highlights that AI initiatives in low- and middle-income countries’ health care, such as clinical decision support and chatbots, face persistent challenges, including reliability, user-friendliness, context misalignment, and distrust of AI systems [29]. Recent empirical evidence further emphasizes that digital readiness, cultural values, and governance frameworks substantially moderate AI acceptance across diverse health care environments, suggesting the need for region-specific analyses to guide equitable adoption [30-33].

Specifically in China, the socioeconomic diversity between urban and rural regions suggests that attitudes toward LLMs may vary significantly. Insights from information management research emphasize the digital divide in low- and middle-income countries, pointing to disparities not merely in access but in digital capacity, health literacy, and sociotechnical support systems [34,35]. Therefore, examining LLM adoption in China becomes an inquiry not just into technology implementation, but into information behavior and management within complex sociocultural and infrastructural contexts [36]. Given China’s unique ecosystem, where rapid digital transformation intersects with uneven regional literacy and regulatory evolution, understanding user trust and institutional readiness offers critical insights for both national policy and global information systems theory [37-40].

Against this backdrop, this study aims to systematically examine how trust, information behavior, and sociotechnical readiness influence the willingness of HCPs and PCs in China to adopt LLMs for medical information access and decision support. Specifically, we investigate the relative importance of psychological, informational, and institutional factors in shaping adoption intentions. We hypothesize that trust and digital readiness, rather than demographic attributes, serve as the strongest predictors of willingness to adopt LLMs. By using a multicenter mixed methods design, this study provides empirical evidence that bridges theory, governance, and real-world implementation, offering insights for both information management theory and responsible AI deployment in health care.


Reporting Standards

The study was reported in accordance with the Good Reporting of A Mixed Methods Study (GRAMMS) guideline [41], which provides a comprehensive framework for ensuring transparency, integration, and methodological rigor in mixed methods research (Multimedia Appendix 1). This approach was chosen to better reflect the interconnected design and analytical structure of the quantitative and qualitative components of this study.

Study Design and Setting

We conducted a multicenter, cross-sectional mixed methods observational study across 5 tertiary hospitals located in 4 provinces of China—Guangdong, Hunan, Jiangxi, and Guizhou—between August 2024 and March 2025. The study was initiated and centrally designed by the research team at Guizhou Provincial People’s Hospital, with all other hospitals participating as independent data collection sites under a unified protocol. The participating institutions were the Third Affiliated Hospital of Sun Yat-sen University, the Third Xiangya Hospital of Central South University, the University of Hong Kong–Shenzhen Hospital, and Jiangxi Cancer Hospital.

These sites were purposively selected to ensure diversity in regional economic development, digital infrastructure, and patient population characteristics. Each hospital facilitated participant recruitment and on-site data collection through its own clinical departments, while all data management, integration, and analysis were performed by the coordinating research team at Guizhou Provincial People’s Hospital.

Data collection occurred in parallel across all sites to minimize temporal bias and included both quantitative surveys and qualitative interviews conducted on-site or through secure teleconferencing platforms.

Participant Recruitment

Participants were recruited consecutively from outpatient and inpatient departments of the 5 hospitals. Health care professionals were invited through internal departmental briefings, institutional WeChat (Tencent) groups, and email lists, while PCs were approached in clinical waiting areas and wards by trained research assistants.

Eligibility criteria required health care professionals to be licensed physicians or registered nurses currently engaged in clinical duties, and patients or caregivers to be adults (≥18 years) receiving or accompanying medical care during the study period. Individuals unable to communicate in Mandarin or with cognitive impairment were excluded.

Recruitment followed a purposive-consecutive hybrid approach to ensure representation across different departments, hospital types, and regional economic levels. The final sample size (240 health care professionals and 480 patients or caregivers) was consistent with power analysis targets and provided sufficient heterogeneity for multivariate modeling and subgroup comparisons.

Data collection was conducted on-site using either paper-based or secure QR code–linked electronic questionnaires, according to local logistical capacity. Research assistants were trained to ensure standardized recruitment, eligibility screening, and data integrity across all centers.

Assessments and Data Sources

The overall study flow is illustrated in Figure 1. In total, 2 structured questionnaires were developed separately for health care professionals and patients or caregivers (Tables S1 and S2 in Multimedia Appendix 2). Variables included demographic and contextual characteristics (age, gender, region, education, role, and clinical experience), digital tool use, previous exposure to AI or LLMs, and perceptions related to (1) perceived usefulness (expected benefits of LLMs for information tasks), (2) trust and perceived ease of use (confidence in reliability and accessibility of LLM outputs), (3) behavioral intention (willingness to use LLMs in information-seeking or communication scenarios), and (4) facilitating conditions (eg, privacy concerns, legal accountability, and digital infrastructure).

Figure 1. Overview of the study design, methods, and analytical framework.HCP: health care professional; LLM: large language models; PC: patients or caregivers.

Figure 1 illustrates the overall workflow of a multicenter, cross-sectional mixed methods observational study conducted across 5 tertiary hospitals in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China, between August 2024 and March 2025. The study included two participant groups, HCPs and PCs, with parallel questionnaire development and data collection. A total of 260 HCPs and 520 PCs were screened, of whom 240 HCPs and 480 PCs met eligibility criteria and completed the quantitative survey. From these, 15 HCPs and 15 PCs were purposively sampled for qualitative interviews. Quantitative data were analyzed using statistical and regression modeling, while qualitative data underwent thematic coding and interpretation. Findings were integrated through mixed methods triangulation to generate consolidated insights and practical recommendations for trust-based and equitable adoption of LLMs in health care.

All items were rated on 5-point Likert scales (1=strongly disagree and 5=strongly agree). The outcome variable is willingness to adopt LLMs (binary). Predictors include trust, perceived usefulness, digital literacy, previous exposure, and socioeconomic context. Potential confounders include age, gender, education, and institutional type.

Data were collected using secure QR code–based online surveys or paper forms, depending on site logistics. Trained assistants supported participants with limited literacy to ensure data integrity. The instruments were validated via expert review, pilot testing (n=30), and Cronbach α reliability (>0.85 across constructs).

Qualitative Interviews

To complement quantitative findings, we conducted semistructured interviews with 30 participants (15 HCPs and 15 PCs) purposively selected based on demographic and digital literacy diversity. Interviews lasted 30‐45 minutes, were conducted in private rooms or encrypted online sessions, and followed a guide covering (1) perceived benefits and risks of LLMs; (2) trust, accuracy, and data privacy; (3) expectations for accountability and ethical safeguards; and (4) anticipated impact on patient-provider communication.

Interviews were conducted on-site by trained research assistants at each participating hospital, following a standardized interview guide. A total of two external qualitative experts—a public health researcher and a clinical psychologist, both experienced in health communication research—subsequently reviewed the audio recordings and deidentified transcripts to evaluate thematic coherence and methodological rigor. The interview guide was pilot-tested among 3 participants to ensure clarity and relevance. No repeat interviews were conducted. Data saturation was reached when no new themes emerged after approximately 25 interviews. Transcripts were not returned to participants for comment.

All interviews were audio recorded with consent, transcribed verbatim, deidentified, and analyzed inductively using NVivo (NVivo qualitative data analysis software, QSR International) and the R Qualitative Data Analysis package. In total, 2 coders independently conducted open and axial coding, achieving intercoder reliability (κ=0.82). An inductive thematic analysis approach guided by grounded theory principles was adopted to identify key patterns and relationships. Thematic analysis yielded major categories and subthemes, supported by visualization techniques (node diagrams, frequency plots, and Sankey diagrams).

Quantitative and Predictive Analyses

Quantitative analyses were performed using Python (version 3.12) and R software (version 4.3, R Foundation for Statistical Computing). Descriptive statistics were reported as means (SDs) for continuous variables and frequencies (%) for categorical variables. Group differences were assessed using Chi-square tests and independent-sample t tests (2-sided) as appropriate.

No missing data were identified in the dataset. All questionnaire items were mandatory in the electronic version, and paper-based responses were reviewed immediately by trained research assistants to ensure completeness. As such, the final dataset contained 100% valid observations across all variables, and no imputation or missing completely at random testing was required.

To identify determinants of adoption intent, we estimated multivariable logistic regression (LR) models, controlling for demographic and contextual covariates. To complement regression, 2 machine learning approaches—random forest (RF) and extreme gradient boosting (XGBoost)—were implemented to explore nonlinear patterns and improve the prediction of willingness to use LLMs. Model training used stratified sampling (HCPs: 168/240 and PCs: 336/480, 70% training and HCPs 72/240 and PCs 144/480, 30% validation), with 5-fold cross-validation applied to reduce overfitting.

Model performance was assessed primarily through discrimination (area under the receiver operating characteristic curve [AUC]) and calibration (Brier score and calibration plots). While conventional clinical decision-making metrics (eg, decision curve analysis) were not central to this study, additional evaluation measures such as F1-score, precision, recall, and accuracy were included to examine classification quality. Variable importance rankings and Shapley additive explanation (SHAP) values were used to enhance interpretability, highlighting the most influential factors shaping adoption intent (eg, digital literacy, trust, and privacy concerns). An overview of the statistical modeling and analytical procedures is provided in Multimedia Appendix 3.

Ethical Considerations

Ethics Review and Approval

This study was reviewed and approved by the institutional review boards or ethics committees of all participating hospitals, including Guizhou Provincial People’s Hospital (ethics number: 2024004), the Third Affiliated Hospital of Sun Yat-sen University (ethics number: B2023074), the Third Xiangya Hospital of Central South University (ethics number: 2024011), the University of Hong Kong–Shenzhen Hospital (ethics number: B2024021), and Jiangxi Cancer Hospital (ethics number: JC2024006). The research protocol complied with the Declaration of Helsinki (2013 revision) and local institutional regulations governing research involving human participants.

Informed Consent

All participants provided written informed consent before participation. The consent process included clear explanations of study aims, data collection procedures, voluntary participation, and withdrawal rights. For the qualitative component, participants were also informed of audio recording and transcription procedures. No waiver of consent was applied.

Privacy and Confidentiality

Participant data were anonymized and deidentified before analysis. No personally identifying information (eg, names, contact details, and hospital ID numbers) was collected. Data were stored on password-protected institutional servers with access restricted to authorized research team members only. All analyses were conducted using deidentified datasets to ensure confidentiality and compliance with institutional data protection standards.

Participant Compensation

Participants did not receive financial compensation for participation. However, small tokens of appreciation (eg, stationery or refreshments of RMB ¥30, approximately US $4) were provided in accordance with local ethical norms to acknowledge participants’ time.

Image and Identifiable Data Protection

No images or materials containing identifiable personal information are included in the manuscript or supplementary materials. All figures are based on aggregated or simulated data. If any future publication materials include identifiable content, appropriate written consent will be obtained and submitted to the editorial office in accordance with Journal of Medical Internet Research policies.


Participant Characteristics

Table 1 summarizes the sociodemographic and professional characteristics of HCPs (eg, role, department, years of experience, and attitudes toward digital tools), while Table 2 presents PCs demographic features (eg, age, gender, education, digital literacy, and perspectives on AI-assisted interpretation).

Table 1. Characteristics of health care professionals (HCPs, n=240).
VariablesValues
Sex, n (%)
 Male141 (58.8)
 Female99 (41.3)
Age (years), mean (SD)39.9 (6.5)
Profession, n (%)
 Physician159 (66.2)
 Nurse81 (33.8)
Years of clinical experience, mean (SD)9.9 (4.8)
Department, n (%)
 Internal medicine54 (22.5)
 Surgery41 (17.1)
 ICUa47 (19.6)
 Emergency60 (24.0)
 Other38 (15.8)
Hospital affiliation, n (%)
 Nonaffiliated101 (42.1)
 Affiliated139 (57.9)
Hospital type, n (%)
 Oncology44 (18.3)
 General196 (81.7)
Province economic level, n (%)
 Low46 (19.2)
 Medium101 (42.1)
 High93 (38.7)
Frequency of digital tool use, n (%)
 Rarely38 (15.8)
 Occasionally85 (35.4)
 Often117 (48.8)
Prior awareness of LLMsb, n (%)
 Yes166 (69.2)
 No74 (30.8)
Prior use of LLMs, n (%)
 Yes88 (36.7)
 No152 (63.3)
Received AIc or informatics training, n (%)
 Yes92 (38.3)
 No148 (61.7)
Trust in LLMs (1-5), mean (SD)3.09 (1.21)
Perceived usefulness (1-5), mean (SD)2.98 (1.25)
Privacy concerns (1-5), mean (SD)3.23 (1.33)
Clarity on legal responsibility (1-5), mean (SD)3.11 (1.32)

aICU: intensive care unit.

bLLM: large language model.

cAI: artificial intelligence.

Table 2. Characteristics of patients and caregivers (PCs, n=480).
VariablesValues
Sex, n (%)
 Male242 (50.4)
 Female238 (49.6)
Age (years), mean (SD)50.1 (12.6)
Role, n (%)
 Patient360 (75.0)
 Caregiver120 (25.0)
Education level, n (%)
 Primary school or below121 (25.2)
 Junior high school100 (20.8)
 High school96 (20.0)
 College91 (19.0)
 Bachelor’s degree or above72 (15.0)
Hospital affiliation, n (%)
 Nonaffiliated197 (41.0)
 Affiliated283 (59.0)
Hospital type, n (%)
 Oncology94 (19.6)
 General386 (80.4)
Province economic level, n (%)
 Low88 (18.3)
 Medium202 (42.1)
 High190 (39.6)
Frequency of digital tool use, n (%)
 Rarely95 (19.8)
 Occasionally162 (33.7)
 Often223 (46.5)
Previous awareness of LLMsa, n (%)
 Yes247 (51.5)
 No233 (48.5)
Previous use of LLMs, n (%)
 Yes125 (26.0)
 No355 (74.0)
Trust in LLMs (1-5), mean (SD)2.83 (1.22)
Perceived usefulness (1-5), mean (SD)2.61 (1.14)
Privacy concerns (1-5), mean (SD)2.72 (1.29)

aLLM: large language model.

In Table 1, data were collected from 5 tertiary hospitals located in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China, between August 2024 and March 2025. Variables include age, gender, profession, years of clinical experience, department type, hospital affiliation and type, provincial economic level, frequency of digital tool use, prior awareness and use of LLMs, and receipt of AI or informatics training. In addition, self-reported perceptions—trust in LLMs, perceived usefulness, privacy concern, and clarity on legal responsibility—were measured using a 5-point Likert scale (1=strongly disagree and 5=strongly agree).

Table 2 presents the demographic, educational, and digital behavior characteristics of patients or caregivers who participated in a multicenter, cross-sectional mixed methods observational study on the adoption of LLMs in health care. Data were collected from 5 tertiary hospitals located in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China, between August 2024 and March 2025. Variables include age, gender, education level, role (patient or caregiver), hospital affiliation and type, provincial economic level, frequency of digital tool use, previous awareness and use of LLMs, and attitudinal measures—including trust in LLMs, perceived usefulness, and privacy concerns—assessed using a 5-point Likert scale (1=strongly disagree and 5=strongly agree).

A total of 720 individuals (240/720 HCPs and 480/720 PCs) participated in the study (Tables 1 and 2). Among HCPs, two-thirds were physicians (159/240, 66.2%) and one-third nurses (81/240, 33.8%), with an average age of 39.9 (SD 6.5) years and 9.9 (SD 4.8) years of clinical experience. Participants were drawn from diverse departments, with representation from internal medicine, surgery, intensive care, and emergency care. Almost half (117/240, 48.8%) reported frequent use of digital tools, and 166/240 (69.2%) were already aware of LLMs, though only 36.7% (88/240) had previous experience using them.

Patients or caregivers were evenly distributed by sex (242/480, 50.4% male), with a mean age of 50.1 (SD 12.6) years. Educational attainment varied considerably: 121/480 (25.2%) PCs had only primary schooling or below, while 15.0% (72/480) held a Bachelor’s degree or higher. Frequent digital tool use was reported by 15.0% (72/480), and about half (247/480, 51.5%) had heard of LLMs, though only 26.0% (125/480) had ever used them. These characteristics reflected broad heterogeneity in digital readiness and information familiarity across both user groups.

Group Differences in Perceptions of LLMs

Perceptions of LLMs varied significantly between HCPs and PCs. On a 5-point Likert scale, HCPs rated perceived usefulness higher (mean 2.98, 95% CI 2.86‐3.10 vs mean 2.61, 95% CI 2.52‐2.70; P<.001) and expressed stronger privacy concerns (mean 3.23, 95% CI 3.12‐3.34 vs mean 2.72, 95% CI 2.63‐2.81; P<.001). Trust in LLMs, however, was comparable between groups (mean 3.09, 95% CI 2.97‐3.21 vs mean 2.83, 95% CI 2.72‐2.94; P=.20). These results suggest that HCPs view LLMs as potentially valuable but remain cautious about risks, while PCs perceive fewer benefits but also fewer risks (Multimedia Appendix 4).

Correlates of Adoption Willingness

Among HCPs, willingness to adopt LLMs was positively associated with previous awareness and use (P<.01), frequent digital tool use (P=.02), AI or informatics training (P=.03), higher trust (P<.001), and perceived usefulness (P<.001). In contrast, privacy concerns (P=.04) and uncertainty about legal accountability (P=.02) reduced willingness (Table 3).

Table 3. Bivariate comparisons of characteristics by willingness to adopt large language models among health care professionals. Associations were tested using t tests for continuous variables and Chi-square tests for categorical variables.
VariablesWillingUnwilling or uncertainP valuea
Sex, n (%).927
 Male69/116 (59.5)72/124 (58.1)
 Female47/116 (40.5)52/124 (41.9)
Profession, n (%).712
 Physician75/116 (64.7)84/124 (67.7)
 Nurse41/116 (35.3)40/124 (32.3)
Department, n (%).751
 Internal medicine27/116 (23.3)27/124 (21.8)
 Surgery22/116 (19)19/124 (15.3)
 ICUb22/116 (19)25/124 (20.2)
 Emergency25/116 (21.6)35/124 (28.2)
 Other20/116 (17.2)18/124 (14.5)
Hospital affiliation, n (%).137
 Nonaffiliated55/116 (47.4)46/124 (37.1)
 Affiliated61/116 (52.6)78/124 (62.9)
Hospital type, n (%).938
 Oncology22/116 (19)22/124 (17.7)
 General94/116 (81)102/124 (82.3)
Province’s economic level, n (%)<.001
 Low11/116 (9.5)35/124 (28.2)
 Medium34/116 (29.3)59/124 (47.6)
 High71/116 (61.2)30/124 (24.2)
Frequency of digital tool use, n (%)<.001
 Rarely5/116 (4.3)33/124 (26.6)
 Occasionally36/116 (31)49/124 (39.5)
 Often75/116 (64.7)42/124 (33.9)
Previous awareness of LLMsc, n (%)<.001
 Yes93/116 (80.2)73/124 (58.9)
 No23/116 (19.8)51/124 (41.1)
Previous use of LLMs, n (%)<.001
 Yes56/116 (48.3)32/124 (25.8)
 No60/116 (51.7)92/124 (74.2)
Received AId or informatics training, n (%).030
 Yes53/116 (45.7)39/124 (31.5)
 No63/116 (54.3)85/124 (68.5)
Age (years), mean (SD)39.5 (6.4)40.3 (6.6).382
Years of clinical experience, mean (SD)10.7 (5.1)9.2 (4.4).017
Trust in LLMs (1-5), mean (SD)3.20 (1.22)2.78 (1.26).011
Perceived usefulness (1-5), mean (SD)3.70 (1.11)2.52 (1.02)<.001
Privacy concerns (1-5), mean (SD)2.82 (1.20)3.62 (1.33)<.001
Clarity on legal responsibility (1-5), mean (SD)3.41 (1.29)2.84 (1.29)<.001

aP<.05 was considered statistically significant.

bICU: intensive care unit.

cLLM: large language model.

dAI: artificial intelligence.

Table 3 presents bivariate comparisons (Chi-square tests for categorical variables and independent-sample t tests for continuous variables) examining factors associated with willingness to adopt LLMs among health care professionals. Data were derived from a multicenter, cross-sectional mixed methods observational study conducted in 5 tertiary hospitals located in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China, between August 2024 and March 2025. Variables analyzed include age, gender, profession, department, hospital characteristics, provincial economic level, digital tool use frequency, previous awareness and use of LLMs, receipt of AI or informatics training, and attitudinal measures—trust in LLMs, perceived usefulness, privacy concern, and clarity on legal responsibility—assessed on a 5-point Likert scale (1=strongly disagree and 5=strongly agree).

Among PCs, willingness was more strongly shaped by sociodemographic and contextual factors: higher education level (P<.001), better regional economic conditions (P<.001), and frequent digital tool use (P<.001) were all associated with stronger adoption intent. Trust (P<.001) and perceived usefulness (P<.001) were again positive predictors, while privacy concerns acted as deterrents (P=.01). Unlike HCPs, previous LLM exposure was not a significant factor for PCs (Table 4).

Table 4. Bivariate comparisons of characteristics by willingness to adopt large language models among patients and caregivers. Statistical comparisons were performed using t tests for continuous variables and Chi-square tests for categorical variables.
VariablesParticipantsP valuea
Willing (n=187)Unwilling/uncertain (n=293)
Sex, n (%).547
 Male98 (52.4)144 (49.1)
 Female89 (47.6)149 (50.9)
Role, n (%).787
 Patient142 (75.9)218 (74.4)
 Caregiver45 (24.1)75 (25.6)
Education level, n (%)<.001
 Primary school or below19 (10.2)86 (29.4)
 Junior high school28 (15)66 (22.5)
 High school40 (21.4)65 (22.2)
 College58 (31)45 (15.4)
 Bachelor’s degree or above42 (22.5)31 (10.6)
Hospital affiliation, n (%).886
 Nonaffiliated78 (41.7)119 (40.6)
 Affiliated109 (58.3)174 (59.4)
Hospital type, n (%).462
 Oncology33 (17.6)61 (20.8)
 General154 (82.4)232 (79.2)
Province economic level, n (%)<.001
 Low23 (12.3)67 (22.9)
 Medium97 (51.9)61 (20.8)
 High67 (35.8)165 (56.3)
Frequency of digital tool use, n (%)<.001
 Rarely4 (2.1)90 (30.7)
 Occasionally24 (12.8)138 (47.1)
 Often159 (85)65 (22.2)
Previous awareness of LLMsb, n (%).177
 Yes102 (54.5)140 (47.8)
 No85 (45.5)153 (52.2)
Previous use of LLMs, n (%).118
 Yes91 (48.7)120 (41)
 No96 (51.3)173 (59)
Age (years), mean (SD)49.5 (12.0)50.5 (13.0).410
Trust in LLMs (1-5), mean (SD)3.02 (1.18)2.36 (1.04)<.001
Perceived usefulness (1-5), mean (SD)3.97 (0.70)2.10 (0.88)<.001
Privacy concerns (1-5), mean (SD)2.21 (1.77)3.04 (1.25)<.001

aP<.05 considered statistically significant.

bLLM: large language model.

Table 4 presents bivariate comparisons (Chi-square tests for categorical variables and independent-sample t tests, 2-tailed for continuous variables) examining factors associated with willingness to adopt LLMs among PCs. Data were obtained from a multicenter, cross-sectional mixed methods observational study conducted in 5 tertiary hospitals located in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China, between August 2024 and March 2025. Variables analyzed include age, gender, role (PC), education level, hospital affiliation and type, provincial economic level, frequency of digital tool use, previous awareness and use of LLMs, and attitudinal measures—trust in LLMs, perceived usefulness, and privacy concern—assessed on a 5-point Likert scale (1=strongly disagree and 5=strongly agree).

Independent Predictors From Multivariate Modeling

Multivariable LR highlighted both common and distinct predictors across groups.

For HCPs, adoption was most strongly predicted by trust (odds ratio [OR] 3.78, 95% CI 2.15‐6.63; P<.001), previous use (OR 5.61, 95% CI 3.02‐10.44; P<.001), and clarity regarding legal responsibility (OR 1.56, 95% CI 1.07‐2.27; P=.02). Privacy concerns were a negative predictor (OR 0.72, 95% CI 0.53‐0.97; P=.03) (Table S3 in Multimedia Appendix 2 and Multimedia Appendix 5; Figure 2A).

Figure 2. Multivariate logistic regression models identifying predictors of willingness to adopt large language models (A) healthcare professionals, (B) patients or caregivers. Values are shown on a logarithmic scale, with the red dashed line indicating an OR of 1 (no effect). Variables positioned to the right of the line indicate positive associations with large language model adoption willingness, while those to the left indicate negative associations. HCP: health care professional; LLM: large language model; OR: odds ratio; PC: patient or caregiver.

Figure 2 presents forest plots from multivariate logistic regression analyses examining predictors of willingness to adopt LLMs among 2 user groups: HCPs and PCs. Data were derived from a multicenter, cross-sectional mixed methods observational study conducted in 5 tertiary hospitals across Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China, between August 2024 and March 2025. Each plot displays the ORs with 95% CIs for independent variables including demographic characteristics (eg, age, gender, and education), digital behavior (eg, frequency of digital tool use, previous use or awareness of LLMs), and attitudinal factors (eg, trust, perceived usefulness, privacy concern, and legal clarity).

For patients or caregivers, trust (OR 36.34, 95% CI 18.41‐71.74; P<.001) dominated as the single strongest determinant, alongside perceived usefulness (OR 2.01, 95% CI 1.52‐2.67; P<.001). Lower education, infrequent digital tool use, and disadvantaged economic background significantly reduced adoption likelihood (Table S4 in Multimedia Appendix 2; Figure 2B).

These results underscore that attitudinal and informational readiness factors (trust, usefulness, and digital literacy) are more predictive of willingness than static demographic variables.

Model Performance and Interpretability

Across LR, RF, and XGBoost, predictive performance was strong, with AUCs ranging from 0.83‐0.85 among HCPs and 0.93‐0.96 among PCs. Classification metrics (precision, recall, and F1-score) confirmed robustness across models (Figures 3-5; Multimedia Appendices 6 and 7; Table 5; and Tables S5-S8 in Multimedia Appendix 2).

Figure 3. Discriminative and calibration performance of logistic regression, random forest, and extreme gradient boosting models in predicting willingness to adopt large language models. (A) Receiver operating characteristic curves for health care professionals, (B) Receiver operating characteristic curves for patients or caregivers, (C) Calibration curves with 10-bin error bars for health care professionals, (D) Calibration curves with 10-bin error bars for patients or caregivers. (A) and (B) depict ROC curves comparing training, testing, and CV performance across models, with AUC values presented for each subset. Panels (C) and (D) display calibration curves with 10-bin error bars, comparing predicted probabilities to observed outcomes; the dashed diagonal line represents perfect calibration. AUC: area under the receiver operating characteristic curve ; CV: cross-validation.
Figure 4. Model interpretability and cross-model agreement in predicting large language model adoption willingness among health care professionals. Shapley additive explanations bar plot (A) and beeswarm plot (B) for logistic regression. Shapley additive explanations bar plot (C) and beeswarm plot (D) for random forest. SHAP bar plot (E) and beeswarm plot (F) for extreme gradient boosting. (G) Heatmap of feature importance rank consistency across the 3 models. (H) Overlap in the top 3, 5, and 7 ranked features across models. (I) Kendall tau rank correlation matrix assessing agreement in feature importance rankings across models. High concordance of feature rankings (G-I) confirmed model robustness and cross-model agreement, with correlation coefficients exceeding 0.9 across methods. AI: artificial intelligence; HCP: health care professional; ICU: intensive care unit; LLM: large language models; SHAP: Shapley additive explanations; XGBoost: extreme gradient boosting.
Figure 5. Model interpretability and cross-model agreement in predicting large language models’ adoption willingness among patients or caregivers. (A) Shapley additive explanations bar plot and (B) beeswarm plot for logistic regression. (C) Shapley additive explanations bar plot and (D) beeswarm plot for random forest. (E) Shapley additive explanations bar plot and (F) beeswarm plot for extreme gradient boosting. (G) Heatmap of feature importance rank consistency across the three models. (H) Overlap in the top 3, 5, and 7 ranked features across models. (I) Kendall tau rank correlation matrix assessing agreement in feature importance rankings across models. Cross-model comparisons (G-I) demonstrated strong consistency in feature rankings, with high Spearman and Kendall correlation coefficients indicating robust agreement in variable importance across the 3 algorithms. LLM: large language model; PC: patients or caregivers; SHAP: Shapley additive explanations; XGBoost: extreme gradient boosting.

Figure 3 summarizes the discrimination and calibration performance of three predictive models—LR, RF, and XGBoost—used to estimate willingness to adopt LLMs among HCPs and PCs. Data were obtained from a multicenter, cross-sectional mixed methods observational study conducted in five tertiary hospitals across Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China, between August 2024 and March 2025. Brier scores are reported to quantify prediction error. All models were trained using stratified 5-fold CV with a held-out validation set, ensuring balanced sampling and robust generalizability across user groups.

Figure 4 illustrates model interpretability and feature importance comparisons for predicting HCPs' willingness to adopt LLMs in clinical and research contexts. Data were derived from a multicenter cross-sectional survey of 240 physicians and nurses conducted between August 2024 and March 2025 across 5 tertiary hospitals in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China. Three supervised machine learning algorithms—logistic regression, random forest, and XGBoost—were trained to identify key determinants of LLM adoption willingness. Predictor variables encompassed demographic, professional, and attitudinal factors, including trust in LLMs, privacy concern, perceived usefulness, frequency of digital-tool use, prior exposure to AI or informatics training, and clarity of legal responsibility. Across all models, trust in LLMs and privacy concern emerged as the most influential predictors, indicating that sociotechnical and psychological readiness outweighed demographic or institutional attributes in shaping adoption intent.

Figure 5 presents model interpretability and feature-importance comparisons for predicting PCs' willingness to adopt LLMs for medical information seeking and decision support. Data were obtained from a multicenter cross-sectional survey of 480 adult patients and caregivers conducted between August 2024 and March 2025 across 5 tertiary hospitals in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China. Three supervised machine learning algorithms—logistic regression, random forest, and XGBoost—were trained to identify key determinants of LLM adoption willingness. Predictor variables included trust in LLMs, privacy concern, perceived usefulness, digital tool use frequency, education level, prior exposure to AI or LLMs, and regional economic context. Across all models, trust in LLMs consistently emerged as the dominant positive predictor of adoption willingness, whereas privacy concern and low education level were the primary barriers. Frequent digital tool use and higher perceived usefulness further increased the likelihood of adoption. These findings highlight that, among patients and caregivers, trust and digital readiness outweighed demographic or institutional factors in determining openness to adopting LLM-based health information tools.

Table 5. Pairwise DeLong test results for the areas under the receiver operating characteristic curves comparison across models (health care professionals and patients or caregivers).
Model 1Model 2AUCa 1AUC 2ΔAUCP value
Health care professionals
 Logistic regressionXGBoostb0.8310.8250.006.90
 Logistic regressionRandom forest0.8310.846–0.015.70
 XGBoostRandom forest0.8250.846–0.022.30
Patients or caregivers
 Logistic regressionXGBoost0.9530.9380.014.33
 Logistic regressionRandom forest0.9530.957–0.004.78
 XGBoostRandom forest0.9380.957–0.018.02

aAUC: area under the receiver operating characteristic curve.

bXGBoost: extreme gradient boosting.

Table 5 summarizes the results of pairwise DeLong tests comparing the AUCs for 3 predictive models—logistic regression, random forest, and XGBoost—used to estimate the willingness to adopt large language models (LLMs) in health care. Analyses were performed separately for HCPs and PCs based on data from a multicenter, cross-sectional mixed methods observational study conducted in 5 tertiary hospitals across Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China, between August 2024 and March 2025. The DeLong test was applied to assess whether differences in model discrimination (AUC values) were statistically significant, providing a measure of the comparative robustness and generalizability of predictive models across user groups.

Interpretability analyses (SHAP values) consistently identified trust in LLMs as the most influential predictor of adoption across both groups. Among HCPs, previous use, legal clarity, and digital tool use also ranked highly. For PCs, education and regional economic status were more prominent, reflecting underlying disparities in digital readiness. Demographic and institutional factors (eg, age, department, and hospital type) had minimal predictive power.

Marginal Effects and Interaction Patterns

Marginal effects analysis confirmed that trust in LLMs exerted the strongest positive effect on adoption across both groups, while privacy concern consistently deterred adoption. Among HCPs, previous use amplified the effect of trust, whereas among PCs, education and digital literacy had stronger moderating roles. Interaction effects were limited but indicated that combinations such as trust×legal clarity (for HCPs) and usefulness×previous exposure (for PCs) enhanced adoption intent (Figures 67, Multimedia Appendices 5, 8-11).

Figure 6. Marginal effects of key predictors on willingness to adopt large language models among health care professionals. (A) Interaction between trust in large language models and prior use of large language models. (B) Interaction between trust in large language models and receipt of AI or informatics training. (C) Interaction between prior use of large language models and perceived usefulness. (D) Interaction between prior use of large language models and privacy concerns. (E) Interaction between perceived usefulness and privacy concern. (F) Interaction between prior use of large language models and clarity on legal responsibility. LLM: large language models.
Figure 7. Marginal effects of key predictors on willingness to adopt large language models among patients or caregivers. (A) Interaction between trust in large language models and prior use of LLMs. (B) Interaction between trust in large language models and prior awareness of large language models. (C) Interaction between trust in large language models and low education level (primary school or below). (D) Interaction between prior use of large language models and perceived usefulness. (E) Interaction between prior use of large language models and privacy concern. (F) Interaction between privacy concern and education level (high school). LLM: large language models.

Figure 6 depicts the marginal effects and interaction patterns of key predictors on HCPs’ willingness to adopt LLMs for clinical and research purposes. Data were derived from a multicenter cross-sectional survey of 240 physicians and nurses conducted between August 2024 and March 2025 across 5 tertiary hospitals in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China. Multivariable logistic regression models were estimated with average marginal effects to visualize how trust, prior exposure, perceived usefulness, privacy concern, and legal clarity interact to influence adoption intent. Positive marginal effects (points to the right of zero) indicate increased probability of adoption. Across all interactions, trust in LLMs consistently amplified the effect of favorable experience or clear governance conditions, whereas privacy concern attenuated these relationships. These findings suggest that institutional training and regulatory clarity strengthen the trust–usefulness pathway underlying professionals’ readiness to integrate LLMs into practice.

Figure 7 illustrates the marginal effects and interaction patterns of major predictors influencing PCs’ willingness to adopt LLMs for health information seeking and decision support. Data were obtained from a multicenter cross-sectional survey of 480 adults conducted between August 2024 and March 2025 across 5 tertiary hospitals in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China. Average marginal effects were derived from multivariable logistic regression models to visualize how trust, previous exposure, perceived usefulness, privacy concern, and education level jointly shape adoption intent. Positive marginal effects represent increased probability of LLM adoption, whereas negative values indicate deterrent effects. Across models, trust in LLMs remained the strongest positive driver, amplifying the influence of prior awareness and perceived usefulness, while privacy concern and low educational attainment significantly weakened adoption intent. These results underscore that, among PCs, cognitive accessibility and trust interact closely with privacy literacy and digital readiness to determine willingness to engage with AI-enabled health information systems.

Qualitative Insights

Semistructured interviews (n=30) revealed 11 thematic categories providing depth to the survey findings (Figure 8). Representative quotations from participants (eg, “HCP#3, female, oncology day care unit nurse;” “PC#8, male, caregiver”) were included to illustrate key themes and ensure analytic transparency.

Figure 8. Thematic structure, key subthemes, and motivational patterns shaping willingness to adopt large language models across health care professionals and patients or caregivers. (A) Theme frequency distribution. (B) Subtheme node diagram. (C) Top 20 subthemes in health care professionals. (D) Top 20 subthemes in patients or caregivers. (E) Subtheme co-occurrence in health care professionals. (F) Subtheme co-occurrence in patients or caregivers. (G) Motivational pathways for large language model adoption (Sankey). HCP: health care professional; LLM: large language model; PC: patients or caregivers.

Figure 8 summarizes the qualitative thematic analysis describing how HCPs and PCs conceptualize the benefits, risks, and conditions influencing their willingness to adopt LLMs in health care. Data were drawn from semistructured interviews with 30 participants (15 HCPs and 15 PCs) conducted between August 2024 and March 2025 across 5 tertiary hospitals in Guangdong, Hunan, Jiangxi, and Guizhou Provinces, China. Transcripts were coded inductively following thematic and co-occurrence analysis to identify dominant categories, subthemes, and motivational linkages. Overall, qualitative findings indicate that both HCPs and PCs perceive LLM adoption as a trust-dependent, literacy-mediated process shaped by previous exposure, perceived utility, and contextual reassurance through institutional safeguards.

Health care professionals emphasized risks and responsibilities, focusing on concerns about misinformation, privacy, and institutional accountability. They also stressed the need for structured onboarding, training, and workflow integration.

Patients or caregivers prioritized accessibility and comprehensibility, particularly among participants with lower digital literacy. Themes such as plain-language explanations, conversational interaction, and support for understanding medical information emerged strongly.

Across both groups, trust surfaced repeatedly as the foundation for adoption, while digital readiness shaped whether users perceived LLMs as beneficial or burdensome. Thematic co-occurrence analyses highlighted that risk-related concerns were often tied to calls for oversight and institutional safeguards. Patients or caregivers’ adoption intent was more closely tied to communication benefits, while HCPs’ intent was more closely tied to efficiency and professional utility.


Summary of Main Findings

This multicenter mixed methods study investigated how trust, perceived usefulness, and sociotechnical readiness influence willingness to adopt LLMs among HCPs and PCs in China. Consistent with our study objectives, quantitative analyses and multivariate models demonstrated that trust in LLMs was the most significant predictor for both groups. Model performance remained robust across approaches, and interpretability analyses consistently identified trust as the most influential feature. Interaction analyses further revealed that trust, privacy concern, and digital literacy jointly shaped adoption trajectories.

Complementary qualitative findings echoed these patterns: HCPs emphasized professional accountability and workflow integration, whereas PCs valued comprehension, reassurance, and accessible communication. Together, the mixed methods evidence indicates that the adoption of generative AI in health care depends less on algorithmic performance and more on human-centered and ethical readiness.

Trust as a Central Determinant

Trust emerged as the cornerstone of adoption intentions for both clinicians and patients [42,43]. From a sociotechnical perspective, trust encompasses confidence in technical reliability, institutional endorsement, and perceived integrity of AI outputs. Among HCPs, trust was reinforced by prior experience, legal clarity, and professional training, while among PCs, it was closely linked to transparency, interpretability, and educational background [44,45].

These findings suggest that trust functions as a bridge between perceived usefulness and behavioral intention, shaped by previous exposure and regulatory assurance. Therefore, cultivating trust requires not only transparent algorithms but also institutional validation and explainability-by-design—principles essential for ensuring safe AI integration in clinical environments [46,47].

Perceived Usefulness and Role-Specific Value

Perceived usefulness was another important enabler of adoption, reflecting both functional and cognitive dimensions. For HCPs, usefulness aligned with workflow efficiency gains and decision support, while PCs associated usefulness with plain-language explanations and empathy in communication.

This role-specific distinction illustrates that functional usefulness (improved professional performance) and cognitive usefulness (enhanced understanding and reassurance) are complementary but context-dependent. Designers should therefore tailor AI interfaces to user goals—embedding documentation tools and audit trails for clinicians, while simplifying outputs and promoting conversational transparency for PCs [48-53].

Barriers: Privacy Concerns, Literacy, and Socioeconomic Inequality

Privacy concerns consistently acted as a barrier to adoption, reflecting apprehensions about data misuse and accountability. Clinicians expressed worries about data liability and auditability, while PCs feared data misuse or unintended disclosure. These findings align with global evidence that perceived privacy risk undermines trust formation and information sharing [54,55].

Furthermore, education and digital tools use moderated adoption, highlighting how the digital divide extends beyond access to include literacy and contextual capability. Participants with lower education or from less developed provinces demonstrated markedly reduced adoption intent. This supports a sociotechnical interpretation of digital inequality—as a function of literacy, institutional support, and participatory capacity rather than infrastructure alone [34,54,55].

To mitigate these barriers, our findings suggest several actionable design and policy interventions. First, privacy-by-design features—such as decentralized data processing, built-in anonymization, and user-controllable consent dashboards—should be integrated into LLM-based systems to minimize exposure risks while preserving functionality. Second, transparent data-use and auditability frameworks are critical for both HCPs and PCs; clear communication of how data are stored, used, and protected can strengthen perceived institutional credibility and accountability.

Third, targeted privacy literacy and digital readiness education programs can empower users to understand their data rights and engage more confidently with AI systems. Finally, context-aware system design, including adaptive privacy settings and local-language interfaces, can bridge literacy and regional disparities.

Together, these strategies operationalize privacy protection as a foundation for trust, transforming user apprehension into informed confidence and facilitating equitable LLM adoption across diverse health care settings.

Model Performance and Interpretability

Comparative analyses demonstrated that all predictive models—LR, RF, and XGBoost—performed consistently well, confirming the robustness of observed associations. RF achieved the best calibration and interpretability balance, and ensemble approaches offered marginal accuracy gains. Overall, these results reinforce that the identified determinants, especially trust, usefulness, and privacy, are statistically and behaviorally meaningful.

Qualitative Integration and Thematic Insights

The qualitative component contextualized quantitative predictors by illustrating how users experience them in practice. Health care professionals’ motivation was efficiency- and accountability-driven; PCs’ motivation centered on clarity, empathy, and reduced uncertainty. These qualitative data contextualize the quantitative predictors by showing why these factors matter—trust is experienced as emotional reassurance and institutional safety rather than abstract confidence. This integration underscores the mixed methods value: statistical predictors explain “what matters,” while qualitative insights explain “why it matters.”

Interpretation of Variability Across Measures

Variability in clinical experience and digital familiarity among participants was expected, given the multicenter design, and enhances generalizability. Differences in perceived usefulness reflected heterogeneity in exposure to digital tools and institutional environments. Such real-world diversity strengthens, rather than weakens, interpretive validity by highlighting subgroup differences that warrant attention in future implementation studies.

Theoretical Contributions

This study contributes to information management theory in several ways.

First, it extends adoption models such as Technology Acceptance Model and Unified Theory of Acceptance and Use of Technology by empirically demonstrating that trust and ethical risk outweigh demographic factors in shaping adoption intent for AI-driven information tools. By quantifying these relationships across both HCPs and PCs, this research offers a novel, data-driven framework for understanding sociotechnical readiness in digital health [56,57].

Second, it introduces sociotechnical readiness as a critical construct, capturing how digital literacy, institutional support, and privacy concerns jointly influence information behaviors. This theoretical construct bridges individual psychology and institutional systems, offering a holistic model for studying complex technology ecosystems in health care [58-60].

Third, the findings highlight the multidimensional nature of information use, where the same technology offers distinct value propositions depending on user roles and contexts—pointing to the need for adaptive system design and role-based evaluation. This supports a paradigm shift in information systems research: from performance-centric models toward equity- and trust-centered frameworks that emphasize human values and ethical alignment.

Collectively, these contributions support a broader agenda for sociotechnical information systems research, emphasizing equity, governance, and the coconstruction of trust—reinforcing that effective information tools require integration of social factors, institutional policies, and technological affordances rather than standalone technical performance.

Practical and Policy Implications

From a practical standpoint, the findings of this study point to several priorities for the responsible deployment of LLMs as information systems.

First, trust must be intentionally engineered through transparent interfaces, verifiable information sources, and institutional endorsement mechanisms. Explainable and auditable AI systems can strengthen confidence among HCPs and PCs alike [46].

Second, system design should be role-sensitive—integrating workflow automation and accountability tools for HCPs while offering plain-language, empathetic interfaces for PCs. This aligns with information systems research emphasizing that system value is contingent on matching technologies to user roles and contexts [61].

Third, bridging the digital divide is imperative. Accessibility-by-design, multilingual support, and targeted digital literacy programs are essential for equitable participation. Evidence shows that the digital divide in AI use is not only about infrastructure, but also about literacy and sociotechnical support systems, requiring intentional interventions to foster equitable participation [35].

Finally, governance frameworks must evolve to clarify liability, privacy protection, and data sovereignty, ensuring that trust is institutionally maintained, not individually assumed. Collectively, these strategies translate theoretical findings into actionable design and policy recommendations, advancing the responsible and inclusive adoption of AI in health care [62,63].

Model Robustness and Potential Overfitting

Although the sample size for both HCPs and PCs was modest, multiple strategies were employed to minimize overfitting and enhance the generalizability of our models. Nested five-fold cross-validation and independent test evaluation demonstrated that model performance was stable across folds, with minimal variance between test and cross-validation AUCs. Regularization and feature selection procedures further reduced dimensionality and noise. The consistent performance of LR, RF, and XGBoost models—alongside similar AUC and Brier scores—supports the robustness of findings.

Model Selection and Interpretability Considerations

Although all 3 models—LR, RF, and XGBoost—demonstrated comparable discriminative performance and stable calibration, LR was identified as the optimal model for interpretation and practical deployment.

Its parsimony, transparency, and statistical stability make it particularly suitable for health care apps, where explainability and auditability are critical for clinical trust and ethical accountability.

Machine learning models such as RF and XGBoost offered marginal performance gains but introduced complexity without substantial improvement in predictive accuracy or calibration.

Therefore, LR was prioritized in interpretive analyses and policy recommendations as the most trustworthy and generalizable framework for understanding adoption behavior.

Limitations and Future Research

Several limitations should be noted. First, although trust consistently emerged as the strongest determinant, the causal mechanisms underlying trust formation require further longitudinal validation [62,64].

Second, the study focused on adoption intention rather than actual behavioral engagement, leaving room for future longitudinal and experimental studies [65].

Third, despite the use of nested cross-validation and calibration analyses to minimize overfitting, residual model instability cannot be entirely excluded—particularly among HCPs, where the sample size was relatively smaller and calibration plots showed wider error bars. The variability in secondary feature rankings across models indicates that predictors beyond trust should be interpreted with caution. Future research with larger, multicenter samples and external validation using independent datasets will be essential to confirm model robustness and generalizability.

Fourth, our analysis was conducted within China; cross-cultural replication is necessary to test the generalizability of sociotechnical readiness frameworks across differing regulatory and cultural contexts.

Finally, broader ethical issues—such as algorithmic bias, consent transparency, and localized governance—should be systematically integrated into future research agendas to ensure AI systems advance not only efficiency but also justice and inclusion.

Conclusion

Overall, this study advances the understanding of how trust, literacy, and institutional readiness jointly shape LLM adoption in health care. It establishes empirical evidence that trust functions as the primary mechanism linking technical credibility with social legitimacy. By conceptualizing sociotechnical readiness as the foundation for equitable AI adoption, this research offers both theoretical enrichment and practical direction. Theoretically, it advances existing technology adoption models by embedding ethical trust, literacy, and governance as core constructs of readiness, thereby bridging behavioral and institutional dimensions of digital transformation. Practically, it informs the design and policy development of AI implementation in health care—emphasizing trust-building mechanisms, literacy training, and transparent governance as prerequisites for sustainable integration. This study advances current understanding by shifting the focus from algorithmic performance toward the social, ethical, and institutional factors—particularly trust, literacy, and organizational readiness—that determine equitable AI adoption.

Acknowledgments

We thank all participating hospitals, health care professionals, and patients and caregivers for their valuable contributions to this multicenter study. The authors also appreciate the coordination and data management support provided by research teams across Guangdong, Hunan, Jiangxi, and Guizhou provinces, China. We particularly acknowledge the valuable assistance of Dr. Cheng Yan, public health researcher, and Dr. Chunqing Liu, clinical psychologist, who provided methodological support and independent evaluation of the qualitative data. Their review of interview transcripts and feedback on thematic coding enhanced the analytic rigor and reliability of the findings. The authors used the generative artificial intelligence tool GPT-5 (OpenAI) for language refinement, grammar correction, and consistency editing of the English manuscript. The tool was not used for conceptualization, data analysis, interpretation, or reference generation. All scientific content, results, and conclusions were developed, verified, and approved by the authors. Original GPT-5 interaction transcripts have been retained by the authors and are available upon request in accordance with JMIR Publications’ editorial policy on generative artificial intelligence use (Multimedia Appendix 12).

Funding

This study was supported by the Talent Fund of Guizhou Provincial People’s Hospital (grant number: [2024-‐16]) and the Basic Research Program of the Guizhou Provincial Department of Science and Technology (Qian Ke He Ji Chu) [Grant No. QKH-JC-MS (2025)492]. The funders had no involvement in the study design, data collection, data analysis, interpretation of results, or the writing of the manuscript.

Data Availability

The data sets generated and analyzed during this study are not publicly available due to institutional data protection policies and participant confidentiality regulations. Deidentified summary data and analysis code (Python and R scripts) used for statistical modeling and figure generation are available in the supplementary materials. Additional data may be made available from the corresponding author upon reasonable request by qualified researchers for academic purposes.

Authors' Contributions

XY: Conceptualization, methodology, software, validation, formal analysis, investigation, resources, data curation, writing - original draft.

YX: Validation, formal analysis, investigation, resources, data curation, writing - original draft.

DL: Validation, formal analysis, investigation, resources, data curation, writing - original draft.

HD: Validation, formal analysis, investigation, resources, data curation, writing - original draft.

JH: Validation, formal analysis, investigation, resources, data curation, writing - original draft.

YZ: Validation, formal analysis, investigation, resources, data curation, writing - original draft.

ML: Validation, formal analysis, investigation, resources, data curation, writing - original draft.

LD: Validation, formal analysis, investigation, resources, data curation, writing - original draft.

ZY: Data curation, writing - original draft.

BZ: Data curation, writing - original draft.

JY: Data curation, writing - original draft.

XZ: Data curation, writing - original draft.

WG: Data curation, writing - original draft.

CX: Conceptualization, methodology, writing - review and editing, visualization supervision, project administration, funding acquisition.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Good Reporting of A Mixed Methods Study (GRAMMS) reporting summary.

PDF File, 146 KB

Multimedia Appendix 2

Health care professionals’ and patients’ perceptions of large language models, multivariate correlates of large language model adoption, model performance and classification metrics, net reclassification improvement and integrated discrimination improvement comparisons, and calibration metrics across models.

DOCX File, 32 KB

Multimedia Appendix 3

Overview of statistical modeling and analytical procedures.

PDF File, 106 KB

Multimedia Appendix 4

Group comparisons of trust, perceived usefulness, and privacy concerns related to large language models.

PDF File, 17 KB

Multimedia Appendix 5

Interaction effects of key predictors on willingness to adopt large language models among health care professionals. (A) Trust in large language models × prior use of large language models. (B) Trust in large language models × artificial intelligence or informatics training. (C) Prior use of large language models × perceived usefulness. (D) Prior use of large language models × privacy concern. (E) Perceived usefulness × privacy concern. (F) Prior use of large language models × clarity on legal responsibility.

PDF File, 601 KB

Multimedia Appendix 6

Comparative clinical utility of large language models’ adoption prediction models based on cumulative incremental accuracy and net reclassification improvement. (A and B) Cumulative incremental accuracy curves across decision thresholds in health care professionals and patients or caregivers, demonstrating relative model performance in identifying true positives. (C and D) Net reclassification improvement analyses for health care professionals and patients or caregivers, highlighting the added value of the random forest model compared to logistic regression and extreme gradient boosting.

PDF File, 564 KB

Multimedia Appendix 7

Calibration assessment of large language models’ adoption prediction models across user groups using Hosmer-Lemeshow plots. (A and B) Hosmer-Lemeshow plots for health care professionals across logistic regression, random forest, and extreme gradient boosting models. (C and D) Hosmer-Lemeshow plots for patients or caregivers across logistic regression, random forest, and extreme gradient boosting models, showing calibration performance and deviations from perfect prediction.

PDF File, 665 KB

Multimedia Appendix 8

Interaction effects of key predictors on willingness to adopt large language models among patients or caregivers. (A) Trust in large language models × prior use of large language models. (B) Trust in large language models × prior awareness of large language models. (C) Trust in large language models × low education level (Primary school or below). (D) Prior use of large language models × perceived usefulness. (E) Prior use of large language models × privacy concern. (F) Privacy concern × education level (high school).

PDF File, 579 KB

Multimedia Appendix 9

Sensitivity analysis of predictor perturbations in health care professionals. (A-C) Probability distribution shifts in logistic regression, random forest, and extreme gradient boosting models when trust in large language models is perturbed (D-F). Heatmaps of model performance changes (ΔProba, ΔAUC, ΔLogLoss) across top predictors.

PDF File, 1389 KB

Multimedia Appendix 10

Sensitivity analysis of predictor perturbations in patients or caregivers. (A-C) Probability distribution shifts in logistic regression, random forest, and extreme gradient boosting models when trust in large language models is perturbed. (D-F) Heatmaps of model performance changes (ΔProba, ΔAUC, ΔLogLoss) across top predictors.

PDF File, 1188 KB

Multimedia Appendix 11

Decision curve analysis comparing clinical utility of predictive models for large language models adoption across user groups. (A) Net benefit curves for logistic regression, random forest, and extreme gradient boosting in health care professionals. (B) Net benefit curves for the same models in patients or caregivers.

PDF File, 138 KB

Multimedia Appendix 12

GPT-5 Interaction summary.

PDF File, 194 KB

  1. Thirunavukarasu AJ, Ting DSJ, Elangovan K, Gutierrez L, Tan TF, Ting DSW. Large language models in medicine. Nat Med. Aug 2023;29(8):1930-1940. [CrossRef] [Medline]
  2. Singhal K, Azizi S, Tu T, et al. Large language models encode clinical knowledge. Nature New Biol. Aug 2023;620(7972):172-180. [CrossRef] [Medline]
  3. Peng Y, Malin BA, Rousseau JF, et al. From GPT to DeepSeek: significant gaps remain in realizing AI in healthcare. J Biomed Inform. Mar 2025;163(104791):104791. [CrossRef] [Medline]
  4. Yang X, Zhang Y, Jiang J, et al. Harnessing GPT-4 for automated error detection in pathology reports: implications for oncology diagnostics. Digit Health. 2025;11:20552076251346703. [CrossRef] [Medline]
  5. Yang X, Xiao Y, Liu D, et al. Enhancing doctor-patient communication using large language models for pathology report interpretation. BMC Med Inform Decis Mak. Jan 23, 2025;25(1):36. [CrossRef] [Medline]
  6. Steimetz E, Minkowitz J, Gabutan EC, et al. Use of artificial intelligence chatbots in interpretation of pathology reports. JAMA Netw Open. May 1, 2024;7(5):e2412767. [CrossRef] [Medline]
  7. Menezes MCS, Hoffmann AF, Tan ALM, et al. The potential of generative pre-trained transformer 4 (GPT-4) to analyse medical notes in three different languages: a retrospective model-evaluation study. Lancet Digit Health. Jan 2025;7(1):e35-e43. [CrossRef] [Medline]
  8. Yang X, Chu XP, Huang S, et al. A novel image deep learning–based sub-centimeter pulmonary nodule management algorithm to expedite resection of the malignant and avoid over-diagnosis of the benign. Eur Radiol. 2024;34(3):2048-2061. [CrossRef]
  9. Yang X, Xiao Y, Liu D, et al. Enhancing physician-patient communication in oncology using GPT-4 through simplified radiology reports: multicenter quantitative study. J Med Internet Res. Apr 17, 2025;27:e63786. [CrossRef] [Medline]
  10. Yang X, Xiao Y, Liu D, et al. Cross language transformation of free text into structured lobectomy surgical records from a multi center study. Sci Rep. May 2, 2025;15(1):15417. [CrossRef] [Medline]
  11. Palaniappan K, Lin EYT, Vogel S. Global regulatory frameworks for the use of artificial intelligence (AI) in the healthcare services sector. Healthcare (Basel). Feb 28, 2024;12(5):562. [CrossRef] [Medline]
  12. Mathew JG, Monti F. Applications of large language models in information systems. In: De Luzi F, Monti F, Mecella M, editors. Engineering Information Systems with Large Language Models. Springer Nature; 2025:207-231. [CrossRef]
  13. Pan S, Luo L, Wang Y, Chen C, Wang J, Wu X. Unifying large language models and knowledge graphs: a roadmap. IEEE Trans Knowl Data Eng. 2024;36(7):3580-3599. [CrossRef]
  14. Wang S, Zhu Y, Liu H, Zheng Z, Chen C, Li J. Knowledge editing for large language models: a survey. ACM Comput Surv. Mar 31, 2025;57(3):1-37. [CrossRef]
  15. Ampel B, Yang CH, Hu J, Chen H. Large language models for conducting advanced text analytics information systems research. ACM Trans Manage Inf Syst. Mar 31, 2025;16(1):1-27. [CrossRef]
  16. Wagman KB, Dearing MT, Chetty M. Generative AI uses and risks for knowledge workers in a science organization. Presented at: Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems; Apr 26, 2025:1-17; Yokohama Japan. [CrossRef]
  17. Chen Z, Chan J. Large language model in creative work: the role of collaboration modality and user expertise. Manage Sci. Dec 2024;70(12):9101-9117. [CrossRef]
  18. Ning Y, Liu M, Liu N. ⁠Advancing ethical AI in healthcare through interpretability. Patterns (N Y). Jun 13, 2025;6(6):101290. [CrossRef] [Medline]
  19. Fehr J, Citro B, Malpani R, Lippert C, Madai VI. A trustworthy AI reality-check: the lack of transparency of artificial intelligence products in healthcare. Front Digit Health. 2024;6(1267290):1267290. [CrossRef] [Medline]
  20. Goktas P. Ethics, transparency, and explainability in generative AI decision-making systems: a comprehensive bibliometric study. J Decis Syst. 2024:1-29. [CrossRef]
  21. Goisauf M, Cano Abadía M, Akyüz K, et al. Trust, trustworthiness, and the future of medical AI: outcomes of an interdisciplinary expert workshop. J Med Internet Res. Jun 2, 2025;27:e71236. [CrossRef] [Medline]
  22. Mohsin Khan M, Shah N, Shaikh N, Thabet A, Alrabayah T, Belkhair S. Towards secure and trusted AI in healthcare: a systematic review of emerging innovations and ethical challenges. Int J Med Inform. Mar 2025;195(105780):105780. [CrossRef] [Medline]
  23. You Z, Wang Y, Xiao Y. Analysing the suitability of artificial intelligence in healthcare and the role of AI governance. Health Care Anal. Mar 6, 2025. [CrossRef] [Medline]
  24. Vo V, Chen G, Aquino YSJ, Carter SM, Do QN, Woode ME. Multi-stakeholder preferences for the use of artificial intelligence in healthcare: a systematic review and thematic analysis. Soc Sci Med. Dec 2023;338(116357):116357. [CrossRef] [Medline]
  25. Moldt JA, Festl-Wietek T, Fuhl W, et al. Exploring the social dimensions of AI integration in healthcare: a qualitative study of stakeholder views on challenges and opportunities. BMJ Open. Jun 27, 2025;15(6):e096208. [CrossRef] [Medline]
  26. Starke G, Gille F, Termine A, et al. Finding consensus on trust in AI in health care: recommendations from a panel of international experts. J Med Internet Res. Feb 19, 2025;27:e56306. [CrossRef] [Medline]
  27. Liu Y, Liu C, Zheng J, Xu C, Wang D. Improving explainability and integrability of medical AI to promote health care professional acceptance and use: mixed systematic review. J Med Internet Res. 2025;27:e73374. [CrossRef]
  28. Huang X, Ruan W, Huang W, et al. A survey of safety and trustworthiness of large language models through the lens of verification and validation. Artif Intell Rev. 2024;57(7):175. [CrossRef]
  29. Ciecierski-Holmes T, Singh R, Axt M, Brenner S, Barteit S. Artificial intelligence for strengthening healthcare systems in low- and middle-income countries: a systematic scoping review. NPJ Digit Med. Oct 28, 2022;5(1):162. [CrossRef] [Medline]
  30. Kim JY, Hasan A, Kueper J, et al. Establishing organizational AI governance in healthcare: a case study in Canada. NPJ Digit Med. Aug 15, 2025;8(1):522. [CrossRef] [Medline]
  31. Ramadan OME, Alruwaili MM, Alruwaili AN, Elsehrawy MG, Alanazi S. Facilitators and barriers to AI adoption in nursing practice: a qualitative study of registered nurses’ perspectives. BMC Nurs. Dec 18, 2024;23(1):891. [CrossRef] [Medline]
  32. Hendricks-Sturrup R, Simmons M, Anders S, et al. Developing ethics and equity principles, terms, and engagement tools to advance health equity and researcher diversity in AI and machine learning: modified delphi approach. JMIR AI. Dec 6, 2023;2:e52888. [CrossRef] [Medline]
  33. Ball Dunlap PA, Michalowski M. Advancing AI data ethics in nursing: future directions for nursing practice, research, and education. JMIR Nurs. Oct 25, 2024;7:e62678. [CrossRef] [Medline]
  34. Biga R, Nottebaum S, Kozlakidis Z, Psomiadis S. Digitalization of healthcare in lmics: digital health and the digital divide based on technological availability and development. In: Kozlakidis Z, Muradyan A, Sargsyan K, editors. Digitalization of Medicine in Low- and Middle-Income Countries. Springer International Publishing; 2024:185-193.
  35. Osonuga A, Osonuga AA, Fidelis SC, Osonuga GC, Juckes J, Olawade DB. Bridging the digital divide: artificial intelligence as a catalyst for health equity in primary care settings. Int J Med Inform. Dec 2025;204(106051):106051. [CrossRef] [Medline]
  36. Wang D, Wang L, Zhang Z. “Brilliant AI doctor” in rural clinics: challenges in AI-powered clinical decision support system deployment. 2021. Presented at: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems; May 6, 2021:1-18; Yokohama, Japan. [CrossRef]
  37. Zhang Y, Bilawal Khaskheli M. The role of digital technologies in advancing sustainable economic development into intersections of policy, law, environmental economics, and a comparative study of China, the EU, and the USA. Sustainability. 2025;17(19):8666. [CrossRef]
  38. Bandeira A, Gonçalves LH, Holl F, et al. Viewpoint on the intersection among health information, misinformation, and generative AI technologies. JMIR Infodemiology. Sep 15, 2025;5:e69474. [CrossRef] [Medline]
  39. Jiang N, Yu X, Yang Y, et al. Digital therapeutics in China: comprehensive review. J Med Internet Res. May 27, 2025;27:e70955. [CrossRef] [Medline]
  40. Zhi Z, Zhao J, Li Q, et al. Evolving perceptions and attitudes to adopting generative AI in professional settings: multicenter longitudinal qualitative study of senior Chinese hospital leaders. J Med Internet Res. Jun 27, 2025;27:e75531. [CrossRef] [Medline]
  41. O’Cathain A, Murphy E, Nicholl J. The quality of mixed methods studies in health services research. J Health Serv Res Policy. Apr 2008;13(2):92-98. [CrossRef] [Medline]
  42. Asif M, Sarwar F. The essential position of trust in the technology acceptance model: a comprehensive review and bibliometric analysis in management studies. Journal of Modelling in Management. Oct 1, 2025;20(5):1635-1655. [CrossRef]
  43. Choudhury A, Shamszare H. Investigating the impact of user trust on the adoption and use of ChatGPT: survey analysis. J Med Internet Res. Jun 14, 2023;25:e47184. [CrossRef] [Medline]
  44. Söllner M, Hoffmann A, Leimeister JM. Why different trust relationships matter for information systems users. Eur J Inf Syst. May 2016;25(3):274-287. [CrossRef]
  45. Wu K, Zhao Y, Zhu Q, Tan X, Zheng H. A meta-analysis of the impact of trust on technology acceptance model: Investigation of moderating influence of subject and context type. Int J Inf Manage. Dec 2011;31(6):572-581. [CrossRef]
  46. Bach TA, Khan A, Hallock H, Beltrão G, Sousa S. A systematic literature review of user trust in AI-enabled systems: an HCI perspective. International Journal of Human–Computer Interaction. Mar 3, 2024;40(5):1251-1266. [CrossRef]
  47. Gulati S, McDonagh J, Sousa S, Lamas D. Trust models and theories in human–computer interaction: a systematic literature review. Computers in Human Behavior Reports. Dec 2024;16:100495. [CrossRef]
  48. Hafiza Razami H, Ibrahim R. Models and constructs to predict students’ digital educational games acceptance: a systematic literature review. Telematics and Informatics. Sep 2022;73:101874. [CrossRef]
  49. Aljarrah E, Elrehail H, Aababneh B. E-voting in Jordan: assessing readiness and developing a system. Comput Human Behav. Oct 2016;63:860-867. [CrossRef]
  50. Davis FD. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Q. Sep 1, 1989;13(3):319-340. [CrossRef]
  51. Ambalov IA. Decomposition of perceived usefulness: a theoretical perspective and empirical test. Technol Soc. Feb 2021;64:101520. [CrossRef]
  52. Pal A, Wangmo T, Bharadia T, et al. Generative AI/LLMs for plain language medical information for patients, caregivers and general public: opportunities, risks and ethics. Patient Prefer Adherence. 2025;19:2227-2249. [CrossRef] [Medline]
  53. Warde F, Papadakos J, Papadakos T, Rodin D, Salhia M, Giuliani M. Plain language communication as a priority competency for medical professionals in a globalized world. Can Med Educ J. May 2018;9(2):e52-e59. [Medline]
  54. Vassilakopoulou P, Hustad E. Bridging digital divides: a literature review and research agenda for information systems research. Inf Syst Front. 2023;25(3):955-969. [CrossRef] [Medline]
  55. Lumare N, Muradyan L, Jansberg C. Behind the screen: the relationship between privacy concerns and social media usage. Journal of Marketing Communications. 2024:1-16. [CrossRef]
  56. Lai CY, Cheung KY, Chan CS, Law KK. Integrating the adapted UTAUT model with moral obligation, trust and perceived risk to predict ChatGPT adoption for assessment support: a survey with students. Comput Educ Artif Intell. Jun 2024;6:100246. [CrossRef]
  57. Balaskas S, Tsiantos V, Chatzifotiou S, Rigou M. Determinants of ChatGPT adoption intention in higher education: expanding on TAM with the mediating roles of trust and risk. Information. 2025;16(2):82. [CrossRef]
  58. Chan J, Peko G, Sundaram D, Sadovykh V. Application and adoption of emerging technologies: from AI to metaverse. Inf Syst Front. Apr 2025;27(2):405-408. [CrossRef]
  59. Verma A, Allen T. A sociotechnical readiness level framework for the development of advanced nuclear technologies. Nucl Technol. Sep 2024;210(9):1722-1739. [CrossRef]
  60. Yigzaw S, Jormanainen I, Tukiainen M. Information systems in developing countries: opportunities and challenges. Presented at: Ninth International Conference on Technological Ecosystems for Enhancing Multiculturality (TEEM’21); Oct 26, 2021:812-818; Barcelona Spain. [CrossRef]
  61. Glassberg I, Ilan YB, Zwilling M. The key role of design and transparency in enhancing trust in AI-powered digital agents. Journal of Innovation & Knowledge. Sep 2025;10(5):100770. [CrossRef]
  62. Emery-Xu N, Jordan R, Trager R. International governance of advancing artificial intelligence. AI & Soc. Apr 2025;40(4):3019-3044. [CrossRef]
  63. Bozkurt Y, Rossmann A, Pervez Z, Ramzan N. Assessing data governance models for smart cities: benchmarking data governance models on the basis of European urban requirements. Sustainable Cities and Society. Jul 2025;130:106528. [CrossRef]
  64. Shaw H, Ellis DA, Ziegler FV. The technology integration model (TIM). predicting the continued use of technology. Comput Human Behav. Jun 2018;83:204-214. [CrossRef]
  65. Jeyaraj A, Dwivedi YK, Venkatesh V. Intention in information systems adoption and use: current state and research directions. Int J Inf Manage. Dec 2023;73:102680. [CrossRef]


AI: artificial intelligence
AUC: area under the receiver operating characteristic curve
GRAMMS: Good Reporting of A Mixed Methods Study
LLM: large language model
LR: logistic regression
OR: odds ratio
RF: random forest
SHAP: Shapley additive explanations
XGBoost: extreme gradient boosting


Edited by Stefano Brini; submitted 27.Sep.2025; peer-reviewed by Asef Raiyan Hoque, Jinyu Guo; final revised version received 15.Oct.2025; accepted 15.Oct.2025; published 11.Dec.2025.

Copyright

© Xiongwen Yang, Yi Xiao, Di Liu, Huiyin Deng, Jian Huang, Yubin Zhou, Maoli Liang, Longyan Dong, Zihao Yuan, Jing Yao, Wankai Guo, Chuan Xu. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 11.Dec.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.