Published on in Vol 27 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/66083, first published .
Impact of AI-Assisted Diagnosis on American Patients’ Trust in and Intention to Seek Help From Health Care Professionals: Randomized, Web-Based Survey Experiment

Impact of AI-Assisted Diagnosis on American Patients’ Trust in and Intention to Seek Help From Health Care Professionals: Randomized, Web-Based Survey Experiment

Impact of AI-Assisted Diagnosis on American Patients’ Trust in and Intention to Seek Help From Health Care Professionals: Randomized, Web-Based Survey Experiment

Authors of this article:

Catherine Chen1, 2 Author Orcid Image ;   Zhihan Cui3, 4 Author Orcid Image

1Manship School of Mass Communication, Louisiana State University, Baton Rouge, LA, United States

2Department of Political Science, Louisiana State University, Baton Rouge, LA, United States

3Institute of Global Health and Development, Peking University, 5 Yiheyuan Road, Haidian District, Beijing, China

4Anderson School of Management, University of California, Los Angeles, Los Angeles, CA, United States

*all authors contributed equally

Corresponding Author:

Zhihan Cui, PhD


Background: Artificial intelligence (AI) technologies are increasingly integrated into medical practice, with AI-assisted diagnosis showing promise. However, patient acceptance of AI-assisted diagnosis, compared with human-only procedures, remains understudied, especially in the wake of generative AI advancements such as ChatGPT.

Objective: This research examines patient preferences for doctors using AI assistance versus those relying solely on human expertise. It also studies demographic, social, and experiential factors influencing these preferences.

Methods: We conducted a preregistered 4-group randomized survey experiment among a national sample representative of the US population on several demographic benchmarks (n=1762). Participants viewed identical doctor profiles, with varying AI usage descriptions: no AI mention (control, n=421), explicit nonuse (No AI, n=435), moderate use (Moderate AI, n=481), and extensive use (Extensive AI, n=425). Respondents reported their tendency to seek help, trust in the doctor as a person and a professional, knowledge of AI, frequency of using AI in their daily lives, demographics, and partisan identification. We analyzed the results with ordinary least squares regression (controlling for sociodemographic factors), mediation analysis, and moderation analysis. We also explored the moderating effect of past AI experiences on the tendency to seek help and trust in the doctor.

Results: Mentioning that the doctor uses AI to assist in diagnosis consistently decreased trust and intention to seek help. Trust and intention to seek help (measured with a 5-point Likert scale and coded as 0‐1 with equal intervals in between) were highest when AI was explicitly absent (control group: mean 0.50; No AI group: mean 0.63) and lowest when AI was extensively used (Extensive AI group: mean 0.30; Moderate AI group: mean 0.34). A linear regression controlling for demographics suggested that the negative effect of AI assistance was significant with a large effect size (β=−.45, 95% CI −0.49 to −0.40, t1740=−20.81; P<.001). This pattern was consistent for trust in the doctor as a person (β=−.33, 95% CI −0.37 to −0.28, t1733=−14.41; P<.001) and as a professional (β=−.40, 95% CI −0.45 to −0.36, t1735=−18.54; P<.001). Results were consistent across age, gender, education, and partisanship, indicating a broad aversion to AI-assisted diagnosis. Moderation analyses suggested that the “AI trust gap” shrank as AI use frequency increased (interaction term: β=.09, 95% CI 0.04-0.13, t1735=4.06; P<.001) but expanded as self-reported knowledge increased (interaction term: β=−.04, 95% CI −0.08 to 0.00, t1736=−1.75; P=.08).

Conclusions: Despite AI’s growing role in medicine, patients still prefer human-only expertise, regardless of partisanship and demographics, underscoring the need for strategies to build trust in AI technologies in health care.

Trial Registration: OSF Registries osf.io/5vcdg; https://osf.io/5vcdg

J Med Internet Res 2025;27:e66083

doi:10.2196/66083

Keywords



In the United States, the Food and Drug Administration has approved artificial intelligence (AI) use in various health care scenarios [1]. AI has facilitated medical processes by reducing decision-making errors, lowering costs, and speeding up diagnoses [2,3]. Despite these benefits, AI diagnosis faces psychological and structural barriers to patient acceptance. These include beliefs that AI cannot consider individual-specific factors [4], uncertainty about liability [5,6], and concerns about data privacy [2,6]. As a result, many consumers remain reluctant to accept fully automated diagnosis [2,4,7].

These challenges have led to a more moderate approach, AI-assisted diagnosis, where AI is a supportive tool for human doctors rather than a replacement [2,7]. Studies show that patients prefer this collaborative approach over fully automated systems [2,7,8]. For instance, Jutzi et al [9] found that 94% of patients would accept AI assistance in skin cancer screening, but this number shrank to 41% when AI stands alone. Palmisciano et al [10] revealed that the acceptance of fully automated neurosurgery was only 17.7% compared with 47.7% for AI-assisted neurosurgery. This is possibly because AI-assisted diagnosis combines the credibility of human expertise with the efficiency and cost-effectiveness of AI [2,11].

However, most prior research focused on earlier forms of AI, such as image readers [7,12], high-risk patient identifiers [13,14], and virtual assistant chatbots [15,16], and was conducted before the emergence of large language model–based generative AI. The emergence of generative AI has introduced a new dimension to medical AI [17]. Generative AI models excel in conducting clinical reasoning and helping medical education [18,19], diagnosing complex cases [3], and enabling text-to-protein generation [20]. These applications have marked a considerable paradigm change in medical AI because of their controllability, adaptability, and applicability [21].

Nonetheless, several concerns limit the public’s acceptance of generative models in health care. First, generative AI models can produce hallucinations, instances where they generate information that sounds plausible but is factually incorrect or entirely fabricated [22]. This is especially concerning because false diagnoses, incorrect treatment recommendations, or fabricated medical facts could seriously affect patients’ health and well-being. Moreover, because commercial firms develop most generative AI products, their lack of transparency and ethical risks to privacy and data protection are potential challenges [3]. Finally, as generative AI is still a relatively novel technology, the general public may have limited familiarity with it and may not yet associate it with medical applications, which could pose a barrier to its acceptability [23].

Because of these issues, whether patients accept generative AI as a tool to support diagnosis remains an open question that requires empirical exploration. To fill this gap, this research is interested in the extent to which people would prefer doctors using AI assistance to traditional doctors relying solely on human expertise. This research question is essential because understanding preferences for AI-assisted doctors can inform health care policy [24], guide the development of AI technologies in health care, and improve patient health outcomes [6] and satisfaction [25].

Previous research has applied the technology acceptance model (TAM) [26-28] and risk perception theory [29] to study the acceptance of AI-assisted diagnosis. Both theories suggest a cost-benefit analysis of perceived gains and hazards from the signals observed about AI-assisted diagnosis as a technology. These theories have implied several factors associated with the acceptance of AI-assisted diagnosis. First, trust is a crucial psychological determinant for AI acceptance, as users must believe the potential benefits and overcome initial skepticism or fear of harmful risks to begin engaging with and ultimately accepting new technological innovations such as AI [23,28,30-32]. For instance, a recent study by Choung et al [28] directly applies the TAM theory on AI acceptance and teases out trust as a robustly significant factor. Juravle et al [31] and Shevtsova et al [32] both demonstrated the importance of trust in the acceptance of AI-assisted medical care. The role of trust is consistent across many studies, and its determinants have been revealed to include socioethical considerations, technical and design features, and user characteristics [33,34].

The TAM also suggests that the perceived ease of use is a key factor for consumers to accept a new technology [26,35]. Familiarity, as predicted by intuition, breeds perceived ease of use [36,37]. Also, it reduces the concern of unexpected risks or harms, aligning with the risk perception theory. Familiarity with AI or new technology also matters because it may breed more comfort and trust in the technology adoption [23]. For instance, continued experience with AI-based algorithms may help develop trust over time [38]. These 2 pathways both determine the importance of technological familiarity.

Beyond the insights from TAM and risk perception theory, communication strategies also shape how people perceive both the benefits and potential downsides of AI [29,39,40]. When AI is framed as an adaptive entity capable of learning from user interactions, a humanized agent endowed with relatable traits (eg, conversational tone and empathy cues), and a collaborative partner designed to augment—rather than replace—human agency [8,11], users are more likely to perceive it as trustworthy and predictable.

Last but not least, user attributes are crucial determinants in AI trust and acceptance [34]. For instance, past studies have suggested that Republicans (conservatives) are typically more skeptical of science and technology [41], signaling a potentially stronger aversion to AI-assisted health care. Preliminary evidence supports this, showing an exceptionally high aversion to AI in diagnosis among conservatives [42,43], with some past research linking conservative and risk aversion to AI’s uncertainty [42].

This study examines an AI-assisted diagnostic approach designed to work alongside existing health care practices rather than function as a stand-alone tool. Patients are presented with a doctor who may or may not rely on AI for evaluating and diagnosing moderate symptoms. Because the study focuses on people’s general responses to the idea of AI in medical settings, it does not target a specific patient population or condition. We hypothesized that Republicans distrust AI-assisted health care more. However, we did not have a directional hypothesis regarding the general public’s perception of AI-assisted diagnosis. Additionally, we explored the impact of various demographic characteristics on AI acceptance.


Experimental Design and Data Sources

From June 17 to 20, 2024, we recruited 1762 participants via Prolific, a web-based subject platform suitable for survey and experimental research. This platform has a high response quality and a low dishonesty rate [44,45]. The sample represented the US population based on gender, age, and political affiliation.

Participants were randomly assigned to 1 of 4 treatment arms. In all 4 groups, participants were presented with identical information about an illness and a typical doctor (“You’re having some moderate symptoms and are thinking about getting advice from a doctor. Here is Doctor M, a doctor with average experience and expertise in treating your illness.”). Figure 1 shows a flowchart illustrating the experimental design.

Figure 1. Flowchart of the experiment design. AI: artificial intelligence.

The control group did not mention AI-related content. In the “No AI” group, participants were informed that the doctor does not use AI assistance for diagnosis or treatment (“Doctor M doesn’t use artificial intelligence (AI, such as ChatGPT) at all when evaluating symptoms and deciding on treatments.”). These 2 groups are coded as “without AI use.” The “Moderate AI” group noted that the doctor uses AI, such as ChatGPT, to a moderate extent when evaluating symptoms and deciding on treatments. The information in the “High AI” group indicated that the doctor uses AI, such as ChatGPT, extensively when evaluating symptoms and deciding on treatments. These 2 groups are coded as “with AI use.”

Next, respondents reported their tendency to seek help, trust in the doctor as a person and a professional, knowledge of AI, and frequency of using AI in their daily lives. All these measures were measured from a 5-point Likert scale and coded into 0, 1/4, 1/2, 3/4, and 1. Then, we collected the participants’ demographics and partisan identification. The interface that a surveyee may face during the experiment is demonstrated in Multimedia Appendix 1. The left panel suggests a PC/Mac interface, and the right panel suggests a smartphone interface.

Analysis of the Main Effect

Following the preregistration (OSF ID: v8kzs) [46], the main results were obtained through ordinary least squares (OLS) regression. We controlled for gender, age, income, education, race, and Hispanic ethnicity. Preregistered OLS regressions were performed on self-identified Republicans. The main regression is specified as follows:

Yi=β×AI_Usei+Controlsi+ui(1)

where Yi is the outcome variable (trust as a professional or an individual, and the intention to seek help), and AI_Usei is a binary variable with 1 indicating that the doctor uses AI-assisted diagnosis.

When examining the effects of AI use intensity within the Republican sample (as preregistered), we conducted the following regression analysis among participants who were assigned to an AI-assisted diagnosis condition:

Yi=β×Extensivei+Controlsi+ui(2)

where Extensivei is a binary variable with 1 indicating that the doctor uses AI-assisted diagnosis extensively and 0 indicating that the doctor uses AI assistance moderately. It was hypothesized that Republicans are less likely to seek help from doctors who use AI extensively compared with those who use AI moderately. We report the standardized OLS regression coefficients as indicators of the effect size [47].

The same regression analyses (1) and (2) were also conducted on the full sample as part of an exploratory analysis. While we did not preregister a directional hypothesis, examining the sign and magnitude of the effect in the general population remains valuable. Additionally, we assess how AI use impacts individuals across different demographic backgrounds.

Moderation and Mediation Analysis

To test the hypothesis that Republican partisanship leads to a greater aversion to AI-assisted diagnosis, we conducted a moderation analysis within the full sample. This analysis examined whether partisanship moderates the effect of AI use on trust and the intention to seek help. Moderation was assessed by analyzing the coefficients of interaction terms [48]:

Yi=β1×AI_Usei+β2×Repi+β3×(AI_Usei×Repi)+Controlsi+ui

where Yi is the intention to seek help, trust as a person, or trust as a professional. We coded Repi in three ways: (1) a binary variable indicating self-reported Republican partisanship (Republican=1; Democrat and Independent=0), (2) a factor with 3 levels (Democrat, Republican, and Independent), and (3) a continuous variable of self-identified partisanship (“Strong Republican”=1, “Not Strong Republican”=5/6, “Lean toward the Republican Party”=2/3, “Don’t lean either way”=1/2, “Lean toward the Democratic Party”=1/3, “Not Strong Democrat”=1/6, and “Strong Democrat”=0).

Then, we conducted a preregistered mediation analysis to test the hypotheses that trust as a professional or an individual mediates the treatment effect of AI use on the intention to seek help. Mediation analyses were performed with the R package (R Foundation for Statistical Computing) mediation [49]. We also controlled for gender, age, income, education, Hispanic ethnicity, and race in the mediation analysis.

Finally, we conducted an exploratory moderation analysis to test the effect of AI familiarity (frequency of use and self-reported AI knowledge) in the full sample. The regression function is:

Yi=β1×AI_Usei+β2×Fami+β3×(AI_Usei×Fami)+Controlsi+ui

Data, Materials, and Software Availability

Preregistration, data, and related analysis (including code files) are available in the Open Science Framework repository (#v8kzs). The complete text of the treatments and survey questions is shown in Multimedia Appendix 2.

Ethical Considerations

This research was conducted under the approval of the University of California Los Angeles institutional review board (No. 24‐000594). Informed consent procedures were implemented in accordance with University of California Los Angeles institutional review board regulations and were administered through Prolific. All data were fully anonymized at the time of collection. No personally identifiable information was collected. Participant responses were stored and analyzed in deidentified format. Participants received compensation consistent with Prolific’s fair pay guidelines. Each participant was compensated at or above the equivalent of US $9.78 per hour, consistent with fair wage standards. No identifiable participant images or materials are included in this manuscript or multimedia appendices.


Summary Statistics

Summary statistics of main outcome variables and covariates are listed in Table 1. Balanced tests are shown in Table S1 in Multimedia Appendix 2. Demographic characteristics are well balanced across groups, with gender, age, and political affiliation distributions generally reflecting national averages. See Multimedia Appendix 2 for demographic comparisons between the current sample and national statistics.

Table 1. Summary statistics of main outcome variables and covariates.
VariablesSummary statistics (n=1762)
Outcome variables, 0‐1 continuous scale, mean (SD)
 Intention to visit the doctor0.44 (0.28)
 Trust of the doctor as a person0.47 (0.23)
 Trust of the doctor as a professional0.48 (0.26)
Group allocation, n (%)
 Does not mention AIa421 (23.9)
 Mention no AI use435 (24.7)
 Moderate AI use481 (27.3)
 Extensive AI use425 (24.1)
Demographics
 Female, n (%)916 (52.0)
 Age (years), n (%)
  18‐29312 (17.7)
  30‐39335 (19.0)
  40‐49278 (15.8)
  50‐59384 (21.8)
  60‐69337 (19.1)
  ≥70116 (6.6)
  Mean age (SD), years47.0 (15.7)
 Race, n (%)
  White1455 (82.6)
  Black180 (10.2)
  Asian129 (7.3)
  Native American32 (1.8)
  Hispanic141 (8.0)
Other34 (1.9)
 Annual income cohort, n (%)
  <20,000152 (8.6)
  20,000‐50,000422 (24.0)
  50,000‐75,000339 (19.2)
  75,000-100,000305 (17.3)
  >100,000544 (30.9)
 Education level, n (%)
  Less than high school14 (0.8)
  High school graduate189 (10.7)
  Some college/associate562 (31.9)
  College graduate657 (37.3)
  Postgraduate340 (19.3)
 Partisanship identity, n (%)
  Strong Democrat350 (19.9)
  Democrat208 (11.8)
  Independent leaning Democrat297 (16.9)
  Independent214 (12.2)
  Independent leaning Republican178 (10.1)
  Republican264 (15.0)
  Strong Republican251 (14.3)
  Continuous partisan score (SD)0.47 (0.35)
 Self-identity as Republican, n (%)515 (29.2)
 AI use experience, n (%)
  Not frequently at all578 (32.9)
  Slightly frequently555 (31.6)
  Moderately frequently330 (18.8)
  Very frequently211 (12.0)
  Extremely frequently85 (4.8)
  Continuous AI use score, mean (SD)0.31 (0.29)
 Self-reported AI knowledge level, n (%)
  None at all49 (2.8)
  A little476 (27.1)
  A moderate amount739 (42.0)
  A lot358 (20.3)
  A great deal138 (7.8)
  Continuous knowledge score, mean (SD)0.51 (0.24)

aAI: artificial intelligence.

Main Effect

Figure 2 shows the effect of doctor’s AI usage on patients’ intention to seek help in different partisan groups. Among Republicans, the average intention to seek help is 0.56 in groups where the doctor does not use AI, but it drops to 0.32 in groups where the doctor uses AI. Consistent with our hypothesis, mentioning a doctor’s use of AI reduced respondents’ trust in the doctor (as a person: β=−.31, 95% CI −0.39 to −0.22, t491=−7.20, P<.001, adjusted R2=0.11; as a professional: β=−.38, 95% CI −0.46 to −0.30, t491=−9.28, P<.001, adjusted R2=0.13) and decreased respondents’ intention to seek help (β=−.43, 95% CI −0.51 to −0.35, t494=−10.58, P<.001, adjusted R2=0.19).

We also preregistered that Republicans would be less likely to seek help from doctors who use AI extensively than from those who use AI moderately. However, as demonstrated in Figure 3 (Moderate Use group: mean 0.34; Extensive Use group: mean 0.29), the intention to seek help did not differ significantly among Republicans, regardless of whether the doctor used AI extensively or moderately (β=.20, 95% CI −0.04 to 0.45, t255=1.64, P=.102, adjusted R2=0.006). Similarly, Republicans’ trust in the doctor, both as a person and as a professional, did not differ significantly based on the extent of AI use (as a person: β=.19, 95% CI −0.05 to 0.42, t254=1.56, P=.12, adjusted R2=0.006; as a professional: β=.23, 95% CI −0.01 to 0.47, t254=1.88, P=.06, adjusted R2=0.032).

Next, we conducted exploratory analyses using the full sample. The 2 panels in Figure 4 show the effect of the doctor’s AI use on a patient’s visit intention in the full sample. Similarly to what we found among Republicans, across the entire sample, AI use reduced trust (as a person: β=−.33, 95% CI −0.37 to −0.28, t1733=−14.41, P<.001, adjusted R2=0.20; as a professional: β=−0.40, 95% CI −0.37 to −0.28, t1735=−18.54, P<.001, adjusted R2=0.11) and decreased respondents’ intention to seek help (β=−0.45, 95% CI −0.45 to −0.36, t1740=−20.81, P<.001, adjusted R2=0.17).

Interestingly, respondents’ trust in the doctor and their intention to seek help were highest when the prompt explicitly stated that the doctor does not use AI at all in their diagnosis (control group: mean 0.50; No AI group: mean 0.63; Extensive AI group: mean 0.31; and Moderate AI group: mean 0.34). Taken together, Figures 2-4 suggest that the aversion to AI assistance was consistent across party identifications.

Moreover, the 3 panels of Figure 5 further showed that the aversion to AI assistance was consistent across age, gender, and education level. The distrust was distributed evenly across the population.

As another robustness check, we tested 3-way interactions (the treatment condition interacted with 2 demographic and political variables) to examine whether the treatment effects varied across intersecting subgroup combinations. We ran 10 ANOVA tests on all 10 combinations of 5 demographic and political variables (age, gender, income, education level, and partisanship). None of the 3-way interactions was significant at the .05 level. This adds further evidence to our argument that the AI aversion is uniform.

Figure 2. The effect of doctor’s AI usage on patients’ intention to seek help in different partisan groups. Error bars represent SE. AI: artificial intelligence.
Figure 3. Republican patients’ intention to seek help in all 4 treatment groups. Error bars represent SE. AI: artificial intelligence.
Figure 4. Patients’ intention to seek help under different treatment conditions (full sample, by AI use status, and by treatment condition). Error bars represent SE. AI: artificial intelligence.
Figure 5. Intention to seek help among patients from different demographic backgrounds (full sample). Error bars represent SE. AI: artificial intelligence.

Moderation and Mediation

We preregistered a moderation analysis to test whether Republican partisanship negatively moderates the effect of AI use on the intention to seek help. However, Republican partisanship did not have a significant moderating effect. This is supported by the interaction regression with 3 different specifications of Republican partisanship. The coefficient of the interaction term is not significant regardless of how partisanship was coded (binary variable: β=.01, t1738=0.34, P=.73; 3-level factor: β=−.01, t1736=−0.18, P=.86; and continuous variable: β=.001, t1738=0.07, P=.94). The aversion to AI assistance was consistent across party identification.

We preregistered that the effect of AI usage on visiting intention is mediated by trust in the doctor. This was supported both among Republicans (mediation by the trust as a person: average causal mediation effect [ACME]=−0.11, P<.001, average direct effect [ADE]=−0.12, P<.001; mediation by the trust as a professional: ACME=−0.16, P<.001, ADE=−0.08, P<.001) and among the full sample (mediation by trust as a person: ACME=−0.11, P<.001, ADE=−0.13, P<.001; mediation by trust as a professional: ACME=−0.17, P<.001, ADE=−0.08, P<.001).

Surprisingly, despite the high correlation between AI knowledge and AI use frequency (r=0.61), these 2 variables had opposite moderating effects (Figures 6A and B). In the entire sample, the negative effect of AI assistance diminished as participants’ AI use frequency increased (interaction term β=.09, 95% CI 0.04 to 0.13, t1735=4.06; P<.001). In other words, the “trust gap” was smallest among those who used AI most frequently. Conversely, the “trust gap” expanded as participants’ self-reported AI knowledge increased, with interaction term β=−.04, 95% CI −0.08 to 0.00, t1736=−1.75; P=.08. Participants who reported the highest AI knowledge exhibited the largest trust gap.

Figure 6. Panel A: The impact of a doctor’s AI usage on a patient’s visit intention was moderated by the patient’s self-reported AI knowledge. Panel B: The impact of a doctor’s AI usage on a patient’s visit intention was moderated by the patient’s self-reported AI use frequency. AI: artificial intelligence.

Summary of Findings

This study investigated the impact of AI-assisted diagnosis on patients’ trust in doctors and their intention to seek medical help. Mentioning a doctor’s use of AI significantly decreased respondents’ trust in the doctor, both as a person and as a professional, and in turn, reduced their intention to seek help (partial mediation in both Republicans and the full sample).

We preregistered that Republicans would be more averse to AI-assisted diagnosis because previous literature suggested that conservatives are more likely to resist changes and new technology [41,43]. However, our results revealed that Americans’ distrust of AI-assisted diagnosis surpassed partisanship and demographic lines. The distrust was consistent across age, gender, education, and party identification. In other words, Americans’ distrust of AI-assisted diagnosis is general rather than a conservatism-specific phenomenon. Furthermore, trust in the doctor and the intention to seek help were highest when the doctor explicitly renounced AI in their diagnosis, suggesting that patients may still prefer human-only expertise in medical care.

Why Is the Aversion Significant and Widespread?

A major finding of this paper is the significant and widespread aversion toward AI-assisted diagnosis in the United States. This result deviates from previous research on specific AI health care products [2,8]. We propose 2 potential explanations for this discrepancy.

First, generative AI systems differ fundamentally from earlier AI health care products, such as image recognition [14,50] and wearable devices [51]. These legacy tools operate within predictable parameters (eg, identifying tumors and tracking heart rates), while generative AI engages in more open-ended interactions that blur the human-machine boundary. Prior AI health care products have already achieved some maturity through years of public exposure, and people may already be familiar with them.

Second, ChatGPT and other large language models are mostly general purpose AI. Different from AI models specifically trained for medical purposes, these models are not intended to be used in clinical scenarios without human oversight [52]. Given their broad applicability and direct accessibility, patients may question the professionalism or judgment of doctors who use such tools. This skepticism may be a result of valid concerns: there is evidence that AI-assisted teams may perform inefficiently because of cognitive challenges such as overreliance on the AI [53] and the lack of metaknowledge [54]. This finding may extend to other nonmedical, unfamiliar technologies: when tools lack a clear medical focus and public familiarity, their integration into health care may lower patient acceptance and satisfaction.

Why Do AI Knowledge and AI Experience Shape Trust Differently?

As an exploratory analysis, we found that the “trust gap” was smallest among those who used AI most frequently. One explanation is that regular AI use fosters comfort and familiarity [14,23], which, in turn, enhances trust in AI-assisted health care [55,56]. Previous literature has indicated that familiarity increases perceived ease of use [35,38] and thus increases support for the adoption of AI in areas including autonomous systems [23], retail chatbots [55], and travel planning [56,57].

Yet, participants who reported the highest AI knowledge exhibited the largest trust gap. This finding seems counterintuitive at first glance because greater knowledge of technology is typically associated with increased understanding and trust, rather than heightened skepticism [58]. Nonetheless, recent studies [23] do suggest that “knowledge” does not correlate with acceptance in AI systems as “using experience” does. While this finding provides some support for our results, it does not fully explain why we observed a negative correlation. To address this, we propose the following reasons why the negative effect of AI assistance expanded as self-reported AI knowledge increased.

First, a potential channel is that the higher the self-reported AI knowledge, the more aware respondents were of generative AI’s limitations, risks, and ethical concerns [23]. As general AI is not specifically designed for diagnostics, those with greater awareness may distrust its use in high-stakes contexts. This heightened awareness can amplify concerns about reliability, accuracy, and maturity in medical settings.

Second, the awareness of AI’s limitations may stem from perceived rather than actual risks. People confident in their belief that health care AI is inadequate may misunderstand its true capabilities. Common concerns, such as AI’s rigidity and inability to personalize care, often drive aversion [4,39], although AI can sometimes outperform humans in these areas [3,19]. Those who strongly believe in these misconceptions may report high AI knowledge while simultaneously exhibiting AI aversion.

There is also the possibility of overconfidence, given respondents self-reported their perceived level of AI knowledge [59,60]. For instance, recent research on women’s attitudes toward medical AI [61] found that respondents with lower educational levels (which correlates with less real knowledge) reported a higher self-perceived accuracy of AI knowledge, consistent with our overconfidence hypothesis. Similarly, studies on health literacy suggested that individuals with higher self-reported knowledge may possess lower actual literacy [60,62]. These patterns align with our findings, as self-reported AI knowledge may not reflect true AI literacy but rather an overestimation of one’s understanding coupled with an underestimation of AI’s capabilities. Such overconfidence can lead to an exaggerated perception of AI’s weaknesses, contributing to the observed trust gap. Unlike self-reported knowledge, experience reflects real interactions (eg, using AI at work or in daily life). This practical experience likely offers a more accurate view of AI’s strengths and weaknesses, which may explain why frequent users showed less aversion.

Theoretical Contributions

This research has the following theoretical implications. First, we extended previous literature on medical AI aversion [4,39,40] and found that aversion widely persists even if AI is used as an assisting tool rather than a sole decision maker in the era of generative AI. We also found that the aversion to AI in health care is not confined to any specific demographic or ideological group, suggesting that concerns about AI-driven diagnoses are broadly held rather than politically polarized.

Second, we challenge assumptions about the role of AI familiarity in mitigating skepticism. While prior studies suggest that increased exposure to AI enhances trust and we obtained similar results, we found that self-reported AI knowledge widened the trust gap, whereas frequent AI use narrowed it. The proposed reasons in the previous section contribute to theories of algorithmic trust, highlighting differential roles of perceived expertise (knowledge) versus experiential familiarity. This necessitates better measures for AI familiarity.

Third, our findings contribute to the literature on trust and decision-making. Trust in the doctor’s personal integrity and professional competence are key mediators of AI-assistance aversion. In our case, trust acts as a barrier to AI-assisted health care adoption, as lower trust in doctors who use AI directly reduces patients’ intention to seek medical help. Trust calibration (ensuring alignment between perceived and actual AI reliability) remains a challenge for medical AI integration.

Practical Implications

The findings of this study offer important practical insights for health care professionals, policy makers, and AI developers working to integrate AI into decision-making.

First, we found that distrust and aversion appear when patients know that the doctor uses AI (such as ChatGPT), even if it serves as a supporting tool rather than a primary decision maker. Therefore, merely incorporating AI into medical diagnoses without solid trust-building ethics may inadvertently reduce patient’s trust. Understanding and addressing this trust barrier is essential for ensuring the successful adoption of AI-assisted health care.

Second, our results highlight the possible need for differentiating between self-reported AI knowledge and AI experience when designing patient education programs. Hands-on experiences with AI can demystify AI and increase trust and visiting intention [38]. Conversely, people with higher self-reported knowledge of AI showed larger gaps in trust and visiting intention. If individuals distrust AI based on self-perceived (but possibly inaccurate) knowledge, traditional knowledge-based educational interventions, especially if coming from sources already viewed with skepticism, may backfire. Communication strategies by governments and medical practitioners should be carefully designed to account for these biases, potentially incorporating experiential learning opportunities rather than relying solely on informational campaigns.

Third, we found that trust in the doctor and intention to seek help were highest when patients knew that the doctor does not use AI. This highlights the need for AI transparency in clinical settings.

Limitations and Future Directions

One limitation of this study is sample selection. First, it used data only from the United States. Due to differences in economic development and culture, findings from the United States may not fully apply to other countries. Factors influencing trust in AI-assisted health care could vary across regions, and similar patterns of AI skepticism may not hold in different social environments [63] and cultural backgrounds [63-66]. Future research should examine cross-national patterns of AI acceptance.

Second, the Prolific sample may not reflect the general population, as its users are more inclined to use web-based services. However, we argue that in this study, this distortion is less problematic. Time spent online is positively associated with openness to experience [67,68], a trait linked to greater acceptance of AI [69]. Therefore, our estimate of AI aversion may represent a lower bound for the general population.

A recent study by the University of Queensland, covering 17 countries [64], found that US attitudes toward AI are similar to those in other high-income nations but more skeptical than in countries such as China, India, and South Africa. This supports partial generalizability to advanced health systems but calls for further research in transitional contexts.

We also used ChatGPT only as a demonstration for AI-assisted diagnostic tools. It is valuable to study different forms of AI assistance, especially those with growing popularity in recent years. Also, a clear comparison between generative AI and other categories of medical AI is pending further exploration.

This study did not investigate ethical design and transparent communication strategies; yet, a successful integration of AI into health care requires attention to those aspects. Prior literature [40] suggested 5 possible strategies to improve AI acceptance: using explainable AI, anthropomorphizing, advertising AI as flexible, restoring user control, and convincing users of AI’s human-like consciousness. Without applying these strategies, AI-assisted diagnosis, particularly involving general purpose tools such as ChatGPT, may receive less acceptance. Future studies should explore how ethical framing, design transparency, and tailored communication can jointly reduce distrust and support safe AI adoption.

Conclusions

Our research found a common aversion to AI-assisted diagnosis, especially within the domain of generative AI, and this aversion does not significantly vary across demographic and ideological cohorts. This highlights the potential challenges to the application of AI-assisted health care. We find that despite the benefits of AI in health care, there is still a strong preference for human expertise, which could slow the adoption of AI-assisted tools.

Acknowledgments

The authors thank Eugene Caruso, Jiuchang Wei, and Ming Xu for their insightful comments. All faults remain the authors' own.

Authors' Contributions

According to the CRediT taxonomy, the detailed author contributions are listed here. CC and ZC contributed to conceptualization, data curation, funding acquisition, investigation, methodology, project administration, resources, software, validation, writing—original draft, and writing—review and editing. CC led the formal analysis and supervision. ZC participated in visualization.

Conflicts of Interest

None declared.

Multimedia Appendix 1

A Screenshot of the intervention. The left panel indicates the screen seen by a PC user, and the right panel indicates the screen seen by a smartphone user.

PNG File, 278 KB

Multimedia Appendix 2

Extended results and methods.

DOCX File, 45 KB

Checklist 1

CONSORT-eHEALTH checklist (V 1.6.1).

PDF File, 4064 KB

  1. Benjamens S, Dhunnoo P, Meskó B. The state of artificial intelligence-based FDA-approved medical devices and algorithms: an online database. NPJ Digit Med. 2020;3(1):118. [CrossRef] [Medline]
  2. Young AT, Amara D, Bhattacharya A, Wei ML. Patient and general public attitudes towards clinical artificial intelligence: a mixed methods systematic review. Lancet Digit Health. Sep 2021;3(9):e599-e611. [CrossRef] [Medline]
  3. Eriksen AV, Möller S, Ryg J. Use of GPT-4 to diagnose complex clinical cases. NEJM AI. Jan 2024;1(1). [CrossRef]
  4. Longoni C, Bonezzi A, Morewedge CK. Resistance to medical artificial intelligence. J Consum Res. Dec 1, 2019;46(4):629-650. [CrossRef]
  5. Mello MM, Guha N. Understanding liability risk from using health care artificial intelligence tools. N Engl J Med. Jan 18, 2024;390(3):271-278. [CrossRef] [Medline]
  6. Alowais SA, Alghamdi SS, Alsuhebany N, et al. Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med Educ. Sep 22, 2023;23(1):689. [CrossRef] [Medline]
  7. Nelson CA, Pérez-Chada LM, Creadore A, et al. Patient perspectives on the use of artificial intelligence for skin cancer screening: a qualitative study. JAMA Dermatol. May 1, 2020;156(5):501-512. [CrossRef] [Medline]
  8. Beets B, Newman TP, Howell EL, Bao L, Yang S. Surveying public perceptions of artificial intelligence in health care in the United States: systematic review. J Med Internet Res. 25:e40337. [CrossRef]
  9. Jutzi TB, Krieghoff-Henning EI, Holland-Letz T, et al. Artificial intelligence in skin cancer diagnostics: the patients’ perspective. Front Med (Lausanne). 2020;7:233. [CrossRef] [Medline]
  10. Palmisciano P, Jamjoom AAB, Taylor D, Stoyanov D, Marcus HJ. Attitudes of patients and their relatives toward artificial intelligence in neurosurgery. World Neurosurg. Jun 2020;138:e627-e633. [CrossRef] [Medline]
  11. Chew HSJ, Achananuparp P. Perceptions and needs of artificial intelligence in health care to increase adoption: scoping review. J Med Internet Res. Jan 14, 2022;24(1):e32939. [CrossRef] [Medline]
  12. Keel S, Lee PY, Scheetz J, et al. Feasibility and patient acceptability of a novel artificial intelligence-based screening model for diabetic retinopathy at endocrinology outpatient services: a pilot study. Sci Rep. Mar 12, 2018;8(1):4330. [CrossRef] [Medline]
  13. Yarborough BJH, Stumbo SP. Patient perspectives on acceptability of, and implementation preferences for, use of electronic health records and machine learning to identify suicide risk. Gen Hosp Psychiatry. 2021;70:31-37. [CrossRef] [Medline]
  14. Ho V, Brown Johnson C, Ghanzouri I, Amal S, Asch S, Ross E. Physician- and patient-elicited barriers and facilitators to implementation of a machine learning-based screening tool for peripheral arterial disease: preimplementation study with physician and patient stakeholders. JMIR Cardio. Nov 6, 2023;7:e44732. [CrossRef] [Medline]
  15. Nadarzynski T, Miles O, Cowie A, Ridge D. Acceptability of artificial intelligence (AI)-led chatbot services in healthcare: a mixed-methods study. Digit Health. 2019;5:2055207619871808. [CrossRef] [Medline]
  16. Davis CR, Murphy KJ, Curtis RG, Maher CA. A process evaluation examining the performance, adherence, and acceptability of a physical activity and diet artificial intelligence virtual health assistant. Int J Environ Res Public Health. Dec 7, 2020;17(23):9137. [CrossRef] [Medline]
  17. Thirunavukarasu AJ, Ting DSJ, Elangovan K, Gutierrez L, Tan TF, Ting DSW. Large language models in medicine. Nat Med. Aug 2023;29(8):1930-1940. [CrossRef] [Medline]
  18. Kung TH, Cheatham M, Medenilla A, et al. Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digit Health. Feb 2023;2(2):e0000198. [CrossRef] [Medline]
  19. Nori H, King N, McKinney SM, Carignan D, Horvitz E. Capabilities of GPT-4 on medical challenge problems. arXiv. Preprint posted online on 2023. [CrossRef]
  20. Kucera T, Togninalli M, Meng-Papaxanthos L. Conditional generative modeling for de novo protein design with hierarchical functions. Bioinformatics. Preprint posted online on 2022. [CrossRef]
  21. Moor M, Banerjee O, Abad ZSH, et al. Foundation models for generalist medical artificial intelligence. Nature New Biol. Apr 2023;616(7956):259-265. [CrossRef] [Medline]
  22. Lee P, Bubeck S, Petro J. Benefits, limits, and risks of GPT-4 as an AI chatbot for medicine. N Engl J Med. Mar 30, 2023;388(13):1233-1239. [CrossRef]
  23. Horowitz MC, Kahn L, Macdonald J, Schneider J. Adopting AI: how familiarity breeds both trust and contempt. AI Soc. May 12, 2023;39(4):1-15. [CrossRef] [Medline]
  24. Khanna NN, Maindarkar MA, Viswanathan V, et al. Economics of artificial intelligence in healthcare: diagnosis vs. treatment. Healthcare (Basel). Dec 9, 2022;10(12):2493. [CrossRef]
  25. Jabbar MA, Iqbal H, Chawla U. Patient satisfaction: the role of artificial intelligence in healthcare. J Health Manag. May 15, 2024. [CrossRef]
  26. Holden RJ, Karsh BT. The technology acceptance model: its past and its future in health care. J Biomed Inform. Feb 2010;43(1):159-172. [CrossRef] [Medline]
  27. Beldad AD, Hegner SM. Expanding the technology acceptance model with the inclusion of trust, social influence, and health valuation to determine the predictors of German users’ willingness to continue using a fitness app: a structural equation modeling approach. Int J Hum Comput Interact. Sep 2, 2018;34(9):882-893. [CrossRef]
  28. Choung H, David P, Ross A. Trust in AI and its role in the acceptance of AI technologies. Int J Hum Comput Interact. May 28, 2023;39(9):1727-1739. [CrossRef]
  29. Renn O, Benighaus C. Perception of technological risk: insights from research and lessons for risk communication and management. J Risk Res. Apr 2013;16(3-4):293-313. [CrossRef]
  30. Bahmanziari T, Pearson JM, Crosby L. Is trust important in technology adoption? A policy capturing approach. J Comput Inform Syst. Sep 2003;43(4):46-54. [CrossRef]
  31. Juravle G, Boudouraki A, Terziyska M, Rezlescu C. Trust in artificial intelligence for medical diagnoses. In: Progress in Brain Research. Elsevier; 2020:263-282. [CrossRef] ISBN: 978-0-12-820723-9
  32. Shevtsova D, Ahmed A, Boot IWA, et al. Trust in and acceptance of artificial intelligence applications in medicine: mixed methods study. JMIR Hum Factors. Jan 17, 2024;11:e47031. [CrossRef] [Medline]
  33. Ye T, Xue J, He M, et al. Psychosocial factors affecting artificial intelligence adoption in health care in China: cross-sectional study. J Med Internet Res. Oct 17, 2019;21(10):e14316. [CrossRef] [Medline]
  34. Bach TA, Khan A, Hallock H, Beltrão G, Sousa S. A systematic literature review of user trust in AI-enabled systems: an HCI perspective. Int J Hum Comput Interact. Mar 3, 2024;40(5):1251-1266. [CrossRef]
  35. Davis FD. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Q. Sep 1989;13(3):319. [CrossRef]
  36. Quispel A, Maes A, Schilperoord J. Graph and chart aesthetics for experts and laymen in design: the role of familiarity and perceived ease of use. Inf Visualization. Jul 2016;15(3):238-252. [CrossRef]
  37. Choi JC. User familiarity and satisfaction with food delivery mobile apps. Sage Open. Oct 2020;10(4):2158244020970563. [CrossRef]
  38. Cabiddu F, Moi L, Patriotta G, Allen DG. Why do users trust algorithms? A review and conceptualization of initial trust and trust over time. Eur Manag J. Oct 2022;40(5):685-706. [CrossRef]
  39. Cadario R, Longoni C, Morewedge CK. Understanding, explaining, and utilizing medical artificial intelligence. Nat Hum Behav. Dec 2021;5(12):1636-1642. [CrossRef] [Medline]
  40. De Freitas J, Agarwal S, Schmitt B, Haslam N. Psychological factors underlying attitudes toward AI tools. Nat Hum Behav. Nov 2023;7(11):1845-1854. [CrossRef] [Medline]
  41. Rutjens BT, Sutton RM, van der Lee R. Not all skepticism is equal: exploring the ideological antecedents of science acceptance and rejection. Pers Soc Psychol Bull. Mar 2018;44(3):384-405. [CrossRef] [Medline]
  42. Castelo N, Ward AF. Conservatism predicts aversion to consequential artificial intelligence. PLoS One. 2021;16(12):e0261467. [CrossRef] [Medline]
  43. Robertson C, Woods A, Bergstrand K, Findley J, Balser C, Slepian MJ. Diverse patients’ attitudes towards artificial Intelligence (AI) in diagnosis. PLOS Digit Health. May 2023;2(5):e0000237. [CrossRef] [Medline]
  44. Palan S, Schitter C. Prolific.ac—a subject pool for online experiments. J Behav Exp Finance. Mar 2018;17:22-27. [CrossRef]
  45. Douglas BD, Ewell PJ, Brauer M. Data quality in online human-subjects research: comparisons between MTurk, Prolific, CloudResearch, Qualtrics, and SONA. PLoS One. 2023;18(3):e0279720. [CrossRef] [Medline]
  46. Foster ED, Deardorff A. Open Science Framework (OSF). J Med Libr Assoc. Apr 2017;105(2). [CrossRef]
  47. Nieminen P, Lehtiniemi H, Vähäkangas K, Huusko A, Rautio A. Standardised regression coefficient as an effect size index in summarising findings in epidemiological studies. Epidemiol Biostat Public Health. Jul 4, 2022;10(4). [CrossRef]
  48. Dawson JF. Moderation in management research: what, why, when, and how. J Bus Psychol. Mar 2014;29(1):1-19. [CrossRef]
  49. Tingley D, Yamamoto T, Hirose K, Keele L, Imai K. Mediation: R Package for causal mediation analysis. J Stat Softw. 2014;59(5):1-38. [CrossRef]
  50. Shah P, Mishra D, Shanmugam M, Vighnesh MJ, Jayaraj H. Acceptability of artificial intelligence-based retina screening in general population. Indian J Ophthalmol. Apr 2022;70(4):1140-1144. [CrossRef] [Medline]
  51. Park E. User acceptance of smart wearable devices: an expectation-confirmation model approach. Telematics Inf. Apr 2020;47:101318. [CrossRef]
  52. OpenAI. Introducing the Model Spec. 2024. URL: https://openai.com/index/introducing-the-model-spec/ [Accessed 2025-04-21]
  53. Buçinca Z, Malaya MB, Gajos KZ. To trust or to think: cognitive forcing functions can reduce overreliance on AI in AI-assisted decision-making. Proc ACM Hum Comput Interact. 2021;5:1-21. [CrossRef]
  54. Fügener A, Grahl J, Gupta A, Ketter W. Cognitive challenges in human–artificial intelligence collaboration: investigating the path toward productive delegation. Inf Syst Res. Jun 2022;33(2):678-696. [CrossRef]
  55. Arce-Urriza M, Chocarro R, Cortiñas M, Marcos-Matás G. From familiarity to acceptance: the impact of generative artificial intelligence on consumer adoption of retail chatbots. J Retail Consum Serv. May 2025;84:104234. [CrossRef]
  56. Topsakal Y. How familiarity, ease of use, usefulness, and trust influence the acceptance of generative artificial intelligence (AI)-assisted travel planning. Int J Hum Comput Interact. 2024;40:1-14. [CrossRef]
  57. Christensen J, Hansen JM, Wilson P. Understanding the role and impact of generative artificial intelligence (AI) hallucination within consumers’ tourism decision-making processes. Curr Issues Tour. Feb 16, 2025;28(4):545-560. [CrossRef]
  58. Kim S. Perceptions of discriminatory decisions of artificial intelligence: unpacking the role of individual characteristics. Int J Hum Comput Stud. Feb 2025;194:103387. [CrossRef]
  59. Berner ES, Graber ML. Overconfidence as a cause of diagnostic error in medicine. Am J Med. May 2008;121(Suppl 5):S2-S23. [CrossRef] [Medline]
  60. Canady BE, Larzo M. Overconfidence in managing health concerns: the Dunning-Kruger effect and health literacy. J Clin Psychol Med Settings. Jun 2023;30(2):460-468. [CrossRef] [Medline]
  61. Pesapane F, Rotili A, Valconi E, et al. Women’s perceptions and attitudes to the use of AI in breast cancer screening: a survey in a cancer referral centre. Br J Radiol. Jan 1, 2023;96(1141):20220569. [CrossRef] [Medline]
  62. Dunning D. The Dunning-Kruger effect. In: Advances in Experimental Social Psychology. Elsevier; 2011:247-296. [CrossRef] ISBN: 978-0-12-385522-0
  63. Bompelli A, Wang Y, Wan R, et al. Social and behavioral determinants of health in the era of artificial intelligence with electronic health records: a scoping review. Health Data Sci. 2021;2021:9759016. [CrossRef] [Medline]
  64. Gillespie N, Lockey S, Curtis C, Pool J, Akbari A. Trust in Artificial Intelligence: A Global Study. The University of Queensland; KPMG Australia; 2023. [CrossRef]
  65. Yam KC, Tan T, Jackson JC, Shariff A, Gray K. Cultural differences in people’s reactions and applications of robots, algorithms, and artificial intelligence. Manag Organ Rev. Oct 2023;19(5):859-875. [CrossRef]
  66. Agrawal V, Kandul S, Kneer M, Christen M. From OECD to India: exploring cross-cultural differences in perceived trust, responsibility and reliance of AI and human experts. arXiv. Preprint posted online on 2023. [CrossRef]
  67. Svendsen GB, Johnsen JA, Almås-Sørensen L, Vittersø J. Personality and technology acceptance: the influence of personality factors on the core constructs of the Technology Acceptance Model. Behav Inf Technol. Apr 2013;32(4):323-334. [CrossRef]
  68. Özbek V, Alnıaçık Ü, Koc F, Akkılıç ME, Kaş E. The impact of personality on technology acceptance: a study on smart phone users. Procedia Soc Behav Sci. Sep 2014;150:541-551. [CrossRef]
  69. de Winter J, Dodou D, Eisma YB. Personality and acceptance as predictors of ChatGPT use. Discov Psychol. May 14, 2024;4(1):57. [CrossRef]


ACME: average causal mediation effect
ADE: average direct effect
AI: artificial intelligence
OLS: ordinary least squares
TAM: technology acceptance model


Edited by Javad Sarvestan; submitted 03.09.24; peer-reviewed by Alfredo Vellido, Mihaela Vorvoreanu, Soroosh Tayebi Arasteh, Taiwo Akindahunsi, Zarmina Ehtesham; final revised version received 23.04.25; accepted 23.04.25; published 18.06.25.

Copyright

© Catherine Chen, Zhihan Cui. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 18.6.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.