Published on in Vol 25 (2023)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/44602, first published .
Performance-Based Measurement of eHealth Literacy: Systematic Scoping Review

Performance-Based Measurement of eHealth Literacy: Systematic Scoping Review

Performance-Based Measurement of eHealth Literacy: Systematic Scoping Review

Authors of this article:

Bradley Crocker1 Author Orcid Image ;   Olivia Feng1 Author Orcid Image ;   Lindsay R Duncan1 Author Orcid Image

Review

Department of Kinesiology & Physical Education, McGill University, Montreal, QC, Canada

Corresponding Author:

Bradley Crocker, MA

Department of Kinesiology & Physical Education

McGill University

475 Pine Ave W

Montreal, QC, H2W 1S4

Canada

Phone: 1 514 398 4184 ext 00919

Email: bradley.crocker@mail.mcgill.ca


Background: eHealth literacy describes the ability to locate, comprehend, evaluate, and apply web-based health information to a health problem. In studies of eHealth literacy, researchers have primarily assessed participants’ perceived eHealth literacy using a short self-report instrument, for which ample research has shown little to no association with actual performed eHealth-related skills. Performance-based measures of eHealth literacy may be more effective at assessing actual eHealth skills, yet such measures seem to be scarcer in the literature.

Objective: The primary purpose of this study was to identify tools that currently exist to measure eHealth literacy based on objective performance. A secondary purpose of this study was to characterize the prevalence of performance-based measurement of eHealth literacy in the literature compared with subjective measurement.

Methods: We conducted a systematic scoping review of the literature, aligning with the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) checklist, in 3 stages: conducting the search, screening articles, and extracting data into a summary table. The summary table includes terminology for eHealth literacy, description of participants, instrument design, health topics used, and a brief note on the evidence of validity for each performance-based measurement tool. A total of 1444 unique articles retrieved from 6 relevant databases (MEDLINE; PsycINFO; CINAHL; Library and Information Science Abstracts [LISA]; Library, Information Science & Technology Abstracts [LISTA]; and Education Resources Information Center [ERIC]) were considered for inclusion, of which 313 (21.68%) included a measure of eHealth literacy.

Results: Among the 313 articles that included a measure of eHealth literacy, we identified 33 (10.5%) that reported on 29 unique performance-based eHealth literacy measurement tools. The types of tools ranged from having participants answer health-related questions using the internet, having participants engage in simulated internet tasks, and having participants evaluate website quality to quizzing participants on their knowledge of health and the web-based health information–seeking process. In addition, among the 313 articles, we identified 280 (89.5%) that measured eHealth literacy using only a self-rating tool.

Conclusions: This study is the first research synthesis looking specifically at performance-based measures of eHealth literacy and may direct researchers toward existing performance-based measurement tools to be applied in future projects. We discuss some of the key benefits and drawbacks of different approaches to performance-based measurement of eHealth literacy. Researchers with an interest in gauging participants’ actual eHealth literacy (as opposed to perceived eHealth literacy) should make efforts to incorporate tools such as those identified in this systematic scoping review.

J Med Internet Res 2023;25:e44602

doi:10.2196/44602

Keywords



Background

Individuals form beliefs, make decisions, and enact behavior based on what they perceive to be high-quality and relevant information [1]. Ideas and beliefs about healthy and unhealthy behavior are shaped from various sources, but, above all, the internet plays an increasingly prominent role in people’s information diet regarding health-related topics. Research indicates that the internet is the first source people turn to for health information [2] and in some cases is deemed more credible than physician diagnoses [3]. The internet provides several affordances for accessing health information that cannot be matched by other resources: it is available 24 hours a day, queries are answered immediately, queries can be made anonymously, multiple points of view can be accessed and considered by the information seeker, and all of this is offered in the comfort of one’s own home [4]. Improved public access to health-related information concerning treatment and prevention has the potential to make a dramatic positive impact on public health outcomes. However, the realization of this positive impact on health-related behavior largely depends on how the internet is used by the information seeker.

Although the affordances of the internet as a health information resource are plentiful, locating high-quality health information on the internet can be challenging. The volume of health information on the web is massive and perpetually increasing; for example, a Google search for “sore throat treatment” returns nearly 2 billion results. Within this plethora of information, users’ search results are often clouded with misleading, inaccurate, or unsubstantiated information [5]; for example, in a sample of 227 web pages related to “immunity boosting” and “coronavirus,” Rachul et al [6] found that 85.5% of the sources portrayed “immune boosting” as beneficial for preventing COVID-19 infection despite no existent scientific evidence. Information seekers are tasked with sifting through search engine results to locate something relevant, trustworthy, and in an accessible format to inform their decisions. Norman and Skinner [7] describe this ability as eHealth literacy, which they define as “the ability to seek, find, understand, and appraise health information from electronic sources and apply the knowledge gained to addressing or solving a health problem” [7].

The construct of eHealth literacy was initially derived from 6 underlying foundational literacies: traditional literacy, information literacy, media literacy, health literacy, computer literacy, and scientific literacy [7]. Norgaard et al [8] extended this model further to identify 7 domains that ultimately contribute to an individual’s eHealth literacy: ability to process information (locating, interpreting, and applying health information), engagement in own health (interest in, and basic knowledge of, personal health and health care systems), ability to actively engage with digital services (competency with technology and navigating the internet), feel safe and in control (confidence and knowledge of securing personal health information), motivated to engage with digital services (accepting attitude toward web-based health resources and services), access to digital services that work (having the hardware and software to access web-based health resources and services), and digital services that suit individual needs (web-based health resources exist that are accessible and understandable to the individual user). Taken together, these models of eHealth literacy depict the diverse combination of critical skills needed to effectively acquire health information on the web.

Concurrent with their publication coining eHealth literacy, Norman and Skinner [9] published the eHealth Literacy Scale (eHEALS), an 8-item questionnaire wherein each item is rated using a 5-point Likert scale ranging from 1 (strongly disagree) to 5 (strongly agree), with higher scores indicating higher perceived eHealth literacy. In a systematic review of eHealth literacy measurement tools, Karnoe and Kayser [10] noted that this was the only tool used in multiple studies at the time of publication, and it remains the most widely used instrument across the literature to assess eHealth literacy [11]. As a short self-report assessment tool, the eHEALS is attractive for health care providers looking to optimize clinical efficiency and convenient for researchers to minimize participant burden. However, the self-report nature of this measure may limit its usefulness at providing an accurate depiction of how well people critically engage with web-based health information in today’s internet landscape.

Although self-report measurements have functioned as invaluable research tools in several settings, the efficacy of using self-report to evaluate one’s own skills, abilities, or proficiency in many contexts has been criticized for decades [12,13]. In their seminal review and meta-analysis of self-evaluations of ability across 55 studies, Mabe and West [12] found only a very modest (mean r=0.29) relationship between self-reported ability and performance-based measurement. Several psychological mechanisms may contribute to the considerable gap in estimates of ability and actual ability, many of which have been supported by research across a variety of cognitive skills [14]; for example, individuals have a tendency to be confident in skills where they feel fluent [14], meaning people may mistake the ability to fluidly perform a skill with the ability to perform that skill well. In the context of web-based health information seeking, people may mistake the ability to seamlessly locate some health information on the internet with the distinct ability to retrieve accurate, trustworthy, and relevant web-based health information. Many avid followers of the antivaccine movement have spent several hours surfing the internet for health-related information [15,16]. Their relatively high experience of web-based health information seeking may lead them to feel fluent in the skill, yet they are evidently not equipped to judge the quality of web-based information by the standards of evidence-based medicine.

Relating to web-based information seeking, ample literature indicates that people have a tendency to overestimate their ability with computers [17,18] as well as their ability to locate and understand information on the web [19,20]. Frequent overestimation of ability in this context may result from overlap between the cognitive skills necessary for the task itself and the metacognitive skills necessary for accurate self-assessment, meaning that those without the competence to consistently identify trustworthy information on the web are unlikely to have the capacity to recognize their lack of competence. Dunning [14] has put it more plainly: “If they had the skill to know they were making consistent mistakes, they would also have the skill to avoid those blunders in the first place.” Fittingly, in the context of web-based health information seeking, Stellefson et al [21] found that the least proficient searchers commonly exhibited high confidence in their searching abilities, indicating that self-efficacy is not a consistent predictor of eHealth literacy skills. Similarly, Maitz et al [22] found that adolescents who reported higher eHEALS scores did not perform better at selecting higher-quality websites to access health information. Illustrating this point further, Brown and Dickson [23] found that a group of occupational therapy graduate students reported lower average eHEALS scores than the high school participants in the original paper by Norman and Skinner [9]. Taken together, this literature indicates that intellectual hubris may play a bigger role in self-rated eHealth literacy than actual knowledge or ability; therefore, it should come as little surprise that a bulk of studies have noted no or little association between eHEALS scores and demonstrated eHealth skills [24-26]. Measures of eHealth literacy that involve more objective performance-based measurement of eHealth skills and abilities may be more promising at assessing how people actually behave on the web; however, such measures seem to be more scarcely used among researchers to date [11].

Although research syntheses have been undertaken related to eHealth literacy measurement instruments and their properties [10,27], no synthesis to date has looked specifically at performance-based measures of eHealth literacy. This work is needed to assist researchers in identifying performance-based measurement tools in the literature that may be applied in future projects.

Objectives

The purposes of this study were 2-fold. The first purpose was to identify tools that currently exist to measure eHealth literacy based on objective performance. Within this purpose, we were interested in the design of these tools, their intended population, and whether their authors included some evidence of validity. The second purpose of this study was to characterize the prevalence of performance-based eHealth literacy measurement tools in the literature broadly compared with subjective eHealth literacy measurement tools.


Overview

We conducted a systematic scoping review of the literature. A scoping review was deemed appropriate for the purpose of this study, given its broad focus on how research involving eHealth literacy measurement is conducted [28]. This scoping review was conducted in 3 main stages, aligning with the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) checklist [29].

First Stage

In the first stage, we devised an electronic search protocol that was completed by 2 researchers. We used the following search string, modified from the one used in a systematic review conducted by Karnoe and Kayser [10]:

(“eHealth literacy” OR “electronic health literacy” OR “e-health literacy” OR “digital health literacy” OR (“health literacy” AND “digital literacy”) OR (“health literacy” AND “computer literacy”)) AND (scale OR measure OR survey OR questionnaire OR test OR assessment)

We applied this string to search 6 relevant databases (MEDLINE; PsycINFO; CINAHL; Library and Information Science Abstracts [LISA]; Library, Information Science & Technology Abstracts [LISTA]; and Education Resources Information Center [ERIC]) selected in consultation with a university librarian. All searches were conducted in June 2021. Across these 6 databases, 1735 search results were obtained, which we reduced to 1444 (83.23%) unique articles after removing duplicates.

Second Stage

In the second stage, we conducted 3 screening phases to narrow down the search results. In each screening phase, articles were sought that met the following inclusion criteria: (1) written in English; (2) published in peer-reviewed journals; and (3) created, revised, or used an eHealth literacy assessment tool (in accordance with the definition of eHealth literacy proposed by Norman and Skinner [7], quoted earlier in this paper). The first screening phase involved screening all articles using their titles, which was carried out by the first author. After this phase, of the 1444 articles, 788 (54.57%) were retained for possible inclusion. The second screening phase involved screening all remaining articles using their abstracts, which was carried out by the first and second authors. If either author opted to include the article, it was kept for the third screening phase. After this phase, of the 788 articles, 374 (47.5%) were retained for possible inclusion. The third screening phase involved reading the full text of all remaining articles, which was carried out by the first and second authors. During this phase, each author also classified each included article as containing only a subjective eHealth literacy measure (based on self-rated assessment) or as containing a performance-based eHealth literacy measure (based on practical skill or knowledge assessment).

After the third screening phase, of the 374 articles, 311 (83.2%) were retained for inclusion in this scoping review; 31 (10%) of these 311 articles contained a performance-based measure of eHealth literacy. In cases of disagreement between the first and second authors during the third screening phase, the third author was consulted, and discussion ensued until consensus was reached. Throughout the entire screening process, all authors only excluded articles that very evidently did not meet the inclusion criteria. At this point, the first author read the bibliographies of the articles containing a performance-based measure of eHealth literacy (n=31) and highlighted titles that could meet the inclusion criteria of this scoping review. After discarding studies that had already been considered in the screening process, new studies were selected (n=6) and read in full by the first and second authors. Of these 6 new studies, 2 (33%) were included in the scoping review, and both contained a performance-based measure of eHealth literacy. Thus, the total number of articles included in this scoping review is 313, of which 33 (10.5%) contain a performance-based measure of eHealth literacy. The scoping review process we undertook is summarized in Figure 1.

Figure 1. Diagram summary of the systematic scoping review process.

Third Stage

In the third stage, we extracted data from each of the articles containing a performance-based measure and summarized them into a table. The first and second authors each reread the articles containing a performance-based measure of eHealth literacy (n=33) and input information from each into a comprehensive table devised by all 3 authors. This table was later simplified into appropriate headings, selected in the interest of providing a concise and relevant summary of the findings in this paper. In judging evidence of validity, we considered 3 relevant types of validity: ecological validity, criterion validity, and construct validity. Ecological validity describes the extent to which a measure represents or predicts behavior in real-world settings [30]. In the context of measuring performance-based eHealth literacy, we considered a measure to be ecologically valid if it involved participants accessing real or authentically simulated websites to gather or evaluate health information. Criterion validity describes the extent to which the scores of a measure predict scores of another established measure of interest [30]. We stated that a measure had criterion validity if the authors found statistically significant correlations between scores of their objective eHealth literacy measure and scores from another relevant construct or validated measure (eg, higher scores on this measure correlate with higher scores of another validated measure of computer-related skills). Construct validity describes the extent to which the scores of a measure reflect the actual phenomena that measure intends to assess [30]. This concept has some overlap with criterion validity; however, for our purposes, we stated that a measure had construct validity if meaningful differences were reported based on relevant lifestyle or demographic factors or reported among groups that can be reasonably assumed to have different levels of eHealth literacy (eg, those with more health-related education score consistently higher). We also noted whether the performance-based measurement tool described in the article was included in its full form, partial form (examples provided but some parts omitted), or not at all.


Performance-Based eHealth Literacy Measurement Tools

Our scoping review identified 33 peer-reviewed studies using a tool to measure eHealth literacy that incorporated a performance-based aspect. Within these studies, there were just 2 measures used more than once. In total, 3 (9%) of the 33 articles used the eHealth literacy assessment toolkit (eHLA) created by Karnoe et al [31], and another 3 (9%) used the Research Readiness Self-Assessment created by Ivanitskaya et al [32]. We thus identified a total of 29 unique performance-based measures of eHealth literacy. It is notable that many of these studies use differing terminology to describe the construct of eHealth literacy, as outlined in Table 1; however, all studies included in this review measure constructs aligning with the definition of eHealth literacy originally proposed by Norman and Skinner [7] based on our judgment.

We categorized the 29 unique performance-based measurement tools identified in this scoping review into 5 broad categories according to the structure of their main measurement technique. It should be noted that several of the measurement tools have components that fall into ≥2 of these categories; for instance, the Research Readiness Self-Assessment-health (RRSA-h) [53,55] measures declarative knowledge about the internet and has participants respond to questions that simulate using web-based health information. We have described the instrument design of each performance-based measurement tool in Table 1 such that readers can gain a fuller understanding of each tool compared with the brief descriptions within the body of this paper.

Table 1. Summary of included articles containing a performance-based measure of eHealth literacy (organized alphabetically by category).
Authors, yearTerminology for measured constructParticipants and contextInstrument designHealth topicsEvidence of validityInstrument availability
Answering health-related questions using the internet

Agree et al [33], 2015Web-based health literacy323 participants aged 35 to 90 yearsSix health-related questions were answered by performing web-based searches on a computer, limited to 15 minutes per task; answers were coded by 2 researchers for response accuracy (0 or 1) and specificity (0, 1, or 2) to give a score ranging from 0 to 18Diet and nutrition guidelines, skin cancer, alternative medicine, vaccines, assistive health technology, and over-the-counter genetic testingConstruct validity was demonstrated in that having a college degree and daily internet use were positively associated with more successful health information searches, and the oldest age group had lower success scores compared with younger participants; criterion validity was demonstrated in that higher health literacy was positively associated with success on some search tasksPartially

Blakemore et al [34], 2020eHealth literacyA massive open web-based course run 8 timesParticipants responded to 1 health-related question and were asked to list the resources they used to inform that answer; answers were coded according to the extent that quality resources were used in this questionEpigenetics and cancerEcological validity was demonstrated via having participants access web-based resources to inform answers to a health-related questionYes

Chang et al [35], 2021Searching performance11 older adult participantsParticipants were asked to search for specific health information using a web browser on a computer; search completion time and problem correctness were measured by researchers during live observationVaccination for older adults, stroke, and anginaEcological validity was demonstrated via participants accessing real-world web-based health information to answer health-related questionsNo

Freund et al [36], 2017eHealth literacy79 older adult participantsParticipants were asked to answer 6 questions (3 each for 2 health scenarios) while being given the option of using links to web-based medical databases with relevant informationHypertension, high blood pressure, osteoporosis, breast cancer, and prostate cancerConstruct validity was demonstrated in that test scores for the intervention group improved; ecological validity was demonstrated in that participants responded to questions by referencing a web-based resourceYes

Kordovski et al [37], 2020eHealth search skills56 undergraduate students enrolled in psychology coursesParticipants were instructed to find the answer to 5 short questions and 1 vignette-based question using an internet browser of their choice; participants’ accuracy, time to complete each task, and total number of search queries were recordedHeadaches, migraines, and Lyme diseaseCriterion validity was demonstrated in that long-answer accuracy was associated with better performance on a learning and memory composite test; construct validity was demonstrated in that lower performance on short questions was associated with lower maternal education and lower socioeconomic statusYes

Loda et al [38], 2020Information-seeking behavior140 medical studentsStudents were randomly assigned to use a specific search engine (Google, Medisuch, or free choice) and had 10 minutes to fill in a worksheet outlining a diagnostic recommendation; to pass, students needed to give at least 1 of 3 recommendations matching those of a clinical expertHistamine intoleranceNoneNo

Quinn et al [25], 2017eHealth literacy54 adultsParticipants were presented 6 health questions, and they used a browser to search for information to answer the questions; each answer was scored as correct or incorrect, with a final sum score out of 6Various topicsCriterion validity was demonstrated because the scores correlated with health literacyYes

Sharit et al [39], 2008Internet search task performance40 older adultsParticipants were assigned 6 search problems involving health-related information, for which they had to provide an answer using information they found on the internet; participants had 15 minutes to solve each problem, and the problems were progressively more complex; the problem solutions were scored as incorrect, partially correct, or correct by the researcher to create a task performance score; scores were weighted by difficulty, and participants’ completion times for each problem were also measured and factored into the score such that faster times indicated better performanceVariousCriterion validity was demonstrated in that higher performance correlated with higher knowledge of the internet, as well as with measures of reasoning, working memory, and perceptual speedYes

Sharit et al [40], 2015Search accuracy60 adultsParticipants were given a health scenario, followed by a series of questions related to it, and they could use the internet to find answers; to assess accuracy, researchers assigned a score for each question; questions were weighted based on their difficulty (differed in complexity and number of subtasks)Multiple sclerosisCriterion validity was demonstrated as search accuracy significantly correlated with reasoning, verbal ability, visuospatial ability, processing speed, and executive functionYes

van Deursen and van Dijk [41], 2011Internet skills performance88 adultsParticipants completed 9 health-related assignments using a computer with high-speed internet; assignment was deemed successfully completed if a correct answer was provided and deemed unsuccessful if no correct answer was provided in the given time frameVariousEcological validity was demonstrated via participants using unrestricted web-based searching to answer health-related questionsYes

van Deursen [42], 2012Internet skills performance88 adultsParticipants completed 9 health-related assignments using a computer with high-speed internet; assignment was deemed successfully completed if a correct answer was provided and deemed unsuccessful if no correct answer was provided in the given time frameVariousEcological validity was demonstrated via participants using unrestricted web-based searching to answer health-related questions; construct validity was demonstrated as education was predictive for making incorrect decisions based on the information foundYes
Simulated internet tasks

Camiling [43], 2019Actual eHealth literacy (distinct from perceived eHealth literacy)40 grade-10 students from public and private schoolsParticipants completed 10 simulation tasks; 2 researchers used a rubric to rate eHealth literacy based on task performanceNot specifiedEcological validity was demonstrated via use of simulated internet research tasks resembling a realistic environmentNo

Chan and Kaufman [44], 2011eHealth literacy20 adult participants aged between 18 and 65 yearsParticipants completed eHealth tasks while verbalizing their thoughts (think-aloud protocol); researchers observed their performance, rated accuracy, and denoted barriers based on video capture, audio recording, and notes taken during observationComparing hospital ratingsEcological validity was demonstrated via participants actively completing health-related internet tasks in a realistic environmentPartially

Maitz et al [22], 2020Health literacy (the authors note in their study that their understanding of health literacy includes “internet-based information literacy and reading literacy”)14 secondary school students aged 12 to 14 yearsParticipants were asked to provide health-related advice in response to a short narrative text; they were asked to take screenshots of all searches and web pages opened; the web pages were later classified by researchers as good, fair, poor, or badRhinoplasty and skin cancerNoneYes

Neter and Brainin [24], 2017eHealth literacy88 older adultsParticipants completed 15 computerized simulation tasks assessing digital and health literacy skills; tasks were rated as completed or not completed by the researcher upon reviewing the recorded performance; time needed to perform the task was also recorded; 2 researchers provided overall observational judgment on participants’ performance, ranging from 1 (poor) to 5 (good); a third researcher evaluated whether disagreements were presentVarious topicsConstruct validity was demonstrated because lower performers had significantly fewer years of experience using the internetYes

van der Vaart et al [26], 2011eHealth literacy88 adultsParticipants completed 9 health-related assignments using a computer with high-speed internet; assignment was deemed successfully completed if a correct answer was provided and deemed unsuccessful if no correct answer was provided in the given time frameVariousEcological validity was demonstrated via participants using unrestricted web-based searching to answer health-related questionsYes

van der Vaart et al [45], 2013eHealth literacy31 adult patientsIn study 1, participants could use the internet freely to complete 6 health-related assignments; in study 2, participants used specific websites to complete 5 health-related assignments; researchers coded whether the assignment was completed and whether help was needed; in addition, the time needed to perform each assignment was recorded; the performance was ultimately scored as good, reasonable, or poor according to the skills participants used to execute the assignmentVariousConstruct validity was demonstrated through correlations of higher performance with higher educationYes

Witry et al [46], 2018eHealth task performance100 adult patients with COPDaParticipants completed a series of timed eHealth simulation exercises using a laptop computer and 2 different tablet devices; the time taken to complete each task was measured and used to indicate task performance, with faster times indicating better performanceCOPDConstruct validity was demonstrated because those who reported using video chat took less time than nonusers to complete most of the tasksNo
Website evaluation tasks

Kalichman et al [47], 2006Health information evaluation skills448 adults who used the internet <3 times in the month before screeningParticipants rated 2 preselected web pages—1 from a medical association and 1 with scientifically unsupported claims—on 5 dimensions of website quality; a larger difference in scores indicated higher health information evaluation skillsHIV and AIDS treatmentConstruct validity was demonstrated in that those receiving internet skills training had better discriminationYes

Mitsuhashi [48], 2018eHealth literacy evaluation skills300 adult participantsParticipants were shown a search engine results page with 5 websites and asked which should be viewed first; the list included 2 commercial websites, 2 personal health care websites, and 1 government website; participants choosing the government website were assigned 1 point, others were assigned 0 pointsNot specifiedConstruct validity was demonstrated in that evaluation skills improved significantly in an e-learning intervention group compared with the control groupNo

Schulz et al [49], 2021Health literacy362 adultsParticipants rated 2 health information websites (one was of high quality, whereas the other was of low quality) using 3 seven-step semantic differential scales; in addition, participants were asked to choose beneficial depression treatments from a list of relevant and nonrelevant treatmentsDepression treatmentCriterion validity was demonstrated in that those with high health literacy and accurate recognition of the low-quality website demonstrated good judgment for depression treatmentPartially

Trettin et al [50], 2008Website evaluation142 high school studentsTwo measures: first, a brief 2-item pretest of knowledge about how to evaluate a website, and second, participants ranked 2 different websites (assigned to them from a list of 12 websites) using 6 credibility factors, with scores ranging from 1 (very bad) to 5 (very good)Not specifiedEcological validity was demonstrated in that participants ranked authentic health-related websites according to their credibilityYes

Xie [51], 2011eHealth literacy124 older adultsParticipants were asked to evaluate the quality of 20 health websites: 10 selected from the Medical Library Association’s recommended sites and 10 from a commercial search engine; each correct assessment received 1 point, whereas incorrect or uncertain assessments received 0 pointsNot specifiedConstruct validity was demonstrated because scores improved after an educational interventionNo
Knowledge of theweb-basedhealth information–seeking process

Hanik and Stellefson [52], 2011eHealth literacy77 undergraduate health education majorsResearchers used the RRSA-hb, which is a questionnaire that tests participants’ declarative knowledge of concepts, skills, and thinking strategies related to using the internet to find health informationVariousEcological validity was demonstrated in the study by Ivanitskaya et al [53]No

Hanna et al [54], 2017eHealth literacy165 adult dental patientsParticipants were asked to circle the web-based health information quality seals they recognized and report the purpose of 1 circled figureThird molar knowledgeCriterion validity was demonstrated in that the eHEALSc scores correlated with the dental procedural web-based information–seeking measure; construct validity was demonstrated in that web-based dental procedural information seeking was significantly associated with educational attainment and dental decisional control preferenceYes

Ivanitskaya et al [53], 2006Health information competency400 college-aged studentsResearchers used the RRSA-h, which is a web-based quiz that assesses declarative and procedural knowledge related to web-based health information seekingVariousEcological validity was demonstrated in that some questions had participants access real health-related websites to assess their credibilityNo

Ivanitskaya et al [55], 2010eHealth literacy skills1914 undergraduate and graduate students enrolled in health-related coursesResearchers used the RRSA-h, which is a web-based quiz that assesses declarative and procedural knowledge related to web-based health information seeking; a proxy measure of critical judgment skills related to pharmacies was also includedPharmaceuticals and various othersConstruct validity was demonstrated in that evaluation skills positively correlated with the number of earned college credits and being enrolled in a health-related majorPartially

St. Jean et al [56], 2017Digital health literacy19 adolescentsParticipants were given 13 questions related to searching for health information; researchers analyzed responses using thematic analysis; no evident scoring system usedType 1 diabetesNoneYes

van der Vaart and Drossaert [57], 2017Digital health literacy, eHealth literacy200 adultsParticipants completed a 28-item questionnaire: 21 are self-report items, whereas 7 are performance-based items for each of which there is a correct answerVariousNone for performance-based itemsYes
Health-related knowledge

Holt et al [58], 2019eHealth literacy246 adult patients with diabetes, other endocrine conditions, and gastrointestinal diseasesResearchers used eHLAd performance tests (tools 1 and 4); tool 1 is a performance-based health literacy test based on an information leaflet, and tool 4 is a performance test for knowledge of health and health careVariousConstruct validity was demonstrated in that educational level was positively correlated with tool 4Partially

Holt et al [59], 2020eHealth literacy366 nursing studentsResearchers used eHLA performance tests (tools 1 and 4); tool 1 is a performance-based health literacy test based on an information leaflet, and tool 4 is a performance test for knowledge of health and health careVariousConstruct validity was demonstrated in that graduate-level students scored higher than entry-level students, and performance on tools 1 and 4 was correlated with having at least 1 parent with experience in the social or health care systemPartially

Karnoe et al [31], 2018eHealth literacy475 adults used as a validation sampleResearchers used the eHLA, which consists of 7 tools, 2 of which are objective measures: tool 1 is a performance-based health literacy test based on an information leaflet, and tool 4 is a performance test for knowledge of health and health careVariousNonePartially

Liu et al [60], 2020Digital health literacy1588 adult participantsParticipants were provided 5 randomly selected items from a large web-based health information bank and asked whether the information was right or wrong; 2 items were designed to be relatively easy to judge accurately, 2 were moderately easy, and 1 was difficult; participants scored 1 for each accurate judgment and 0 for being incorrect or unsureVariousEcological validity was demonstrated in that the web-based health information bank was generated from real web-based sources; construct validity was demonstrated because participants at high risk for misjudging health information had lower education level, poorer health, and used the internet less Partially

aCOPD: chronic obstructive pulmonary disease.

bRRSA-h: Research Readiness Self-Assessment-health.

ceHEALS: eHealth Literacy Scale.

deHLA: eHealth literacy assessment toolkit.

Answering Health-Related Questions Using the Internet

One of the most common types of performance-based measurements identified in this scoping review involved participants answering health-related questions using the internet; these were questions for which responses could be scored according to correctness, completeness, and specificity. In the majority of the studies using this assessment method participants were allowed open access to the internet to inform their answers. An exception was the study by Freund et al [36] in which participants were provided expert-checked links to helpful resources that they could choose to use for answering health-related questions. In addition to the correctness of participants’ responses, some researchers factored completion time into participant scores such that faster completion time indicated greater proficiency [39,40]. Chang et al [35] similarly considered faster completion times indicative of better web-based search performance, provided the responses were correct. The health questions posed in these measurement tools ranged from encompassing several diverse topics to being focused on 1 topic in depth; for example, both Agree et al [33] and Quinn et al [25] asked participants about 6 diverse health topics, whereas Kordovski et al [37] and Loda et al [38] asked participants to perform 1 in-depth medical diagnosis relating to symptoms of Lyme disease and histamine intolerance, respectively.

An advantage of eHealth literacy measurements where participants use the internet to answer health-related questions is that participants can be assessed without the need for researcher observation and video recording, except in cases where completion time is additionally factored into the scores. With these measurement tools, participants’ responses to questions were graded by ≥1 researchers according to a predetermined rubric. The majority of measurement tools in this category coded responses as simply correct (1 point) or incorrect (0 points); however, Agree et al [33] also examined specificity (on a scale ranging from 0 to 2 points) and Kordovski et al [37] gave partial credit (1 out of 2 points) for coming close to the correct response by diagnosing something similar to Lyme disease. In contrast to considering response correctness, Blakemore et al [34] gauged participants’ eHealth literacy skills based on whether they included a “detailed list of resources,” had “written about using resources,” or had no reference to web-based health resources in their response to a health-related question.

Simulated Internet Tasks

Another common type of performance-based eHealth literacy assessment identified in this scoping review involved web-based information–seeking simulations, where participants’ web-based behavior was observed (either in real time or via video recording) and assessed for proficiency by researchers. Proficiency was determined by assessing whether participants were able to complete a set of tasks; many researchers went further by also assessing the degree of efficiency and correctness that the participants exemplified throughout each simulated task; for example, in the study by van der Vaart et al [26], participants’ task performance was simply coded as successful or unsuccessful, depending on whether they accomplished the end result of the task (eg, adding a specified website as a bookmark or downloading a specific image), whereas in the study conducted by Camiling [43], a rubric was developed and used to rate participants’ proficiency while researchers observed them completing 10 health-related tasks on an internet-connected computer. Similarly, Neter and Brainin [24] viewed recordings of participants completing 15 simulation tasks on an internet-connected computer and rated each task on whether it was completed and additionally on the quality of task performance on a scale ranging from 1 (poor) to 5 (good). Video and audio recordings were commonly used to record participants’ actions and thoughts in simulation-based measurement tools, with the notable exception of Maitz et al [22] who had participants record screenshots of the websites they visited.

eHealth literacy assessments based on simulated internet tasks tended to be the most time-consuming for participants: the 15 simulation tasks in the study by Neter and Brainin [24] took approximately 90 minutes for each participant to complete, and the 5 or 6 simulation tasks in the study by van der Vaart et al [45] took an approximate median time of 30 minutes for each participant to complete. A notable exception is the study by Witry et al [46] in which the assigned eHealth simulation tasks were far simpler and could be completed in 1 minute; however, it should be noted that participants’ completion time was the only metric used to gauge participants’ eHealth literacy in this study. It also stands to reason that evaluating the proficiency of participants’ recorded web-based behavior is quite time-consuming for researchers compared with assessment methods that are scored based on correct or incorrect responses to questions. In the studies conducted by Chan and Kaufman [44] as well as van der Vaart et al [26], researchers factored in the correctness of responses in addition to researcher-observed performance to gauge participants’ eHealth literacy, combining task simulation with the Health-related questions using the internet methods discussed previously.

Website Evaluation Tasks

A third prominent type of performance-based eHealth literacy assessment identified in this scoping review involved participants evaluating the quality of health-related websites. In the studies conducted by Kalichman et al [47], Schulz et al [49], and Trettin et al [50], participants rated 2 different websites on 5, 7, and 6 dimensions related to their credibility, respectively. In each of these studies, the authors presented participants with a website of high quality and another of low quality, and participants’ eHealth literacy was determined based on whether they rated the websites accordingly (with a larger difference between high- and low-quality website ratings indicating greater proficiency).

Deviating from the structure of these studies but still related to evaluating website quality, Mitsuhashi [48] had participants select which website they should view first from a search results page with 5 options (2 commercial websites, 2 personal health care websites, and 1 government website), in which participants who selected the government website were said to have proficient evaluation skills. Xie [51] collected 20 websites to present to participants—10 from a commercial search engine and 10 from a medical association’s recommended websites—and had participants assess each website as high or low quality. Correct assessments earned participants 1 point, and incorrect or uncertain assessments earned 0 points, for a maximum score of 20.

The rating of the quality of real health-related websites by participants significant ecological validity in terms of participants’ knowledge of how to recognize signifiers of credible health information on the web; however, in contrast to the simulated internet task measurement tools, these tools do not require participants to form queries or extract information to apply to health-related problems. Although evaluating the credibility of a web-based source is undoubtedly a pivotal component of eHealth literacy, it does not encompass all components of eHealth literacy as defined by Norman and Skinner [7].

Knowledge of the Web-Based Health Information–Seeking Process

Distinct from measurement tools where participants demonstrate eHealth literacy skills through task completion or by correctly answering health-related questions using the internet, other tools tested participants on their knowledge of procedural internet skills and information-seeking strategies to gauge their eHealth literacy. The most used tool in this category is the RRSA-h [53,55]. The RRSA-h uses multiple choice and true or false questions to assess participants’ declarative knowledge of web-based health information seeking, as well as some participative problem-based exercises to assess elements of their procedural knowledge. Ivanitskaya et al [53,55] as well as Hanik and Stellefson [52] used this measure to assess eHealth literacy skills in undergraduate and graduate students. The RRSA-h takes approximately 30 minutes on average to complete.

In contrast to this measure that takes relatively long to complete, Hanna et al [54] used just 1 item to gauge participants’ ability to recognize high-quality web-based information by asking them to circle web-based health information quality seals that they recognized from a set of 19 images. In addition, participants were asked to explain the purpose of one of the images they circled. Researchers analyzed whether participants were able to identify ≥1 real quality seals, as well as whether they could identify that the image was used to signify credibility. van der Vaart and Drossaert [57] used 7 performance-based multiple-choice quiz items to gauge 7 dimensions of participants’ eHealth literacy skills, which were added to supplement a subjective measure of eHealth literacy. Finally, in this category, St. Jean et al [56] devised a written assignment where adolescents advised a fictional peer about using the internet to find information about type 1 diabetes through 13 open-ended questions. The researchers did not use an evident scoring system to rate the eHealth literacy skills of this population; rather, they applied thematic analysis to characterize the eHealth literacy skills of the participant group as a whole.

Quizzing participants on their knowledge related to eHealth literacy does not require recording or observing participants’ web-based activity, and nor does it require necessarily providing participants with internet access, which makes this method more accessible for some research settings in terms of resources and complexity than other tools mentioned previously. As participants are not actively carrying out web-based health information–seeking tasks with these measurement tools, the tools may not carry the same degree of ecological validity as those mentioned previously. However, basing eHealth literacy assessment on participants’ knowledge of proficient web-based information–seeking skills still avoids many of the pitfalls related to subjective self-report measures of eHealth literacy.

Health-Related Knowledge

We identified 2 measurement tools that examined participants’ baseline knowledge of health and their ability to apply that knowledge as an indicator of eHealth literacy. The most prominent tool is the eHLA [31], which is composed of 7 distinct measurement tools. Tools 1 and 4 within this set are performance-based measures that assess functional health literacy and participant knowledge of health and disease, respectively. The performance-based aspects of this tool (as well as its subjective tools) were used by Holt et al [58,59] to measure eHealth literacy in medical outpatients and nursing students.

Liu et al [60] also created a performance-based eHealth literacy measurement tool that assesses participants’ health knowledge. In their study, the authors generated a 310-item bank of examples of web-based health information, which included items labeled “easy,” “moderate,” and “difficult.” Participants were randomly assigned 5 items (2 easy, 2 moderately easy, and 1 difficult) and asked to rate the information as correct, incorrect, or unsure. Participants were given 1 point for accurately identifying information as correct or incorrect and 0 points otherwise for a score out of 5 representing their eHealth literacy.

In these 2 eHealth literacy measurement tools, there are no components that directly assess computer literacy in a performance-based manner nor do they contain performance-based components related to actively seeking health information (eg, forming a query or selecting a source). We have included them in this scoping review because the authors themselves define these as measures of eHealth literacy (and it should be noted that the eHLA includes subjective measurements that touch upon computer-related components of eHealth literacy). However, judging their performance-based components solely by the definition of eHealth literacy proposed by Norman and Skinner [7], these measures may be more accurately characterized as partial measures of the construct.

Prevalence of Performance-Based eHealth Literacy Measurement Tools

Of the 313 studies included in this scoping review, 33 (10.5%) used a performance-based measurement tool, and 280 (89.5%) used measures of eHealth literacy that only incorporated subjective or self-rating aspects. Furthermore, it is notable that 210 (67.1%) of the 313 studies reported using eHEALS in either its original form or translated into another language. Our findings indicate that current research on eHealth literacy has a strong tendency to rely on self-rated perceptions of eHealth literacy rather than assessing actual ability.


Principal Findings

The primary purpose of this scoping review was to identify and describe tools that measure eHealth literacy based on objective performance (as opposed to subjective self-rating). We identified 29 such measurement tools, of which only 2 had been used in >1 peer-reviewed study as of the date of our search. These measurement tools were the RRSA-h [55], which is a web-based quiz measuring declarative and procedural knowledge related to web-based health information seeking, and the eHLA [31], which assesses eHealth literacy through 7 distinct tools, of which 2 are performance based. It is noteworthy that the 2 performance-based tools within the eHLA do not touch upon computer- or internet-specific skills or knowledge; as such, not all aspects of eHealth literacy are measured in a performance-based fashion. The same critique may be applied to the measurement tool created by Liu et al [60] who similarly do not directly address computer literacy or media literacy in their performance-based assessment tool.

The second purpose of this scoping review was to characterize the prevalence of performance-based eHealth literacy measurement tools in the literature in contrast to subjective measurement tools. Our findings indicate that the vast majority of current research on eHealth literacy relies on self-rated perceptions of eHealth literacy rather than assessing actual ability. This is concerning, considering the limited utility of this short self-report measure for predicting performed eHealth literacy as indicated by a substantial body of literature [21,22,24-26]. If researchers are interested in gauging participants’ true ability to locate, evaluate, and use web-based health information, more efforts should be made to incorporate performance-based measurement tools of eHealth literacy, such as those identified in this scoping review.

Another notable finding of this scoping review is that of the 29 unique eHealth literacy measurement tools with performance-based components identified, only 2 had been used in >1 peer-reviewed study at the time of this research. This comes in stark contrast to the prevalence of eHEALS, which was used by 210 (67.1%) of the 313 studies in various contexts and languages. This could be due in part to several of the studies (8/33, 24%) not including full versions of their performance-based instruments, making it challenging for other researchers to replicate these measures in other projects. Another implicit challenge to creating a performance-based eHealth literacy measure that may be adopted for widespread use is the changing state of scientific consensus on health-related topics, meaning that correct answers to health-related questions may need to be updated over time; for example, dietary guidelines have changed substantially in the past few decades because they have been updated based on our growing scientific knowledge [61,62]. In addition, health-related topics can differ substantially in relevancy or saliency among different populations, meaning a performance-based eHealth literacy measure effective for a particular population may not be as useful for another; for example, Loda et al [38] designed a performance-based measure for use by medical students where they needed to produce a histamine-intolerance diagnosis on par with a clinical expert using the internet for research, which is a task that is likely beyond the abilities of typical users without comparable existing medical knowledge.

Judging these performance-based measurement tools by the definition of eHealth literacy proposed by Norman and Skinner [7], the tools with the greatest perceived ecological validity include those having participants answer health-related questions using the internet and those having participants engage in simulated web-based health–related tasks. These measurement tools evaluate participants’ web-based health information–seeking abilities in settings similar to the ones they encounter when seeking information on their own computers. The time-consuming nature (for both participants and researchers) and the equipment needs of these measures present substantial barriers to their use. Efforts to streamline some of these measurement tools by including only strictly necessary components could assist with broader use. One example of a concise measure is offered by Witry et al [46], who created a set of simple eHealth simulation tasks that could be completed in <1 minute; however, it should be noted that this measure assesses one’s ability to navigate an eHealth platform more so than to actively seek health information on the web. While considering the challenges to using performance-based eHealth literacy measures, researchers should also weigh the major advantages of these tools in producing a more accurate depiction of performed eHealth literacy compared with short self-report measures.

One glaring absence across most performance-based measures of eHealth literacy is any mention of social media. In discussing the modern utility of eHEALS, Norman [63] noted that the shifting nature of the internet, with everyday users routinely contributing information on the web as opposed to only accessing it, necessitates modifications to existing measurement tools to maintain validity. Consumers are more frequently turning to social media platforms with health-related queries, where they stand to gain social and emotional support from peers while gaining crowdsourced wisdom related to their health problem [64]. The validity and trustworthiness of information quality on these platforms has been flagged as a major concern [65]; however, public health organizations have also used these engaging platforms to distribute high-quality and up-to-date information [66,67]. As an acknowledgment that users are able to gain useful and accessible health information using social media, more eHealth literacy measurement tools should incorporate these platforms into their assessment. One example of an assessment tool with a social web-based component is provided in the study by van der Vaart et al [45], where participants were asked to demonstrate interacting with a health care rating website and peer support forum, for which their proficiency was assessed based on independent task completion and overall task performance. Given the prominence of social media in today’s web-based informational landscape, future performance-based measures of eHealth literacy should consider assessing participants’ ability to identify credible health information on social media platforms.

Limitations

This study includes a few notable limitations. This scoping review only considered peer-reviewed journal articles published in English, which may have caused bias in our findings. Furthermore, it is possible that we missed including studies that evaluated skills under the umbrella of eHealth literacy but did not describe this using any of the terminology in our search protocol. We did locate studies using terms such as “health information evaluation skills” [47] and “information-seeking behavior” [38] through our search protocol, indicating some ability to detect studies deviating from our search terms. Finally, it should be noted that the evidence of validity provided in Table 1 provides only a limited surface-level judgment of 3 types of validity (criterion, construct, and ecological validity) based on the evidence presented in each article. It could very well be the case that the authors of the included studies have additional evidence of validity that we did not recognize as falling within these types of validities or that did not make it into their published articles.

Conclusions

This is the first literature review that specifically identifies objective performance-based measurement tools of eHealth literacy in contrast to subjective self-report measurement tools. We identified 29 unique measurement tools of eHealth literacy with performance-based components used in various populations and covering various health-related topics. To better establish the utility and validity of performance-based measurement tools of eHealth literacy, scholars looking to incorporate performance-based measurement of eHealth literacy into their research should look to use and build from some of these existing measurement techniques rather than producing their own. In contexts where research participants can be provided a computer with internet access, measures from the Health-related questions using the internet, Simulated internet tasks, or Website evaluation tasks categories of this scoping review may offer ecologically valid options for assessing eHealth literacy. In contexts where providing such equipment may not be feasible, measures from the Knowledge of the web-based health information–seeking process category may offer a simpler performance-based eHealth literacy assessment tool that still reduces strict reliance on participants’ ability to gauge their own skill level.

In addition, to the best of our knowledge, this is the first literature review that quantifies the approximate prevalence of performance-based measurement versus subjective measurement of eHealth literacy in the literature broadly. In peer-reviewed scholarship, eHealth literacy assessment is predominantly conducted using subjective self-report measurement techniques. Performance-based measures of eHealth literacy are likely to provide a far better picture of how proficiently people find, evaluate, and use web-based health information. As such, researchers should consider using more objective measures of eHealth literacy such as those identified in this scoping review.

Conflicts of Interest

None declared.

  1. Fishbein M, Ajzen I. Predicting and Changing Behavior The Reasoned Action Approach. Milton Park, Oxfordshire. Taylor & Francis; 2011.
  2. Gualtieri L. The doctor as the second opinion and the internet as the first. In: Proceedings of the CHI '09 Extended Abstracts on Human Factors in Computing Systems. Presented at: CHI '09: CHI Conference on Human Factors in Computing Systems; Apr 4 - 9, 2009, 2009; Boston MA USA. [CrossRef]
  3. Sood N, Jimenez DE, Pham TB, Cordrey K, Awadalla N, Milanaik R. Paging Dr. Google: the effect of online health information on trust in pediatricians' diagnoses. Clin Pediatr (Phila). Jul 29, 2019;58(8):889-896. [CrossRef] [Medline]
  4. Lee ST, Lin J. The influence of offline and online intrinsic motivations on online health information seeking. Health Commun. Aug 2020;35(9):1129-1136. [CrossRef] [Medline]
  5. Kitchens B, Harle CA, Li S. Quality of health-related online search results. Decis Support Syst. Jan 2014;57(1):454-462. [CrossRef]
  6. Rachul C, Marcon AR, Collins B, Caulfield T. COVID-19 and 'immune boosting' on the internet: a content analysis of Google search results. BMJ Open. Oct 26, 2020;10(10):e040989. [FREE Full text] [CrossRef] [Medline]
  7. Norman CD, Skinner HA. eHealth literacy: essential skills for consumer health in a networked world. J Med Internet Res. Jun 16, 2006;8(2):e9. [FREE Full text] [CrossRef] [Medline]
  8. Norgaard O, Furstrand D, Klokker L, Karnoe A, Batterham R, Kayser L, et al. The e-health literacy framework: a conceptual framework for characterizing e-health users and their interaction with e-health systems. Knowl Manag E-Learning. 2015;7(4):522-540. [CrossRef]
  9. Norman CD, Skinner HA. eHEALS: the eHealth literacy scale. J Med Internet Res. Nov 14, 2006;8(4):e27. [FREE Full text] [CrossRef] [Medline]
  10. Karnoe A, Kayser L. How is eHealth literacy measured and what do the measurements tell us? A systematic review. Knowl Manag E-Learning. 2015;7(4):576-600. [CrossRef]
  11. Griebel L, Enwald H, Gilstad H, Pohl A, Moreland J, Sedlmayr M. eHealth literacy research-Quo vadis? Informatics Heal Soc Care. Dec 2018;43(4):427-442. [CrossRef] [Medline]
  12. Mabe PA, West SG. Validity of self-evaluation of ability: a review and meta-analysis. J Appl Psychol. Jun 1982;67(3):280-296. [CrossRef]
  13. Dunning D, Heath C, Suls JM. Flawed self-assessment: implications for health, education, and the workplace. Psychol Sci Public Interes Suppl. Dec 01, 2004;5(3):69-106. [CrossRef] [Medline]
  14. Dunning D. Self-Insight Roadblocks and Detours on the Path to Knowing Thyself. Milton Park, Oxfordshire. Taylor & Francis; 2005.
  15. Hussain A, Ali S, Ahmed M, Hussain S. The anti-vaccination movement: a regression in modern medicine. Cureus. Jul 03, 2018;10(7):e2919. [FREE Full text] [CrossRef] [Medline]
  16. Kata A. Anti-vaccine activists, Web 2.0, and the postmodern paradigm--An overview of tactics and tropes used online by the anti-vaccination movement. Vaccine. May 28, 2012;30(25):3778-3789. [CrossRef] [Medline]
  17. Merritt K, Smith KD, Di Renzo, Jr JC. An investigation of self-reported computer literacy: is it reliable? Issues Inf Syst. 2005;6(1):289-295. [CrossRef]
  18. Palczyńska M, Rynko M. ICT skills measurement in social surveys: can we trust self-reports? Qual Quant. Aug 29, 2021;55(3):917-943. [CrossRef]
  19. Mahmood K. Do people overestimate their information literacy skills? A systematic review of empirical evidence on the Dunning-Kruger effect. Commun Inf Lit. 2016;10(2):198-213. [CrossRef]
  20. Eysenbach G, Köhler C. How do consumers search for and appraise health information on the world wide web? Qualitative study using focus groups, usability tests, and in-depth interviews. BMJ. Mar 09, 2002;324(7337):573-577. [FREE Full text] [CrossRef] [Medline]
  21. Stellefson M, Hanik B, Chaney JD, Tennant B. Analysis of eHealth search perspectives among female college students in the health professions using Q methodology. J Med Internet Res. Apr 27, 2012;14(2):e60. [FREE Full text] [CrossRef] [Medline]
  22. Maitz E, Maitz K, Sendlhofer G, Wolfsberger C, Mautner S, Kamolz L, et al. Internet-based health information-seeking behavior of students aged 12 to 14 years: mixed methods study. J Med Internet Res. May 26, 2020;22(5):e16281. [FREE Full text] [CrossRef] [Medline]
  23. Brown CA, Dickson R. Healthcare students' e-literacy skills. J Allied Health. 2010;39(3):179-184. [Medline]
  24. Neter E, Brainin E. Perceived and performed eHealth literacy: survey and simulated performance test. JMIR Hum Factors. Jan 17, 2017;4(1):e2. [FREE Full text] [CrossRef] [Medline]
  25. Quinn S, Bond R, Nugent C. Quantifying health literacy and eHealth literacy using existing instruments and browser-based software for tracking online health information seeking behavior. Comput Human Behav. Apr 2017;69:256-267. [CrossRef]
  26. van der Vaart R, van Deursen AJ, Drossaert CH, Taal E, van Dijk JA, van de Laar MA. Does the eHealth Literacy Scale (eHEALS) measure what it intends to measure? Validation of a Dutch version of the eHEALS in two adult populations. J Med Internet Res. Nov 09, 2011;13(4):e86. [FREE Full text] [CrossRef] [Medline]
  27. Lee J, Lee E, Chae D. eHealth literacy instruments: systematic review of measurement properties. J Med Internet Res. Nov 15, 2021;23(11):e30644. [FREE Full text] [CrossRef] [Medline]
  28. Munn Z, Peters MD, Stern C, Tufanaru C, McArthur A, Aromataris E. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. Nov 19, 2018;18(1):143. [FREE Full text] [CrossRef] [Medline]
  29. Tricco AC, Lillie E, Zarin W, O'Brien KK, Colquhoun H, Levac D, et al. PRISMA extension for scoping reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. Oct 02, 2018;169(7):467-473. [FREE Full text] [CrossRef] [Medline]
  30. Encyclopedia of Research Design. Thousand Oaks, California, United states. SAGE Publications; 2010.
  31. Karnoe A, Furstrand D, Christensen KB, Norgaard O, Kayser L. Assessing competencies needed to engage with digital health services: development of the eHealth literacy assessment toolkit. J Med Internet Res. May 10, 2018;20(5):e178. [FREE Full text] [CrossRef] [Medline]
  32. Ivanitskaya L, Laus R, Casey AM. Research readiness self-assessment. J Library Administration. Jan 20, 2004;41(1-2):167-183. [CrossRef]
  33. Agree EM, King AC, Castro CM, Wiley A, Borzekowski DL. "It's got to be on this page": age and cognitive style in a study of online health information seeking. J Med Internet Res. Mar 24, 2015;17(3):e79. [FREE Full text] [CrossRef] [Medline]
  34. Blakemore LM, Meek SE, Marks LK. Equipping learners to evaluate online health care resources: longitudinal study of learning design strategies in a health care massive open online course. J Med Internet Res. Feb 26, 2020;22(2):e15177. [FREE Full text] [CrossRef] [Medline]
  35. Chang SJ, Yang E, Lee K, Ryu H. Internet health information education for older adults: a pilot study. Geriatr Nurs. Mar 2021;42(2):533-539. [CrossRef] [Medline]
  36. Freund O, Reychav I, McHaney R, Goland E, Azuri J. The ability of older adults to use customized online medical databases to improve their health-related knowledge. Int J Med Inform. Jun 2017;102:1-11. [CrossRef] [Medline]
  37. Kordovski VM, Babicz MA, Ulrich N, Woods SP. Neurocognitive correlates of internet search skills for eHealth fact and symptom information in a young adult sample. Percept Mot Skills. Oct 01, 2020;127(5):960-979. [CrossRef] [Medline]
  38. Loda T, Erschens R, Junne F, Stengel A, Zipfel S, Herrmann-Werner A. Correction: undergraduate medical students' search for health information online: explanatory cross-sectional study. JMIR Med Inform. Aug 11, 2020;8(8):e23253. [FREE Full text] [CrossRef] [Medline]
  39. Sharit J, Hernández MA, Czaja SJ, Pirolli P. Investigating the roles of knowledge and cognitive abilities in older adult information seeking on the web. ACM Trans Comput Hum Interact. May 2008;15(1):3-25. [FREE Full text] [CrossRef] [Medline]
  40. Sharit J, Taha J, Berkowsky RW, Profita H, Czaja SJ. Online information search performance and search strategies in a health problem-solving scenario. J Cogn Eng Decis Mak. May 21, 2015;9(3):211-228. [FREE Full text] [CrossRef] [Medline]
  41. van Deursen AJ, van Dijk JA. Internet skills performance tests: are people ready for eHealth? J Med Internet Res. Apr 29, 2011;13(2):e35. [FREE Full text] [CrossRef] [Medline]
  42. van Deursen A. Internet skill-related problems in accessing online health information. Int J Med Inform. Jan 2012;81(1):61-72. [CrossRef] [Medline]
  43. Camiling MK. eHealth literacy of high school students in the Philippines. IAFOR J Educ. Dec 01, 2019;7(2):69-87. [CrossRef]
  44. Chan CV, Kaufman DR. A framework for characterizing eHealth literacy demands and barriers. J Med Internet Res. Nov 17, 2011;13(4):e94. [FREE Full text] [CrossRef] [Medline]
  45. van der Vaart R, Drossaert CH, de Heus M, Taal E, van de Laar M. Measuring actual eHealth literacy among patients with rheumatic diseases: a qualitative analysis of problems encountered using Health 1.0 and Health 2.0 applications. J Med Internet Res. Feb 11, 2013;15(2):e27. [FREE Full text] [CrossRef] [Medline]
  46. Witry M, Comellas A, Simmering J, Polgreen P. The association between technology use and health status in a chronic obstructive pulmonary disease cohort: multi-method study. J Med Internet Res. Apr 02, 2018;20(4):e125. [FREE Full text] [CrossRef] [Medline]
  47. Kalichman SC, Cherry C, Cain D, Pope H, Kalichman M, Eaton L, et al. Internet-based health information consumer skills intervention for people living with HIV/AIDS. J Consult Clin Psychol. Jun 2006;74(3):545-554. [CrossRef]
  48. Mitsuhashi T. Effects of two-week e-learning on eHealth literacy: a randomized controlled trial of Japanese Internet users. PeerJ. 2018;6:e5251. [FREE Full text] [CrossRef] [Medline]
  49. Schulz PJ, Pessina A, Hartung U, Petrocchi S. Effects of objective and subjective health literacy on patients' accurate judgment of health information and decision-making ability: survey study. J Med Internet Res. Jan 21, 2021;23(1):e20457. [FREE Full text] [CrossRef] [Medline]
  50. Trettin LD, May JC, McKeehan NC. Teaching teens to "Get Net Smart for Good Health": comparing interventions for an Internet training program. J Med Libr Assoc. Oct 2008;96(4):370-374. [FREE Full text] [CrossRef] [Medline]
  51. Xie B. Experimenting on the impact of learning methods and information presentation channels on older adults' e-health literacy. J Am Soc Inf Sci. Jun 06, 2011;62(9):1797-1807. [CrossRef]
  52. Hanik B, Stellefson M. E-health literacy competencies among undergraduate health education students: a preliminary study. Int Electron J Health Educ. 2011;14:46-58.
  53. Ivanitskaya L, O'Boyle I, Casey AM. Health information literacy and competencies of information age students: results from the interactive online Research Readiness Self-Assessment (RRSA). J Med Internet Res. Apr 21, 2006;8(2):e6. [FREE Full text] [CrossRef] [Medline]
  54. Hanna K, Sambrook P, Armfield J, Brennan D. Internet use, online information seeking and knowledge among third molar patients attending public dental services. Aust Dent J. Sep 31, 2017;62(3):323-330. [FREE Full text] [CrossRef] [Medline]
  55. Ivanitskaya L, Brookins-Fisher J, O'Boyle I, Vibbert D, Erofeev D, Fulton L. Dirt cheap and without prescription: how susceptible are young US consumers to purchasing drugs from rogue internet pharmacies? J Med Internet Res. Apr 26, 2010;12(2):e11. [FREE Full text] [CrossRef] [Medline]
  56. St. Jean B, Greene Taylor N, Kodama C, Subramaniam M. Assessing the digital health literacy skills of tween participants in a school-library-based after-school program. J Consum Health Internet. Mar 10, 2017;21(1):40-61. [CrossRef]
  57. van der Vaart R, Drossaert C. Development of the digital health literacy instrument: measuring a broad spectrum of health 1.0 and health 2.0 skills. J Med Internet Res. Jan 24, 2017;19(1):e27. [FREE Full text] [CrossRef] [Medline]
  58. Holt KA, Karnoe A, Overgaard D, Nielsen SE, Kayser L, Røder ME, et al. Differences in the level of electronic health literacy between users and nonusers of digital health services: an exploratory survey of a group of medical outpatients. Interact J Med Res. Apr 05, 2019;8(2):e8423. [FREE Full text] [CrossRef] [Medline]
  59. Holt KA, Overgaard D, Engel LV, Kayser L. Health literacy, digital literacy and eHealth literacy in Danish nursing students at entry and graduate level: a cross sectional study. BMC Nurs. Apr 10, 2020;19(1):22. [FREE Full text] [CrossRef] [Medline]
  60. Liu P, Yeh L, Wang J, Lee S. Relationship between levels of digital health literacy based on the Taiwan digital health literacy assessment and accurate assessment of online health information: cross-sectional questionnaire study. J Med Internet Res. Dec 21, 2020;22(12):e19767. [FREE Full text] [CrossRef] [Medline]
  61. DeSalvo KB, Olson R, Casavale KO. Dietary guidelines for Americans. JAMA. Feb 02, 2016;315(5):457-458. [CrossRef] [Medline]
  62. Herforth A, Arimond M, Álvarez-Sánchez C, Coates J, Christianson K, Muehlhoff E. A global review of food-based dietary guidelines. Adv Nutr. Jul 01, 2019;10(4):590-605. [FREE Full text] [CrossRef] [Medline]
  63. Norman C. eHealth literacy 2.0: problems and opportunities with an evolving concept. J Med Internet Res. Dec 23, 2011;13(4):e125. [FREE Full text] [CrossRef] [Medline]
  64. Zhao Y, Zhang J. Consumer health information seeking in social media: a literature review. Health Info Libr J. Dec 2017;34(4):268-283. [FREE Full text] [CrossRef] [Medline]
  65. De Martino I, D'Apolito R, McLawhorn AS, Fehring KA, Sculco PK, Gasparini G. Social media for patients: benefits and drawbacks. Curr Rev Musculoskelet Med. Mar 21, 2017;10(1):141-145. [FREE Full text] [CrossRef] [Medline]
  66. Merchant RM, South EC, Lurie N. Public health messaging in an era of social media. JAMA. Jan 19, 2021;325(3):223-224. [CrossRef] [Medline]
  67. Moorhead SA, Hazlett DE, Harrison L, Carroll JK, Irwin A, Hoving C. A new dimension of health care: systematic review of the uses, benefits, and limitations of social media for health communication. J Med Internet Res. Apr 23, 2013;15(4):e85. [FREE Full text] [CrossRef] [Medline]


eHEALS: eHealth Literacy Scale
eHLA: eHealth literacy assessment toolkit
ERIC: Education Resources Information Center
LISA: Library and Information Science Abstracts
LISTA: Library, Information Science & Technology Abstracts
PRISMA-ScR: Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews
RRSA-h: Research Readiness Self-Assessment-health


Edited by A Mavragani; submitted 27.11.22; peer-reviewed by J Dratva, T Mitsuhashi; comments to author 13.01.23; revised version received 23.01.23; accepted 01.03.23; published 02.06.23.

Copyright

©Bradley Crocker, Olivia Feng, Lindsay R Duncan. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 02.06.2023.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.