Original Paper
Abstract
Background: In-depth interviews are a common method of qualitative data collection, providing rich data on individuals’ perceptions and behaviors that would be challenging to collect with quantitative methods. Researchers typically need to decide on sample size a priori. Although studies have assessed when saturation has been achieved, there is no agreement on the minimum number of interviews needed to achieve saturation. To date, most research on saturation has been based on in-person data collection. During the COVID-19 pandemic, web-based data collection became increasingly common, as traditional in-person data collection was possible. Researchers continue to use web-based data collection methods post the COVID-19 emergency, making it important to assess whether findings around saturation differ for in-person versus web-based interviews.
Objective: We aimed to identify the number of web-based interviews needed to achieve true code saturation or near code saturation.
Methods: The analyses for this study were based on data from 5 Food and Drug Administration–funded studies conducted through web-based platforms with patients with underlying medical conditions or with health care providers who provide primary or specialty care to patients. We extracted code- and interview-specific data and examined the data summaries to determine when true saturation or near saturation was reached.
Results: The sample size used in the 5 studies ranged from 30 to 70 interviews. True saturation was reached after 91% to 100% (n=30-67) of planned interviews, whereas near saturation was reached after 33% to 60% (n=15-23) of planned interviews. Studies that relied heavily on deductive coding and studies that had a more structured interview guide reached both true saturation and near saturation sooner. We also examined the types of codes applied after near saturation had been reached. In 4 of the 5 studies, most of these codes represented previously established core concepts or themes. Codes representing newly identified concepts, other or miscellaneous responses (eg, “in general”), uncertainty or confusion (eg, “don’t know”), or categorization for analysis (eg, correct as compared with incorrect) were less commonly applied after near saturation had been reached.
Conclusions: This study provides support that near saturation may be a sufficient measure to target and that conducting additional interviews after that point may result in diminishing returns. Factors to consider in determining how many interviews to conduct include the structure and type of questions included in the interview guide, the coding structure, and the population under study. Studies with less structured interview guides, studies that rely heavily on inductive coding and analytic techniques, and studies that include populations that may be less knowledgeable about the topics discussed may require a larger sample size to reach an acceptable level of saturation. Our findings also build on previous studies looking at saturation for in-person data collection conducted at a small number of sites.
doi:10.2196/52998
Keywords
Introduction
Background
In-depth interviews are commonly used to collect qualitative data for a wide variety of research purposes across many subject matter areas. These types of interviews are an ideal approach for examining individuals’ perceptions and behaviors at a level of depth, complexity, and richness that would be challenging to achieve with quantitative data collection methods. Typically, trained interviewers conduct interviews using a guide designed to address the study’s key research aims by asking a series of questions and probes ordered by topic. These interview guides can range from highly structured to completely unstructured (eg, loosely organized conversations). Following the completion of data collection, interview notes and transcripts generated from audio recordings of the interviews are analyzed to assess for patterns in responses among the interviewees or subsets of the participants [
, ].During the COVID-19 pandemic, web-based data collection became increasingly common, as traditional in-person data collection was not possible, and researchers continue to use web-based data collection methods post the COVID-19 emergency, citing advantages such as accessing marginalized populations, achieving greater geographic diversity, being able to offer a more flexible schedule, and saving on travel expenses [
]. Potential concerns about web-based data collection, such as the inability to build rapport and data richness, have been largely unfounded [ , ].While we do not expect web-based data collection to supplant in-person research, it continues to show signs of growth. To date, much of the research on qualitative methods has focused on in-person data collection. Consequently, it will be important to conduct research to determine if previous widely accepted findings hold true for web-based data collection.
Researchers typically make a priori decisions about the number of interviews to conduct with the aim of balancing the need for sufficient data with resource limitations and respondent burden. The concept of saturation is frequently used to justify the study’s rigor with respect to the selected sample size. To provide empirically based recommendations on adequate minimum sample sizes, researchers have conducted studies to assess when saturation occurs. However, multiple types of saturation exist—such as theoretical, thematic, code, and meaning—and within each type of saturation, the definitions and measurement approaches used by investigators vary substantially, as does the level of detail researchers report in publications about their methods for achieving and assessing saturation [
].This study aimed to examine the number of interviews needed to obtain code saturation for 5 recently conducted studies funded by the Food and Drug Administration [
] involving web-based interviews. Specifically, how many web-based interviews are needed to obtain true code saturation (ie, the use of 100% of all codes applied in the study) and how many web-based interviews are needed to achieve near code saturation (ie, the use of 90% of all codes applied in the study)?Literature Review
Multiple authors have defined saturation as the point during data collection and analysis, at which no new additional data are found that reveal a new conceptual category [
- ] or theme related to the research question—an indicator that further data collection is redundant [ ]. Additionally, Coenen et al [ ] specified that no new second-level themes are revealed in 2 consecutive focus groups or interviews.Other authors have distinguished between various types of saturation. One of the most common types of saturation mentioned in the literature is theoretical saturation, which emerges from grounded theory and occurs when the concepts of a theory are fully reflected in the data and no new insights, themes, or issues are identified from the data [
, , , - ]. Hennink et al [ ] expanded this definition, adding that all relevant conceptual categories should have been identified, thus emphasizing the importance of sample adequacy over sample size. Guest et al [ ] operationalized the concept of theoretical saturation as the point in data collection and analysis when new information produces little or no change to the codebook, and van Rijnsoever [ ] operationalized it as being when all the codes have been observed once in the sample.Some authors have defined theoretical saturation, thematic saturation, and data saturation as the same concept [
, ], whereas others have defined these terms differently [ , ]. For example, some authors have defined thematic saturation as the point where no new codes or themes are emerging from the data [ , ]. For thematic saturation to be achieved, data should be collected until nothing new is generated [ , ]. Data saturation has been defined as the level to which new data are repetitive of the data that have been collected [ , , ].Furthermore, Hennink et al [
] distinguished between code saturation and meaning saturation. Code saturation is based on primary or parent codes and relates to the quantity of the data (“hearing it all”). Meaning saturation is based on sub or child codes and relates to the quality or richness of the data (“understanding it all”). Constantinou et al [ ] made the point that it is the categorization of the raw data, rather than the data, that are saturated.The literature reflects multiple methods that have been used to determine saturation [
- , - , , ]. Sim et al [ ] discussed the four general approaches that have been used to determine sample size for qualitative research: (1) rules of thumb, based on a combination of methodological considerations and past experience; (2) conceptual models, based on specific characteristics of the proposed study; (3) numerical guidelines derived from the empirical investigation; and (4) statistical approaches, based on the probability of obtaining a sufficient sample size.For example, Galvin [
] used a statistical approach based on binomial logic to establish the relationship between identifying a theme in a particular sample and within the larger population; for example, number of chances of detecting a theme if that theme exists within number of the population. Using the probability equation, the researcher can determine the number of interviews needed for a stated level of confidence that all relevant themes held by a certain proportion of the population will occur within the interview sample. This method assumes the researcher knows in advance the emergent themes from the study and at what rate they may occur.Constantinou et al [
] used the comparative method for themes saturation, which relies on both a deductive and an inductive approach to generate codes (keywords extracted from the participants’ words) and themes (codes that fall into similar categories). Themes are compared across interviews, and theme saturation is reached when the next interview does not produce any new themes. The sequence of interviews is reordered multiple times to check for order-induced error. When exploring the various methods for determining saturation, researchers reached different conclusions on when saturation was achieved (findings on saturation by other authors are present in ) [ - , - , , , , ].Most studies assessing saturation focused on in-person data collection or did not specify the data collection method. Given recent increases in web-based data collection, studies assessing saturation for web-based interviews are critical to ensure that recommendations regarding sample size are tailored to the mode of data collection [
]. While there is evidence to suggest that the content of data coded from in-person as compared with web-based interviews is conceptually similar [ ], this is a relatively new area of exploration. Rapport may be higher with in-person as compared with web-based interviews [ ], which may impact the amount and type of content generated. Additionally, participants in web-based data collection studies are more geographically diverse and may be more likely to be non-White, less educated, and less healthy than participants in in-person data collection studies [ ].Methods
Study Design
This study was based on analyses from data collected for 5 Food and Drug Administration–funded studies conducted using web-based platforms, such as Zoom (Zoom Video Communications) and Adobe Connect (Adobe Systems), and focused on patients with underlying medical conditions or on health care providers who provide primary or specialty care to patients. All platforms used for these interviews offered audio and video components and allowed for the sharing of stimuli on screen. A brief description of each study is provided in
. Each study’s data had been coded and stored using NVivo software (version 11; QSR International).Study name | Sample size, n | General eligibility criteria | Primary objectives | Summary of topics | Length of interview (minutes) | Number of interview questions | Regions and states covered |
Study A | 30 | Patients diagnosed with a condition treated by biologic medications (eg, cancer, inflammatory bowel disease, and diabetes) | Obtain feedback on multimedia educational materials about biosimilar biologic medications |
| 90 |
|
|
Study B | 48 | Patients diagnosed with vulvovaginal atrophy or type 2 diabetes | Explore how patients use boxed warnings when making decisions about prescription drugs and how well the warnings meet patients’ information needs |
| 30 |
|
|
Study C | 70 | Primary care physicians or specialists who write at least 50 prescriptions per week | Assess how primary care physicians and specialists access, understand, and use prescription drug labeling information, including information on labels for drugs that have multiple indications. |
| 60 |
|
|
Study D | 35 | Patients diagnosed with type 2 diabetes | Understand how patients weigh the potential benefits against possible risks and side effects, dosage and administration characteristics, and costs when selecting treatments for chronic health conditions. |
| 60 |
|
|
Study E | 35 | Patients diagnosed with psoriasis | Understand how patients weigh the potential benefits against possible risks and side effects, dosage and administration characteristics, and costs when selecting treatments for chronic health conditions. |
| 60 |
|
|
Ethical Considerations
This project was determined to not research with human participants by Research Triangle Institute’s institutional review board (STUDY00021985). The original 5 studies that this project is based on were reviewed by Research Triangle Institute’s institutional review board and were determined to be exempt under category 2ii. Participants in these studies were provided information about measures used to protect their privacy and the confidentiality of their data in the study’s consent forms. All participants were provided compensation for their time (the amount and type varied by study).
Data Preparation and Analysis
We established and applied a systematic approach to analyze all 5 study data sets. Our analytic approach was organized into 2 stages—data preparation and data analysis.
Data Preparation
First, because previous interviews sometimes influence moderator probes—for example, the moderator asks a follow-up question based on something they heard in a previous interview—we sorted interviews from each study by interview order. We then extracted code- and interview-specific data from the NVivo databases—including transcript name, code name, number of files coded, number of associated parent and child codes, and number of coding references—and compiled these data in an Excel (Microsoft Corp) file. We then updated the Excel file with important code and interview characteristics, including the order in which interviews were conducted, whether each code was directly (ie, child codes) or indirectly (ie, parent codes) applied to transcripts (in a tiered coding scheme, direct codes are those that have no child codes, whereas indirect codes function as “parents” that have additional codes nested beneath them), and the point at which each code was first applied to an interview. Finally, we created pivot tables within each Excel file to compile the data.
Data Analysis
Once the data were compiled, the data summaries were examined to determine when true saturation and near saturation occurred during data collection. True saturation was defined as 100% of all applied codes being used; near saturation was defined as 90% of all applied codes being used. We calculated saturation separately for each study’s data set, and we calculated saturation separately for all codes (ie, parent and child codes) as compared with direct codes (ie, child codes only). True saturation and near saturation points were identified by calculating the cumulative percentage of new codes for each interview, flagging when 100% and 90% of applied codes had been used.
Results
True and Near Saturation
The number of web-based interviews used across the 5 studies ranged from 30 to 70 (
). True saturation (100% use of all applied codes) was reached in the final or near final interview ( ), suggesting that, even with a large sample size, additional interviews are likely to continue uncovering a small number of new codes or findings.Study | Total interviews, n | Coding: total codes in codebook, n | True saturation: interviews needed, n (%) | Near saturation: interviews needed, n (%) |
Study A | 30 | 657 | 30 (100) | 18 (60) |
Study B | 48 | 313 | 47 (98) | 21 (44) |
Study C | 70 | 362 | 67 (96) | 23 (33) |
Study D | 35 | 205 | 33 (94) | 15 (43) |
Study E | 35 | 200 | 32 (91) | 15 (43) |
Across all studies, near saturation (90% use of all applied codes) was reached near—and often before—the midpoint of data collection. In other words, only a small number of new codes or findings were uncovered once the first half of the sample had been interviewed. In terms of absolute numbers, the point at which near saturation was reached occurred between 33% and 60% (n=15-23) of planned interviews (
). Despite the participants being more geographically, and possibly demographically, diverse compared with typical in-person participants, our findings were similar to previous studies on saturation [ , , ].We examined the types of codes applied after near saturation had been reached. In 4 of the 5 studies, most of these codes (n=8-33, 57%-62%) represented previously established core concepts or themes, such as a trusted source of information, a behavioral intention, or a recommended change to educational material. Codes representing newly identified concepts (n=2-8, 10%-15%), other miscellaneous responses (eg, “in general”; n=6-9, 13%-41%), uncertainty or confusion (eg, “don’t know”; n=0-6, 0%-11%), or categorization for analysis (eg, “correct as compared with incorrect”; n=0-3, 0%-4%) were less commonly applied after near saturation had been reached.
The overwhelming majority of codes applied after near saturation (n=9-41, 73%-82%) had already been established in study codebooks before analysis. Only a small number of codes applied after this point (n=4-20, 18%-27%) were conceptually distinct enough to merit updating the study codebooks by including them. Likewise, most of the codes used after near saturation (n=11-35, 44%-64%) were applied to only a single interview. Far fewer codes were applied to 2 interviews (n=0-13, 0%-27%), 3 interviews (n=0-6, 0%-21%), or 4 or more interviews (n=0-12, 0%-21%).
Study B was an outlier in terms of codes applied after near saturation. This study had fewer codes representing core established concepts (n=8, 28%) and more codes representing newly identified concepts (n=7, 24%) or providing categorization for analysis (n=3, 10%) than other studies. The study also had a much higher proportion of new codes (n=20, 69%) that were added to the study codebook during analysis. These differences may be because the study sampled 2 populations with very different medical conditions (ie, type 2 diabetes as compared with vulvovaginal atrophy), leading to a broader range of applied codes.
In examining the relationship between the number of codes in the codebook for each study, the study with the most codes (study A: 657 codes) required the largest number of interviews to reach both true saturation and near saturation. However, this pattern did not hold true for the remainder of the studies. The study with the next highest number of codes (study C: 362 codes) was third to reach true saturation and last to reach near saturation.
Parent and Child Codes
All 5 study codebooks included both parent (ie, top-level codes) and child codes (ie, subcodes). We examined saturation using two analytic lenses—(1) all codes (parent and child) and (2) parent codes only—to determine if there were differences in when saturation was reached. We found no differences in when true saturation was reached. However, near saturation was reached slightly later (ie, after an additional 3 to 4 interviews) when examining only parent codes (
).Differences by Study
In total, 3 of the studies had codebooks that consisted almost entirely of deductive (ie, concept-driven) codes, whereas the codebooks in the remaining 2 studies contained a mix of both deductive and inductive (ie, data-driven) codes. Although the results were largely consistent across the 5 studies, as expected, the studies that relied heavily on deductive coding reached both true saturation and near saturation sooner. This finding suggests that studies using more inductive coding and analytic techniques may require slightly larger sample sizes to reach saturation.
Structure of an Interview Guide
Although all the studies used a semistructured interview guide, the level of structure varied across studies. The 3 studies (ie, studies C, D, and E) that had a more structured interview guide (eg, questions for which participants were asked their preference among discrete choices or the range of likely answers was limited) reached both true saturation and near saturation sooner. In fact, the study with the most structured guide reached near saturation the soonest, although it fell in the middle for true saturation. This finding suggests that studies using a less structured interview guide may need to conduct more interviews to reach an acceptable level of saturation.
Discussion
Principal Findings
Although true saturation was not reached until the final interview or close to the final interview, near saturation was reached much sooner, ranging from just below to just above the midpoint of data collection, with most of the studies falling just below the midpoint. Although additional interviews conducted after near saturation may result in new information, our findings suggest there may be diminishing returns relative to the resources expended. We have identified several study characteristics that researchers can consider when making decisions on sample size for web-based interviews.
Although our findings were mostly consistent across the 5 studies we examined, near saturation was reached sooner on the studies that consisted of largely deductive codes compared with those that had a greater number of inductive codes. Consequently, researchers should consider their analytic approach when determining sample size. Studies that intend for the coding scheme to be iterative throughout the coding process may want to err on the side of having a slightly higher sample size than if the codebook is expected to consist largely of deductive codes tied to the interview guide.
These studies ranged in length from 30 to 90 minutes, and a majority (n=3) lasted 60 minutes. Although the 90-minute study reached both true saturation and near saturation at the latest point, the shortest interview (at 30 minutes) required the second-highest number of interviews to reach both saturation points. Although the length of the interview may be a minor consideration, the level of structure of the interview guide and the types of codes used seem to be larger drivers.
Our findings point to the need for a slightly higher number of interviews to reach an acceptable level of saturation—categorized by us as near code saturation—than what has been found in other studies. For example, Guest et al [
] found that 6 interviews were enough to get high-level themes, reaching a plateau at 10 to 12 interviews. Similarly, Young and Casey [ ] found that near code saturation was reached at 6 to 9 interviews.Our findings also build on previous studies looking at saturation for in-person data collection conducted at a small number of sites. Data from our studies included participants from all US Census Bureau regions, which provides support that these findings may be more generalizable than previous studies.
Limitations
Our study had several limitations. First, our analysis was conducted on a sample of 5 studies that had similarities. All the studies were related to the medical field, and our study populations (patients with an identified medical condition and health care providers) were knowledgeable about the topics discussed. Second, all the studies were conducted using semistructured interview guides that leaned toward being more structured (ie, interviewers largely stuck to scripted probes as compared with guides that allow for unscripted follow-up probes and unstructured conversations). Additionally, all the studies used a similar approach to coding by using a mix of both deductive and inductive codes (though to varying extents). Consequently, studies with a less structured approach to both the interview and coding process may yield different results. Finally, all our studies are broadly classified as social science research. The findings for other fields of inquiry, such as economic or medical studies, may differ.
Conclusions
Saturation is an important consideration in planning and conducting qualitative research, yet, there is no definitive guidance on how to define and measure saturation, particularly for web-based data collection, which allows for data to be collected from a more geographically diverse sample. Our study provides support that near saturation may be a sufficient measure to target and that conducting additional interviews after that point may result in diminishing returns. Factors to consider in determining how many interviews to conduct include the structure and type of questions included in the interview guide, the coding structure, and the population being studied. Studies with less structured interview guides, studies that rely heavily on inductive coding and analytic techniques, and studies that include populations that may be less knowledgeable about the topics discussed may require a larger sample size to reach an acceptable level of saturation. Rather than trying to reach a consensus on the number of interviews needed to achieve saturation in qualitative research overall, we recommend that future research should explore saturation within different types of studies, such as different fields of inquiry, subject matter, and populations being studied. Creating a robust body of knowledge in this area will allow researchers to identify the guidance that best meets the needs of their work.
Acknowledgments
Research Triangle Institute–affiliated authors received support for the development of this manuscript from the RTI Fellow’s program under RTI Fellow, Leila Kahwati, MPH, MD. All studies included in the analyses were funded by the Food and Drug Administration. The authors would like to thank the following Food and Drug Administration staff for their contribution to this research: Kit Aikin, Kevin Betts, Amie O’Donoghue, and Helen Sullivan.
Data Availability
The data sets analyzed during this study are available from the corresponding author on reasonable request.
Conflicts of Interest
None declared.
Achieving saturation in interviews: saturation type, methods for achieving saturation, and findings by other authors.
DOCX File , 25 KBReferences
- Miles MB, Michael HA, Saldana J. Qualitative Data Analysis: A Methods Sourcebook. Los Angeles, CA. SAGE Publications; 2014.
- Trochim W, Donnelly J. The Research Methods Knowledge Base. 3rd Edition. Mason, OH. Atomic Dog Publishing; 2008.
- Keen S, Lomeli-Rodriguez M, Joffe H. From challenge to opportunity: virtual qualitative research during COVID-19 and beyond. Int J Qual Methods. 2022;21:16094069221105075. [FREE Full text] [CrossRef] [Medline]
- Roberts J, Pavlakis A, Richards M. It’s more complicated than it seems: virtual qualitative research in the COVID-19 era. Int J Qual Methods. Mar 25, 2021;20(20):160940692110029-160940692110020. [CrossRef]
- Vasileiou K, Barnett J, Thorpe S, Young T. Characterising and justifying sample size sufficiency in interview-based studies: systematic analysis of qualitative health research over a 15-year period. BMC Med Res Methodol. 2018;18(1):148. [FREE Full text] [CrossRef] [Medline]
- Lobe B, Morgan DL, Hoffman K. A systematic comparison of in-person and video-based online interviewing. Int J Qual Methods. Sep 15, 2022;21:160940692211270. [CrossRef]
- Constantinou CS, Georgiou M, Perdikogianni M. A comparative method for themes saturation (CoMeTS) in qualitative interviews. Qual Res. Jan 23, 2017;17(5):571-588. [CrossRef]
- Francis JJ, Johnston M, Robertson C, Glidewell L, Entwistle V, Eccles MP, et al. What is an adequate sample size? Operationalising data saturation for theory-based interview studies. Psychol Health. Dec 2010;25(10):1229-1245. [CrossRef] [Medline]
- Galvin R. How many interviews are enough? Do qualitative interviews in building energy consumption research produce reliable knowledge? J Build Eng. Mar 2015;1:2-12. [CrossRef]
- Guest G, Namey E, Chen M. A simple method to assess and report thematic saturation in qualitative research. PLoS One. 2020;15(5):e0232076. [FREE Full text] [CrossRef] [Medline]
- Hennink M, Kaiser BN. Sample sizes for saturation in qualitative research: a systematic review of empirical tests. Soc Sci Med. Jan 2022;292:114523. [FREE Full text] [CrossRef] [Medline]
- Sebele-Mpofu FY, Serpa S. Saturation controversy in qualitative research: complexities and underlying assumptions. A literature review. Cogent Soc Sci. Oct 23, 2020;6(1):1-17. [CrossRef]
- Turner-Bowker DM, Lamoureux RE, Stokes J, Litcher-Kelly L, Galipeau N, Yaworsky A, et al. Informing a priori sample size estimation in qualitative concept elicitation interview studies for clinical outcome assessment instrument development. Value Health. Jul 2018;21(7):839-842. [FREE Full text] [CrossRef] [Medline]
- Coenen M, Stamm TA, Stucki G, Cieza A. Individual interviews and focus groups in patients with rheumatoid arthritis: a comparison of two qualitative methods. Qual Life Res. Mar 2012;21(2):359-370. [CrossRef] [Medline]
- Guest G, Bunce A, Johnson L. How many interviews are enough? Field Methods. Jul 21, 2016;18(1):59-82. [CrossRef]
- Hagaman AK, Wutich A. How many interviews are enough to identify metathemes in multisited and cross-cultural research? Another perspective on Guest, Bunce, and Johnson’s (2006) landmark study. Field Methods. Jul 24, 2016;29(1):23-41. [CrossRef]
- Hennink MM, Kaiser BN, Marconi VC. Code saturation versus meaning saturation: how many interviews are enough? Qual Health Res. Mar 2017;27(4):591-608. [FREE Full text] [CrossRef] [Medline]
- Lowe A, Norris AC, Farris AJ, Babbage DR. Quantifying thematic saturation in qualitative data analysis. Field Methods. Jan 22, 2018;30(3):191-207. [CrossRef]
- van Rijnsoever FJ. (I Can't Get No) Saturation: a simulation and guidelines for sample sizes in qualitative research. PLoS One. 2017;12(7):e0181689. [FREE Full text] [CrossRef] [Medline]
- O’Reilly M, Parker N. ‘Unsatisfactory Saturation’: a critical exploration of the notion of saturated sample sizes in qualitative research. Qual Res. May 17, 2012;13(2):190-197. [CrossRef]
- Weller SC, Vickers B, Bernard HR, Blackburn AM, Borgatti S, Gravlee CC, et al. Open-ended interview questions and saturation. PLoS One. 2018;13(6):e0198606. [FREE Full text] [CrossRef] [Medline]
- Green J, Thorogood N. Chapter 4: In-depth interviews. In: Qualitative Methods for Health Research. 2nd Edition. London. Sage Publications; 2004:198-202.
- Fusch PI, Ness LR. Are we there yet? Data saturation in qualitative research. Qual Rep. 2015:1408-1416. [CrossRef]
- Bowen GA. Naturalistic inquiry and the saturation concept: a research note. Qual Res. Feb 01, 2008;8(1):137-152. [CrossRef]
- Fugard AJB, Potts HW. Supporting thinking on sample sizes for thematic analyses: a quantitative tool. Int J Soc Res Methodol. Feb 10, 2015;18(6):669-684. [CrossRef]
- Sim J, Saunders B, Waterfield J, Kingstone T. Can sample size in qualitative research be determined a priori? Int J Soc Res Methodol. Mar 27, 2018;21(5):619-634. [CrossRef]
- Young DS, Casey EA. An examination of the sufficiency of small qualitative samples. Soc Work Res. 2019;43(1):53-58. [CrossRef]
- Ryan GW, Bernard HR. Techniques to identify themes. Field Methods. Jul 24, 2016;15(1):85-109. [CrossRef]
- Namey E, Guest G, O’Regan A, Godwin CL, Taylor J, Martinez A. How does mode of qualitative data collection affect data and cost? Findings from a quasi-experimental study. Field Methods. Dec 16, 2019;32(1):58-74. [CrossRef]
- Namey E, Guest G, O'Regan A, Godwin CL, Taylor J, Martinez A. How does qualitative data collection modality affect disclosure of sensitive information and participant experience? Findings from a quasi-experimental study. Qual Quant. 2022;56(4):2341-2360. [FREE Full text] [CrossRef] [Medline]
- Rupert DJ, Poehlman JA, Hayes JJ, Ray SE, Moultrie RR. Virtual versus in-person focus groups: comparison of costs, recruitment, and participant logistics. J Med Internet Res. Mar 22, 2017;19(3):e80. [FREE Full text] [CrossRef] [Medline]
Edited by A Mavragani; submitted 22.09.23; peer-reviewed by K Kelly, G Guest; comments to author 24.10.23; revised version received 30.01.24; accepted 09.05.24; published 09.07.24.
Copyright©Claudia M Squire, Kristen C Giombi, Douglas J Rupert, Jacqueline Amoozegar, Peyton Williams. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 09.07.2024.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.