Original Paper
Abstract
Background: Assigning meaningful probabilities of SARS-CoV-2 infection risk presents a diagnostic challenge across the continuum of care.
Objective: The aim of this study was to develop and clinically validate an adaptable, personalized diagnostic model to assist clinicians in ruling in and ruling out COVID-19 in potential patients. We compared the diagnostic performance of probabilistic, graphical, and machine learning models against a previously published benchmark model.
Methods: We integrated patient symptoms and test data using machine learning and Bayesian inference to quantify individual patient risk of SARS-CoV-2 infection. We trained models with 100,000 simulated patient profiles based on 13 symptoms and estimated local prevalence, imaging, and molecular diagnostic performance from published reports. We tested these models with consecutive patients who presented with a COVID-19–compatible illness at the University of California San Diego Medical Center over the course of 14 days starting in March 2020.
Results: We included 55 consecutive patients with fever (n=43, 78%) or cough (n=42, 77%) presenting for ambulatory (n=11, 20%) or hospital care (n=44, 80%). In total, 51% (n=28) were female and 49% (n=27) were aged <60 years. Common comorbidities included diabetes (n=12, 22%), hypertension (n=15, 27%), cancer (n=9, 16%), and cardiovascular disease (n=7, 13%). Of these, 69% (n=38) were confirmed via reverse transcription-polymerase chain reaction (RT-PCR) to be positive for SARS-CoV-2 infection, and 20% (n=11) had repeated negative nucleic acid testing and an alternate diagnosis. Bayesian inference network, distance metric learning, and ensemble models discriminated between patients with SARS-CoV-2 infection and alternate diagnoses with sensitivities of 81.6%-84.2%, specificities of 58.8%-70.6%, and accuracies of 61.4%-71.8%. After integrating imaging and laboratory test statistics with the predictions of the Bayesian inference network, changes in diagnostic uncertainty at each step in the simulated clinical evaluation process were highly sensitive to location, symptom, and diagnostic test choices.
Conclusions: Decision support models that incorporate symptoms and available test results can help providers diagnose SARS-CoV-2 infection in real-world settings.
doi:10.2196/24478
Keywords
Introduction
Despite advances in molecular diagnostics and imaging, ruling in or ruling out COVID-19 infection in an individual patient remains a significant challenge [
]. Current guidelines from the Centers for Disease Control and Prevention (CDC) recommend providers to determine whether signs or symptoms are compatible with COVID-19 infection and to test appropriate patients using nucleic acid amplification tests (NAATs) or antigen detection assays [ ]. However, the diverse clinical presentations of COVID-19 infection may mimic those of common infections, potentially confounding the diagnostic value of presenting symptoms [ ]. Moreover, significant and evolving differences in estimated local disease prevalence for both COVID-19 infection and seasonal respiratory illnesses meaningfully impact differential diagnostic probabilities. Despite the uncertain utility of this symptom and local prevalence information, in low-resource and community settings such as ambulatory clinics, nursing homes, and telemedicine, these may be the only sources of data. In higher-resource settings, NAATs [ ], antibody based lateral flow assays [ ], chest radiography (CXR) [ ], and computed tomography (CT) [ ] may be available, yet published literature notes varied performance. Despite these limitations, clinicians with access to any imaging and testing modalities must optimize diagnostic imaging and testing sequences to appropriately reduce diagnostic uncertainty in a given setting.This complexity highlights the need for reliable and user-friendly clinical decision support systems (CDSS) that suggest optimal testing strategies and quantify SARS-CoV-2 infection risk for patients across the continuum of care. Prior research has demonstrated the potential utility of Bayesian inference [
, ] and machine learning [ , ] methods in diagnostic decision making, but computational clinical decision support has often been underutilized due to a lack of accessibility, transparency, workflow integration, and most importantly, the flexibility to incorporate local provider beliefs into the diagnostic model [ , ].A robust diagnostic risk model should be built on individualized patient data that is easily obtained by patients and health care workers. Menni et al [
] analyzed a large database of smartphone-enabled, self-reported symptom tracker records to predict potential COVID-19 cases using logistic regression models. In the US test set, this approach had a reported sensitivity of 66% and a specificity of 83%. Ahsan et al [ ] used deep learning techniques to differentiate between COVID-19 and non–COVID-19 patients based on open-source imaging and clinical data. However, the performance of this or other machine learning models in clinical settings has not yet been examined.Moreover, in evolving contexts where illness presentation may change depending on host and viral characteristics, large databases of individual patient records may not be available or locally relevant. Constructing inflexible predictive algorithms, such as logistic regression models, based on out-of-date and locally irrelevant data sets would significantly compromise diagnostic accuracy. Addressing these issues, Chishti et al [
] demonstrated the advantages of using flexible probabilistic frameworks built without large-scale clinical data sets to generate ranked differential diagnoses that are more accurate that those developed by physicians.Combining the approaches of this prior work suggests that an appropriate diagnostic support model should rely on easily obtained symptom data, probabilistic frameworks to avoid the need for large-scale data sets, and most importantly, a flexible schema to refine predictions based on provider judgment and the ability to adapt to changes in local prevalence and current diagnostic test performance. To this end, we present a comparison and clinical validation of 3 such quantitative models as well as an ensemble approach to the diagnosis of COVID-19 in ambulatory and acute care settings. We then illustrate how this approach can be employed to help providers optimally reduce diagnostic uncertainty through appropriate diagnostic test choices and update predictions based on local clinical context and test results as that are obtained. Finally, we provide an interactive, online resource to assess COVID-19 infection probability based on user-defined parameters such as local disease prevalence, imaging, and testing performance [
].Methods
Data Acquisition
National and state-specific confirmed cases of COVID-19 as of July 2, 2020, were acquired from the Center for Systems Science and Engineering at Johns Hopkins University [
]. During our model training, validation, and testing process, we assumed a national SARS-CoV-2 infection prevalence of 11.1% based on the total confirmed count of 5,438,325 in the United States as of August 17, 2020 [ ], a population estimate of 328,239,523 [ ], and an estimated reporting rate of 14.9% [ - ]. Prevalence and conditional symptom probabilities for diseases in the differential diagnosis were collected from the CDC and literature estimates ( , Table S1). COVID-19 symptom probabilities were developed primarily from a 393-person consecutive patient series [ ] and supplemented by 3 meta-analyses, which included 3062 [ ], 49,504 [ ], and 53,000 patients [ ]. Where conditional symptom probabilities have not been described in the literature, we used a symptom probability of 1.0% based on our assumption that a higher conditional symptom probability would have been discussed in the literature.To incorporate location and diagnostic test results into risk predictions, we used state-level case figures [
], state-level population data [ ], and the estimated reporting rate [ - ] to compute an estimated SARS-CoV-2 infection prevalence for each state. We sourced imaging diagnostic accuracies from existing literature [ , ] and laboratory test accuracies from the Johns Hopkins Center for Health Security. The reverse transcription-polymerase chain reaction (RT-PCR) sensitivity of 70% is based on published estimates [ ] that take into account operator dependency and variability in viral load across upper respiratory tract sites [ ]. The RT-PCR specificity of 99.8% is based on published data from Abbott Molecular [ ]. Antibody test sensitivity and specificity are based on published figures [ ] for electro-chemiluminescence immunoassay completed between 0-6 days of infection. We computed likelihood ratios and prevalence-adjusted predictive values based on sensitivity, specificity, and our estimated national COVID-19 prevalence of 11.1% ( ).Diagnostic test | Sensitivity (%) | Specificity (%) | Likelihood ratio (%) | Predictive valuea (%) | ||
Positive | Negative | PPVb | NPVc | |||
Chest radiography | 69.0 [ | ]50.0d | 1.4 | 0.6 | 14.7 | 92.8 |
Computed tomography [ | ]97.0 | 57.3 | 2.3 | 0.1 | 22.1 | 99.3 |
RT-PCRe | 70.0 [ | ]97.0 [ | ]23.3 | 0.3 | 74.5 | 96.3 |
Antibody (0-6 days) [ | ]65.5 | 99.8 | 327.5 | 0.3 | 97.6 | 95.9 |
aPrevalence-adjusted predictive values assume a prevalence of 11.1%.
bPPV: positive predictive value.
cNPV: negative predictive value.
dNo published figures available.
eRT-PCR: reverse transcription-polymerase chain reaction.
Training
We developed Bayesian inference network (BN) and set-cover (SC) models from published disease prevalence and conditional symptom probabilities (see
, Table S1). We simulated symptom profiles and diagnoses for 100,000 patients using the published aggregate diagnosis prevalence and conditional symptom probabilities. Of the 100,000 simulated patients, the number of patients assigned to each mutually exclusive diagnosis was proportional to diagnosis prevalence. Within each diagnostic class, we simulated a joint symptom distribution by randomly assigning the presence or absence of each symptom to every patient. The number of patients with a positive symptom within each diagnostic class was proportional to the conditional symptom probability for that symptom and diagnosis. We trained our distance metric learning (DML) and ensemble models on this simulated data.Study Design
We analyzed consecutive ambulatory and hospitalized patients with COVID-19–compatible syndromes presenting to University of California San Diego Medical Center over 14 days in March and April 2020, with institutional review board approval (#200498). Patients were included if they had a recorded presenting illness including fever or cough, and at least a single NAAT in the electronic health record. Patients were labeled “positive” if they had one or more positive RT-PCR tests and a compatible syndrome or findings on radiographic imaging. Patients were labeled “negative” if they had 2 or more consecutive negative NAAT tests (>72 hours apart) or a single negative RT-PCR and a negative antibody test within 14-21 days of symptom onset. Chart review was performed manually by an infectious disease specialist with an anonymized and blinded data set presented for analysis (see
for additional details).Data Analysis
We calculated the sensitivity, specificity, and prevalence-adjusted accuracy as well as the prevalence-adjusted negative predictive value (NPV) and positive predictive value (PPV) of each model on the clinical test data using standard Wald-type CIs [
]. We estimated the 95% CIs for sensitivity and specificity using Clopper-Pearson exact binomial proportion CIs [ ]. We estimated 95% CIs for accuracy using the normal approximation method [ ]. For the imaging and laboratory tests, we computed likelihood ratios based on sensitivity and specificity; and prevalence-adjusted predictive values based on sensitivity, specificity, and an assumed national COVID-19 prevalence of 11.1%.Results
Patient Characteristics
In total, 55 individuals and the presence or absence of 13 symptoms at initial presentation were included in our clinical test data set. Of this, 38 patients (69.1%) were confirmed SARS-CoV-2 infection positive by RT-PCR; 44 subjects were seen via inpatient services, and 11 were seen as outpatients. The majority of subjects (n=43, 78.2%) presented with fever, 63.6% (n=35) with shortness of breath or dyspnea, 54.5% (n=30) with nonproductive cough, 21.8% (n=12) with productive cough, 50.9% (n=28) with fatigue or exhaustion, 9.1% (n=5) with loss of smell, 7.3% (n=4) with sore throat or pharyngalgia, 18.2% (n=10) with body or muscle aches, 16.4% (n=9) with headaches, 16.4% (n=9) with diarrhea, 14.5% (n=8) with nausea, 5.5% (n=3) with vomiting, and 3.6% (n=2) with nasal congestion or rhinorrhea (
; , Table S2).Characteristic | Total, n (%) | SARS-CoV-2 test result | ||
Positive, n (%) | Negative, n (%) | |||
Patients | 55 (100) | 38 (69) | 17 (31) | |
Sex | ||||
Male | 27 (49) | 21 (55) | 6 (35) | |
Female | 28 (51) | 17 (45) | 11 (65) | |
Age (years) | ||||
<60 | 27 (49) | 19 (50) | 8 (47) | |
60-70 | 13 (24) | 9 (24) | 4 (24) | |
70-80 | 9 (16) | 5 (13) | 4 (24) | |
>80 | 6 (11) | 5 (13) | 1 (6) | |
Setting | ||||
Inpatient | 44 (80) | 32 (84) | 12 (71) | |
Outpatient | 11 (20) | 6 (16) | 5 (29) | |
Symptoms | ||||
Fever | 43 (78) | 33 (87) | 10 (59) | |
Dyspnea | 35 (64) | 33 (87) | 2 (12) | |
Dry cough | 30 (55) | 28 (74) | 2 (12) | |
Productive cough | 12 (22) | 4 (11) | 8 (47) | |
Fatigue | 28 (51) | 21 (55) | 7 (41) | |
Loss of smell | 5 (9) | 5 (13) | 0 (0) | |
Sore throat | 4 (7) | 1 (3) | 3 (18) | |
Body/muscle aches | 10 (18) | 8 (21) | 2 (12) | |
Headache | 9 (16) | 6 (16) | 3 (18) | |
Diarrhea | 9 (16) | 5 (13) | 4 (24) | |
Nausea | 8 (15) | 3 (8) | 5 (29) | |
Vomiting | 3 (6) | 1 (3) | 2 (12) | |
Nasal congestion/rhinorrhea | 2 (4) | 0 (0) | 2 (12) | |
Comorbidities | ||||
Cancer | 9 (16) | 5 (13) | 4 (24) | |
Diabetes | 12 (22) | 10 (26) | 2 (12) | |
Cardiovascular disease | 7 (13) | 4 (11) | 3 (18) | |
Hypertension | 15 (27) | 12 (32) | 3 (18) |
Classification Performance in the Clinical Test Data Set
Base models classified SARS-CoV-2 infection in the clinical test data set with sensitivities and specificities of 81.6% (95% CI 65.7 to 92.3) and 58.8% (95% CI 32.9 to 81.6) for the BN model; 0.0% (95% CI 0.0 to 9.3) and 100.0% (95% CI 80.5 to 100.0) for the SC model; 84.2% (95% CI 68.7 to 94.0) and 64.7% (95% CI 38.3 to 85.8) for the DML model; and 81.6% (95% CI 65.7 to 92.3) and 70.6% (95% CI 44.0 to 89.7) for the ensemble model. The overall accuracy of each of these models was 61.4% (95% CI 48.5 to 74.2) for the BN model; 88.9% (95% CI 80.6 to 97.2) for the SC model; 66.9% (95% CI 54.4 to 79.3) for the DML model; and 71.8% (95% CI 59.9 to 83.7) for the ensemble model. The prevalence-adjusted positive and negative predictive values for each model were 19.9% (95% CI 10.5 to 29.2) and 96.2% (95% CI 92.5 to 100.0) for the BN model; 0.0% and 88.9% for the SC model; 23.0% (95% CI 11.3 to 34.7) and 97.0% (95% CI 93.8 to 100.0) for the DML model; and 25.8% (95% CI 11.4 to 40.2) and 96.8% (95% CI 93.7 to 100.0) for the ensemble model (
).Model | Sensitivity (95% CI) (%) | Specificity (95% CI) (%) | Accuracya (95% CI) (%) | Predictive valuea (%) | |
PPVb (95% CI) | NPVc (95% CI) | ||||
Bayesian inference network | 81.6 (65.7-92.3) | 58.8 (32.9-81.6) | 61.4 (48.5-74.2) | 19.9 (10.5-29.2) | 96.2 (92.5-100.0) |
Information-theoretic set cover | 0.0 (0.0-9.3) | 100.0 (80.5-100.0) | 88.9 (80.6-97.2) | 0.0 | 88.9 |
Distance metric learning | 84.2 (68.7-94.0) | 64.7 (38.3-85.8) | 66.9 (54.4-79.3) | 23.0 (11.3-34.7) | 97.0 (93.8-100.0) |
Multinomial logistic regression ensemble | 81.6 (65.7-92.3) | 70.6 (44.0-89.7) | 71.8 (59.9-83.7) | 25.8 (11.4-40.2) | 96.8 (93.7-100.0) |
Logistic regression (Menni et al [ | ])15.8 (6.0-31.3) | 100.0 (80.5-100.0) | 90.6 (82.9-98.3) | 100.0 | 90.5 (88.7-92.2) |
aPrevalence-adjusted metrics assume a COVID-19 prevalence of 11.1%.
bPPV: positive predictive value.
cNPV: negative predictive value.
Incorporation of Location and Diagnostic Test Sequences
We then employed the BN model to evaluate 3 hypothetical patients with 3 different presentations: (1) fever, dry cough, shortness of breath, and anosmia; (2) fever and dry cough; and (3) asymptomatic. We assumed all of these patients presented for care in an area with a local disease prevalence equivalent to the national disease prevalence of 11.1%. For patient 1, we simulated a clinically plausible imaging and test result sequence of negative RT-PCR, negative antibody, and negative CXR. The probability of a COVID-19 diagnosis following symptom collection was 99.8%. Despite negative test results, residual risk due to local disease prevalence and symptoms remained at 97.7%. The change in diagnosis probability, or the reduction in diagnostic uncertainty, was only 2.1% following all 3 negative tests. For patient 2, we simulated the same negative test sequence. In this scenario, the combination of negative test results with nonspecific symptom information resulted in a decrease in residual risk to 12.3%. The reduction in diagnostic uncertainty due to test results was 55.6%, primarily due to negative RT-PCR and negative antibody test results. The negative CXR provided less information as the reduction in diagnostic uncertainty following negative RT-PCR and antibody tests was only 6.2%. For patient 3, we simulated an imaging and test result sequence of negative RT-PCR, positive antibody, and negative CXR. The negative RT-PCR test reduced disease probability by only 0.1%, and the positive antibody test increased the probability of a COVID-19 diagnosis by 8.4%. The CXR results reduced diagnostic uncertainty by 3.0% (
).To illustrate the dependence of risk assessment on local disease prevalence, we simulated a patient with symptoms of only fever and dry cough presenting in 3 locations with significantly different COVID-19 prevalence estimates: Vermont with an estimated statewide prevalence of 1.6%, Utah with an estimated statewide prevalence of 9.8%, and Florida with an estimated statewide prevalence of 18.0% at the time of the simulation. We combined results from 3 common test sequences with our BN pretest probabilities to compute location-dependent risk trajectories. The test sequences included: (1) negative CXR and negative RT-PCR; (2) negative CXR and positive RT-PCR; and (3) positive CXR and negative RT-PCR. Our results indicate that for a pauci-symptomatic patient presenting with identical symptoms in states with significantly different disease prevalence, the posttest probabilities of SARS-CoV-2 infection following common diagnostic test sequences demonstrate marked variation. Moreover, changes in diagnostic probability or reductions in diagnostic uncertainty are highly context and test dependent (
).Discussion
Principal Findings
Our results suggest simple computable models that quantify patient risk of SARS-CoV-2 infection based on key elements of the clinical case can reduce diagnostic uncertainty for providers attempting to rule in or rule out disease with limited or conflicting information.
Building on work by Chishti et al [
], we chose probabilistic models considering the scarcity of detailed, individual patient data and to take advantage of the depth of published literature on aggregate symptom probabilities. Clinicians are trained in evidence-based medicine, and Bayesian reasoning provides a natural framework to understand the impact of incremental information on diagnosis probabilities. Our approaches to making stepwise diagnostic assessments with incremental information mimic clinical workflows and reflect the need for transparency and accommodation of new information critical to clinical decision making. As in Menni et al [ ], we chose clinical indicators that would be easily obtained by patients and providers as well as predictive models that are easily computed and transparent to all users. While other machine learning approaches, such as generative adversarial networks, transfer learning, n-shot learning, and prototypical networks, are also robust for limited data, these methods can be opaque and inaccessible to providers and may be inflexible and fragile in an evolving clinical context.Our most simple model, the Bayesian inference network, is transparent, easily interpreted, and highly modifiable depending on the user’s prior beliefs about location-based prevalence, conditional symptom probabilities, and imaging and laboratory test accuracy. Clinicians, educated in evidence-based medicine and often familiar with Bayesian decision making in diagnostic testing, are ideal users of this model. By developing base models that do not require access to large amounts of patient-level data and can accommodate changes in local provider beliefs and new sources of information, we alert physicians to the utility of using Bayesian reasoning to not only combine multiple data streams in order to make more informed diagnostic decisions but also to guide decisions about use of imaging and testing that will most effectively reduce diagnostic uncertainty.
Limitations
Our study has limitations. First, we used simulated patient data based on prevalence and conditional symptom probabilities to train and validate our DML and ensemble models that biased the ensemble model to heavily weight the DML model predictions. Second, the number of patients in our clinical test data set was relatively small, and this data set was enriched for SARS-CoV-2–positive patients due to the cancellation of all elective procedures and the use of telemedicine for almost all patient visits during the study period, leaving clinics and hospitals open primarily for COVID-19 patients and the acutely ill. Third, 80% of the patients in our clinical test data set were from inpatient services, potentially biasing model accuracy by disease severity. Fourth, we chose as a reference standard the RT-PCR test results for SARS-CoV-2 infection despite outstanding questions about false negative rates in NAAT tests due to operator dependency and patient-level differences in viral loads across upper respiratory tract sites [
, ].Conclusions
Overall, we found that the Bayesian inference network, the metric learning model, and ensemble models trained and validated on a simulated patient data set had sensitivities (81.6%-84.2%) and specificities (58.8%-70.6%) for discriminating between COVID-19 infection and other potential diagnoses in real clinical settings. These models had higher sensitivities than reported for most commonly used diagnostics, and model specificities were higher than those of both imaging modalities. For purposes of comparison, the logistic regression model proposed by Menni et al [
], when applied to our clinical test data set, had a sensitivity of 15.8% and a specificity of 100.0%. Finally, our BN model shows that information acquired by imaging and testing choices is highly dependent on location and symptoms, and emphasizes the utility of a quantitative framework to guide clinical decision making in rapidly changing local environments with potentially unreliable diagnostic tests.Acknowledgments
We would like to thank Mary R Mulcare, MD, at the Weill Cornell School of Medicine; Dana R Levin, MD, MPH, at the Columbia University College of Physicians and Surgeons; and Daniel L Wang, MD, at the University of California San Diego School of Medicine, for scientific guidance, help with data preparation, and manuscript comments.
Conflicts of Interest
None declared.
Supplemental materials.
DOCX File , 315 KBReferences
- Song J, Yun J, Noh J, Cheong H, Kim W. Covid-19 in South Korea — Challenges of Subclinical Manifestations. N Engl J Med 2020 May 07;382(19):1858-1859. [CrossRef]
- Coronavirus Disease 2019 (COVID-19): Overview of Testing for SARS-CoV-2 (COVID-19). Centers for Disease Control and Prevention. 2020. URL: https://www.cdc.gov/coronavirus/2019-ncov/hcp/testing-overview.html [accessed 2020-09-15]
- Bordi L, Nicastri E, Scorzolini L, Di Caro A, Capobianchi MR, Castilletti C, et al. Differential diagnosis of illness in patients under investigation for the novel coronavirus (SARS-CoV-2), Italy, February 2020. Euro Surveill 2020 Feb;25(8):2000170 [FREE Full text] [CrossRef] [Medline]
- Woloshin S, Patel N, Kesselheim AS. False Negative Tests for SARS-CoV-2 Infection — Challenges and Implications. N Engl J Med 2020 Aug 06;383(6):e38. [CrossRef]
- Global Progress on COVID-19 Serology-Based Testing. Johns Hopkins Center for Health Security. 2020. URL: https://www.centerforhealthsecurity.org/resources/COVID-19/serology/Serology-based-tests-for-COVID-19.html [accessed 2020-09-15]
- Wong H, Lam H, Fong A, Leung ST, Chin TWY, Lo CSY, et al. Frequency and Distribution of Chest Radiographic Findings in Patients Positive for COVID-19. Radiology 2020 Aug;296(2):E72-E78 [FREE Full text] [CrossRef] [Medline]
- Ai T, Yang Z, Hou H, Zhan C, Chen C, Lv W, et al. Correlation of Chest CT and RT-PCR Testing for Coronavirus Disease 2019 (COVID-19) in China: A Report of 1014 Cases. Radiology 2020 Aug;296(2):E32-E40 [FREE Full text] [CrossRef] [Medline]
- Grinfeld J, Nangalia J, Baxter EJ, Wedge DC, Angelopoulos N, Cantrill R, et al. Classification and Personalized Prognosis in Myeloproliferative Neoplasms. N Engl J Med 2018 Oct 11;379(15):1416-1430. [CrossRef]
- Palmerini T, Benedetto U, Bacchi-Reggiani L, Riva DD, Biondi-Zoccai G, Feres F, et al. Mortality in patients treated with extended duration dual antiplatelet therapy after drug-eluting stent implantation: a pairwise and Bayesian network meta-analysis of randomised trials. The Lancet 2015 Jun;385(9985):2371-2382. [CrossRef]
- Milea D, Najjar RP, Jiang Z, Ting D, Vasseneix C, Xu X, et al. Artificial Intelligence to Detect Papilledema from Ocular Fundus Photographs. N Engl J Med 2020 Apr 30;382(18):1687-1695. [CrossRef]
- Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017 Jan 25;542(7639):115-118. [CrossRef]
- Liu Y, Chen PC, Krause J, Peng L. How to Read Articles That Use Machine Learning: Users' Guides to the Medical Literature. JAMA 2019 Nov 12;322(18):1806-1816. [CrossRef] [Medline]
- Rosenbaum L, Lamas D. Cents and sensitivity--teaching physicians to think about costs. N Engl J Med 2012 Jul 12;367(2):99-101. [CrossRef] [Medline]
- Menni C, Valdes AM, Freidin MB, Sudre CH, Nguyen LH, Drew DA, et al. Real-time tracking of self-reported symptoms to predict potential COVID-19. Nat Med 2020 Jul 11;26(7):1037-1040. [CrossRef] [Medline]
- Ahsan MM, E. Alam T, Trafalis T, Huebner P. Deep MLP-CNN Model Using Mixed-Data to Distinguish between COVID-19 and Non-COVID-19 Patients. Symmetry 2020 Sep 16;12(9):1526. [CrossRef]
- Chishti S, Jaggi KR, Saini A, Agarwal G, Ranjan A. Artificial Intelligence-Based Differential Diagnosis: Development and Validation of a Probabilistic Model to Address Lack of Large-Scale Clinical Datasets. J Med Internet Res 2020 Apr 28;22(4):e17550 [FREE Full text] [CrossRef] [Medline]
- COVID-19 Infection Probability Assessment. 2020. URL: http://covid-calc.herokuapp.com/ [accessed 2020-09-15]
- Dong E, Du H, Gardner L. An interactive web-based dashboard to track COVID-19 in real time. The Lancet Infectious Diseases 2020 May;20(5):533-534. [CrossRef]
- State Population Totals and Components of Change: 2010-2019. United States Census Bureau. 2019. URL: https://www.census.gov/data/tables/time-series/demo/popest/2010s-state-total.html [accessed 2020-09-15]
- Commercial Laboratory Seroprevalence Data. Centers for Disease Control and Prevention. 2020. URL: https://www.cdc.gov/coronavirus/2019-ncov/cases-updates/commercial-lab-surveys.html [accessed 2020-09-15]
- Havers F, Reed C, Lim T, Montgomery JM, Klena JD, Hall AJ, et al. Seroprevalence of Antibodies to SARS-CoV-2 in Six Sites in the United States, March 23-May 3, 2020. medRxiv Preprint posted online June 26, 2020. [CrossRef]
- Pei S, Shaman J. Initial Simulation of SARS-CoV2 Spread and Intervention Effects in the Continental US. medRxiv Preprint posted online March 27, 2020. [CrossRef]
- Goyal P, Choi JJ, Pinheiro LC, Schenck EJ, Chen R, Jabri A, et al. Clinical Characteristics of Covid-19 in New York City. N Engl J Med 2020 Jun 11;382(24):2372-2374. [CrossRef]
- Zhu J, Ji P, Pang J, Zhong Z, Li H, He C, et al. Clinical characteristics of 3062 COVID-19 patients: A meta-analysis. J Med Virol 2020 Oct 24;92(10):1902-1914 [FREE Full text] [CrossRef] [Medline]
- Heydari K, Rismantab S, Shamshirian A, Lotfi P, Shadmehri N, Houshmand P, et al. Clinical and Paraclinical Characteristics of COVID-19 patients: A systematic review and meta-analysis. medRxiv Preprint posted online March 30, 2020. [CrossRef]
- Ma C, Gu J, Hou P, Zhang L, Bai Y, Guo Z, et al. Incidence, clinical characteristics and prognostic factor of patients with COVID-19: a systematic review and meta-analysis. medRxiv Preprint posted online March 20, 2020.
- Yu F, Yan L, Wang N, Yang S, Wang L, Tang Y, et al. Quantitative Detection and Viral Load Analysis of SARS-CoV-2 in Infected Patients. Clin Infect Dis 2020 Jul 28;71(15):793-798 [FREE Full text] [CrossRef] [Medline]
- Current Molecular and Antigen Tests with FDA EUA Status. Johns Hopkins Center for Health Security. 2020. URL: https://www.centerforhealthsecurity.org/resources/COVID-19/molecular-based-tests/current-molecular-and-antigen-tests.html [accessed 2020-09-15]
- Mercaldo ND, Lau KF, Zhou XH. Confidence intervals for predictive values with an emphasis to case-control studies. Stat Med 2007 May 10;26(10):2170-2183. [CrossRef] [Medline]
Abbreviations
BN: Bayesian network |
CDC: Centers for Disease Control and Prevention |
CDSS: clinical decision support systems |
CT: computed tomography |
CXR: chest radiography |
DML: distance metric learning |
NAAT: nucleic acid amplification test |
NPV: negative predictive value |
PPV: positive predictive value |
RT-PCR: reverse transcription-polymerase chain reaction |
SC: set-cover model |
Edited by G Eysenbach, R Kukafka; submitted 02.10.20; peer-reviewed by M Majurul Ahsan, N Agarwal; comments to author 11.10.20; revised version received 13.10.20; accepted 02.11.20; published 16.12.20
Copyright©Christopher D'Ambrosia, Henrik Christensen, Eliah Aronoff-Spencer. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 16.12.2020.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.