Search Articles

View query in Help articles search

Search Results (1 to 10 of 12 Results)

Download search results: CSV END BibTex RIS


Accurate Modeling of Ejection Fraction and Stroke Volume With Mobile Phone Auscultation: Prospective Case-Control Study

Accurate Modeling of Ejection Fraction and Stroke Volume With Mobile Phone Auscultation: Prospective Case-Control Study

This pilot study describes a novel diagnostic technology using audio recordings from a standard mobile phone. Prior publications have sought both invasive and noninvasive means of describing cardiac function, but very few have moved out of research phases to clinical or practical use [12-16].

Martin Huecker, Craig Schutzman, Joshua French, Karim El-Kersh, Shahab Ghafghazi, Ravi Desai, Daniel Frick, Jarred Jeremy Thomas

JMIR Cardio 2024;8:e57111

Effectiveness of Sensitization Campaigns in Reducing Leprosy-Related Stigma in Rural Togo: Protocol for a Mixed Methods Cluster Randomized Controlled Trial

Effectiveness of Sensitization Campaigns in Reducing Leprosy-Related Stigma in Rural Togo: Protocol for a Mixed Methods Cluster Randomized Controlled Trial

Through the inclusion of audio tools for information dissemination, we aim to acknowledge potential stigma facilitators, such as low literacy rates, in affected groups [30]. Offering learning content in local languages (rather than French), the audio tools could potentially improve campaign effectiveness by increasing self-efficacy through the ability to participate in the intervention as well as being able to understand the content [24-26].

Dominik Jockers, Akila Wimima Bakoubayi, Kate Bärnighausen, P'tanam P'kontème Bando, Stefanie Pechar, Teresia Wamuyu Maina, Jonas Wachinger, Mark Vetter, Yawovi Djakpa, Bayaki Saka, Piham Gnossike, Nora Maike Schröder, Shuyan Liu, Denis Agbenyigan Yawovi Gadah, Christa Kasang, Till Bärnighausen

JMIR Res Protoc 2024;13:e52106

Impact of Audio Data Compression on Feature Extraction for Vocal Biomarker Detection: Validation Study

Impact of Audio Data Compression on Feature Extraction for Vocal Biomarker Detection: Validation Study

Voice data are often captured, transmitted, and stored in various digital formats that may include compression, a common practice used to reduce the size of audio files, making them more manageable and efficient for storage and transmission [9]. It is necessary to consider the potential impact of audio data compression on the overall process of vocal biomarker development as the process can have significant effects on the audio [10].

Jessica Oreskovic, Jaycee Kaufman, Yan Fossat

JMIR Biomed Eng 2024;9:e56246

Investigation of Deepfake Voice Detection Using Speech Pause Patterns: Algorithm Development and Validation

Investigation of Deepfake Voice Detection Using Speech Pause Patterns: Algorithm Development and Validation

This may result in subtle but detectable differences in the way pauses are present in authentic versus cloned audio. Indeed, when humans were asked to distinguish between audio deepfakes and authentic voices, one of the primary justifications for a fake audio classification was unnatural pauses in the recordings [10].

Nikhil Valsan Kulangareth, Jaycee Kaufman, Jessica Oreskovic, Yan Fossat

JMIR Biomed Eng 2024;9:e56245

Real-Time Detection of Sleep Apnea Based on Breathing Sounds and Prediction Reinforcement Using Home Noises: Algorithm Development and Validation

Real-Time Detection of Sleep Apnea Based on Breathing Sounds and Prediction Reinforcement Using Home Noises: Algorithm Development and Validation

The entire night recording of each patient was divided into 30-second segments and converted to a Mel spectrogram for visual representation of audio to visualize how the sound energy in each frequency bin changed over time. The Mel spectrograms were synchronized with manually annotated sleep apnea events from PSG. In this paper, we omitted central apnea and regarded only OSA events as apnea.

Vu Linh Le, Daewoo Kim, Eunsung Cho, Hyeryung Jang, Roben Delos Reyes, Hyunggug Kim, Dongheon Lee, In-Young Yoon, Joonki Hong, Jeong-Whun Kim

J Med Internet Res 2023;25:e44818

Exploring Longitudinal Cough, Breath, and Voice Data for COVID-19 Progression Prediction via Sequential Deep Learning: Model Development and Validation

Exploring Longitudinal Cough, Breath, and Voice Data for COVID-19 Progression Prediction via Sequential Deep Learning: Model Development and Validation

Furthermore, audio characteristics vary among individuals (eg, one COVID-19–positive participant may produce a similar spectrogram as another COVID-19–negative participant). This is not considered in most conventional audio-based COVID-19 detection systems, which so far have only used a single audio sample rather than sequences. This makes automatic detection a difficult task and may lead to wrong predictions.

Ting Dang, Jing Han, Tong Xia, Dimitris Spathis, Erika Bondareva, Chloë Siegele-Brown, Jagmohan Chauhan, Andreas Grammenos, Apinan Hasthanasombat, R Andres Floto, Pietro Cicuta, Cecilia Mascolo

J Med Internet Res 2022;24(6):e37004

Classifying Autism From Crowdsourced Semistructured Speech Recordings: Machine Learning Model Comparison Study

Classifying Autism From Crowdsourced Semistructured Speech Recordings: Machine Learning Model Comparison Study

In this work, we propose a machine learning–based approach to predict signs of autism directly from self-recorded semistructured home audio clips recording a child’s natural behavior. We use random forests, convolutional neural networks (CNNs), and fine-tuned wav2vec 2.0 models to identify differences in speech between children with autism and neurotypical (NT) controls. One strength of our approach is that our models are trained on mobile device audio recordings of varying audio quality.

Nathan A Chi, Peter Washington, Aaron Kline, Arman Husic, Cathy Hou, Chloe He, Kaitlyn Dunlap, Dennis P Wall

JMIR Pediatr Parent 2022;5(2):e35406

Reliability of Commercial Voice Assistants’ Responses to Health-Related Questions in Noncommunicable Disease Management: Factorial Experiment Assessing Response Rate and Source of Information

Reliability of Commercial Voice Assistants’ Responses to Health-Related Questions in Noncommunicable Disease Management: Factorial Experiment Assessing Response Rate and Source of Information

To have a backup of the voice responses, we used an audio recorder (Philips DVT4010, Koninklijke Philips N.V.). Based on Kocaballi et al [28], we tested commonly used unimodal and multimodal VAs. To operationalize the variables developer and modality, we employed the 3 most common Vas (ie, Amazon Alexa, Apple Siri, and Google Assistant) and aimed for the 2 most frequently used devices (ie, smart speaker and smartphone) for each VA [10].

Caterina Bérubé, Zsolt Ferenc Kovacs, Elgar Fleisch, Tobias Kowatsch

J Med Internet Res 2021;23(12):e32161

Complete and Resilient Documentation for Operational Medical Environments Leveraging Mobile Hands-free Technology in a Systems Approach: Experimental Study

Complete and Resilient Documentation for Operational Medical Environments Leveraging Mobile Hands-free Technology in a Systems Approach: Experimental Study

Moreover, the noise in ASR audio input may result in specific types of word errors in the output text interfering with the documentation when extracting relevant medical information. The existing publicly and commercially available ASR models are optimized for the daily conversation and thus may perform poorly when applied to domain-specific clinical speech [30,31]. ASR consists of multiple components to convert input audio to output text.

MinJae Woo, Prabodh Mishra, Ju Lin, Snigdhaswin Kar, Nicholas Deas, Caleb Linduff, Sufeng Niu, Yuzhe Yang, Jerome McClendon, D Hudson Smith, Stephen L Shelton, Christopher E Gainey, William C Gerard, Melissa C Smith, Sarah F Griffin, Ronald W Gimbel, Kuang-Ching Wang

JMIR Mhealth Uhealth 2021;9(10):e32301