Search Articles

View query in Help articles search

Search Results (1 to 10 of 14 Results)

Download search results: CSV END BibTex RIS


A Multimodal Large Language Model as an End-to-End Classifier of Thyroid Nodule Malignancy Risk: Usability Study

A Multimodal Large Language Model as an End-to-End Classifier of Thyroid Nodule Malignancy Risk: Usability Study

Beyond text-based tasks, there has been a growing interest in the use of vision-language multimodal LLMs, such as Microsoft’s Large Language and Visual Assistant (LLa VA), which combines a visual encoder with a general LLM to allow it to synthesize both image and text data [12]. Like text-based LLMs, they can be further fine-tuned with domain-specific knowledge.

Gerald Gui Ren Sng, Yi Xiang, Daniel Yan Zheng Lim, Joshua Yi Min Tung, Jen Hong Tan, Chiaw Ling Chng

JMIR Form Res 2025;9:e70863

A Web-Based Lifestyle-Related Course for People Living With Multiple Sclerosis: Quantitative Evaluation of Course Completion, Satisfaction, and Lifestyle Changes Among Participants Enrolled in a Randomized Controlled Trial

A Web-Based Lifestyle-Related Course for People Living With Multiple Sclerosis: Quantitative Evaluation of Course Completion, Satisfaction, and Lifestyle Changes Among Participants Enrolled in a Randomized Controlled Trial

After confirming the feasibility, acceptability, and learnability of the MSOC [14], a large randomized controlled trial (RCT) was designed and is currently being conducted to investigate the effectiveness of a multimodal lifestyle intervention course (IC) in improving health outcomes and Qo L in people living with MS compared to a standard-care course (SCC) [15].

Maggie Yu, Sandra Neate, Steve Simpson-Yap, Rebekah Davenport, William Bevens, George Jelinek, Jeanette Reece

JMIR Hum Factors 2025;12:e59363

Using Digital Phenotyping to Discriminate Unipolar Depression and Bipolar Disorder: Systematic Review

Using Digital Phenotyping to Discriminate Unipolar Depression and Bipolar Disorder: Systematic Review

Only 1 (5%) out of 21 studies used a multimodal approach, collecting diverse information including text, audio, facial expressions, heart rate, and eye movement during participant interactions with a virtual assistant. This study was classified as “multimodal technology.” Among the 11 (52%) out of 21 studies that directly distinguished between UD and BD, 8 (73%) described the mood states of patients (depressive, manic, mixed, or euthymic state).

Rongrong Zhong, XiaoHui Wu, Jun Chen, Yiru Fang

J Med Internet Res 2025;27:e72229

Acoustic and Natural Language Markers for Bipolar Disorder: A Pilot, mHealth Cross-Sectional Study

Acoustic and Natural Language Markers for Bipolar Disorder: A Pilot, mHealth Cross-Sectional Study

Multimodal, smartphone-integrated digital assessments could serve as powerful tools for clinical purposes to remotely complement standard mental health evaluations, potentially contributing to distinguish clinical conditions in people with BD. Feasibility of similar systems seems promising, though issues related to privacy, intrusiveness, and clinical therapeutic relationships should be carefully considered.

Cristina Crocamo, Riccardo Matteo Cioni, Aurelia Canestro, Christian Nasti, Dario Palpella, Susanna Piacenti, Alessandra Bartoccetti, Martina Re, Valentina Simonetti, Chiara Barattieri di San Pietro, Maria Bulgheroni, Francesco Bartoli, Giuseppe Carrà

JMIR Form Res 2025;9:e65555

Empowering Mental Health Monitoring Using a Macro-Micro Personalization Framework for Multimodal-Multitask Learning: Descriptive Study

Empowering Mental Health Monitoring Using a Macro-Micro Personalization Framework for Multimodal-Multitask Learning: Descriptive Study

Participants provided self-annotated emotional states over 2 weeks, creating a rich, multimodal resource for understanding daily mental health dynamics. Developing a macro-micro framework for personalized daily mental health. Our framework develops a multimodal and multitask learning (MTL) strategy, innovatively built global emotion embeddings with individual personalization embedding.

Meishu Song, Zijiang Yang, Andreas Triantafyllopoulos, Zixing Zhang, Zhe Nan, Muxuan Tang, Hiroki Takeuchi, Toru Nakamura, Akifumi Kishi, Tetsuro Ishizawa, Kazuhiro Yoshiuchi, Björn Schuller, Yoshiharu Yamamoto

JMIR Ment Health 2024;11:e59512

Multidisciplinary Design–Based Multimodal Virtual Reality Simulation in Nursing Education: Mixed Methods Study

Multidisciplinary Design–Based Multimodal Virtual Reality Simulation in Nursing Education: Mixed Methods Study

According to user evaluations, multimodal interaction with voice communication and implementation of direct actions in the environment were viewed as positive and provided a moderate level of presence. However, communication accuracy and technical features related to virtual object manipulation required improvement. To the best of our knowledge, this is the first study to combine voice interactions with direct hand manipulations in practical nursing training.

Ji-Young Yeo, Hyeongil Nam, Jong-Il Park, Soo-Yeon Han

JMIR Med Educ 2024;10:e53106

Multimodal ChatGPT-4V for Electrocardiogram Interpretation: Promise and Limitations

Multimodal ChatGPT-4V for Electrocardiogram Interpretation: Promise and Limitations

Accuracy of the multimodal Chat GPT-4 V model in answering multiple-choice questions related to electrocardiogram (ECG) interpretation. The number of correct responses among 3 attempts for each question are shown from left to right. The accuracy rates with at least 1, 2, and 3 correct responses are annotated on the right from the bottom to the top. Different shapes represent different question types. We evaluated Chat GPT-4 V responses using the official reference answers as a standard for reliability.

Lingxuan Zhu, Weiming Mou, Keren Wu, Yancheng Lai, Anqi Lin, Tao Yang, Jian Zhang, Peng Luo

J Med Internet Res 2024;26:e54607

Framework for Ranking Machine Learning Predictions of Limited, Multimodal, and Longitudinal Behavioral Passive Sensing Data: Combining User-Agnostic and Personalized Modeling

Framework for Ranking Machine Learning Predictions of Limited, Multimodal, and Longitudinal Behavioral Passive Sensing Data: Combining User-Agnostic and Personalized Modeling

The framework is designed to induce improved predictions for multimodal sensing. It balances both user-agnostic and personalized modeling of small data sets encountered often in mental and physical health–based studies.

Tahsin Mullick, Sam Shaaban, Ana Radovic, Afsaneh Doryab

JMIR AI 2024;3:e47805

Efficacy of a WeChat-Based Multimodal Digital Transformation Management Model in New-Onset Mild to Moderate Hypertension: Randomized Clinical Trial

Efficacy of a WeChat-Based Multimodal Digital Transformation Management Model in New-Onset Mild to Moderate Hypertension: Randomized Clinical Trial

However, studies comprehensively evaluating personalized hypertension management (which focuses on individual differences, customized approach, and incorporation of multimodal data, digital transformation, and multimodal intervention) using rigorous assessments to determine its effectiveness are scarce.

Yijun Wang, Fuding Guo, Jun Wang, Zeyan Li, Wuping Tan, Mengjie Xie, Xiaomeng Yang, Shoupeng Duan, Lingpeng Song, Siyi Cheng, Zhihao Liu, Hengyang Liu, Jiaming Qiao, Yueyi Wang, Liping Zhou, Xiaoya Zhou, Hong Jiang, Lilei Yu

J Med Internet Res 2023;25:e52464