Exploring Large-Scale Language Models to Evaluate EEG-Based Multimodal Data for Mental Health

التفاصيل البيبلوغرافية
العنوان: Exploring Large-Scale Language Models to Evaluate EEG-Based Multimodal Data for Mental Health
المؤلفون: Hu, Yongquan, Zhang, Shuning, Dang, Ting, Jia, Hong, Salim, Flora D., Hu, Wen, Quigley, Aaron J.
سنة النشر: 2024
المجموعة: Computer Science
مصطلحات موضوعية: Computer Science - Human-Computer Interaction
الوصف: Integrating physiological signals such as electroencephalogram (EEG), with other data such as interview audio, may offer valuable multimodal insights into psychological states or neurological disorders. Recent advancements with Large Language Models (LLMs) position them as prospective ``health agents'' for mental health assessment. However, current research predominantly focus on single data modalities, presenting an opportunity to advance understanding through multimodal data. Our study aims to advance this approach by investigating multimodal data using LLMs for mental health assessment, specifically through zero-shot and few-shot prompting. Three datasets are adopted for depression and emotion classifications incorporating EEG, facial expressions, and audio (text). The results indicate that multimodal information confers substantial advantages over single modality approaches in mental health assessment. Notably, integrating EEG alongside commonly used LLM modalities such as audio and images demonstrates promising potential. Moreover, our findings reveal that 1-shot learning offers greater benefits compared to zero-shot learning methods.
Comment: 6 pages; UbiComp Companion '24, Companion of the 2024 ACM International Joint Conference on Pervasive and Ubiquitous Computing, October 5--9, 2024}{Melbourne, VIC, Australia
نوع الوثيقة: Working Paper
DOI: 10.1145/3675094.3678494
URL الوصول: http://arxiv.org/abs/2408.07313
رقم الأكسشن: edsarx.2408.07313
قاعدة البيانات: arXiv