دورية أكاديمية

Classifying Autism From Crowdsourced Semistructured Speech Recordings: Machine Learning Model Comparison Study

التفاصيل البيبلوغرافية
العنوان: Classifying Autism From Crowdsourced Semistructured Speech Recordings: Machine Learning Model Comparison Study
المؤلفون: Nathan A Chi, Peter Washington, Aaron Kline, Arman Husic, Cathy Hou, Chloe He, Kaitlyn Dunlap, Dennis P Wall
المصدر: JMIR Pediatrics and Parenting, Vol 5, Iss 2, p e35406 (2022)
بيانات النشر: JMIR Publications, 2022.
سنة النشر: 2022
المجموعة: LCC:Pediatrics
مصطلحات موضوعية: Pediatrics, RJ1-570
الوصف: BackgroundAutism spectrum disorder (ASD) is a neurodevelopmental disorder that results in altered behavior, social development, and communication patterns. In recent years, autism prevalence has tripled, with 1 in 44 children now affected. Given that traditional diagnosis is a lengthy, labor-intensive process that requires the work of trained physicians, significant attention has been given to developing systems that automatically detect autism. We work toward this goal by analyzing audio data, as prosody abnormalities are a signal of autism, with affected children displaying speech idiosyncrasies such as echolalia, monotonous intonation, atypical pitch, and irregular linguistic stress patterns. ObjectiveWe aimed to test the ability for machine learning approaches to aid in detection of autism in self-recorded speech audio captured from children with ASD and neurotypical (NT) children in their home environments. MethodsWe considered three methods to detect autism in child speech: (1) random forests trained on extracted audio features (including Mel-frequency cepstral coefficients); (2) convolutional neural networks trained on spectrograms; and (3) fine-tuned wav2vec 2.0—a state-of-the-art transformer-based speech recognition model. We trained our classifiers on our novel data set of cellphone-recorded child speech audio curated from the Guess What? mobile game, an app designed to crowdsource videos of children with ASD and NT children in a natural home environment. ResultsThe random forest classifier achieved 70% accuracy, the fine-tuned wav2vec 2.0 model achieved 77% accuracy, and the convolutional neural network achieved 79% accuracy when classifying children’s audio as either ASD or NT. We used 5-fold cross-validation to evaluate model performance. ConclusionsOur models were able to predict autism status when trained on a varied selection of home audio clips with inconsistent recording qualities, which may be more representative of real-world conditions. The results demonstrate that machine learning methods offer promise in detecting autism automatically from speech without specialized equipment.
نوع الوثيقة: article
وصف الملف: electronic resource
اللغة: English
تدمد: 2561-6722
Relation: https://pediatrics.jmir.org/2022/2/e35406; https://doaj.org/toc/2561-6722
DOI: 10.2196/35406
URL الوصول: https://doaj.org/article/d85bd7932dfe44bf9cec46a28dc21cf8
رقم الأكسشن: edsdoj.85bd7932dfe44bf9cec46a28dc21cf8
قاعدة البيانات: Directory of Open Access Journals
الوصف
تدمد:25616722
DOI:10.2196/35406