Exploring Deep Transfer Learning Techniques for Alzheimer's Dementia Detection

Zhu, Youxiang and Liang, Xiaohui and Batsis, John A. and Roth, Robert M. (2021) Exploring Deep Transfer Learning Techniques for Alzheimer's Dementia Detection. Frontiers in Computer Science, 3. ISSN 2624-9898

[thumbnail of pubmed-zip/versions/1/package-entries/fcomp-03-624683/fcomp-03-624683.pdf] Text
pubmed-zip/versions/1/package-entries/fcomp-03-624683/fcomp-03-624683.pdf - Published Version

Download (1MB)

Abstract

Examination of speech datasets for detecting dementia, collected via various speech tasks, has revealed links between speech and cognitive abilities. However, the speech dataset available for this research is extremely limited because the collection process of speech and baseline data from patients with dementia in clinical settings is expensive. In this paper, we study the spontaneous speech dataset from a recent ADReSS challenge, a Cookie Theft Picture (CTP) dataset with balanced groups of participants in age, gender, and cognitive status. We explore state-of-the-art deep transfer learning techniques from image, audio, speech, and language domains. We envision that one advantage of transfer learning is to eliminate the design of handcrafted features based on the tasks and datasets. Transfer learning further mitigates the limited dementia-relevant speech data problem by inheriting knowledge from similar but much larger datasets. Specifically, we built a variety of transfer learning models using commonly employed MobileNet (image), YAMNet (audio), Mockingjay (speech), and BERT (text) models. Results indicated that the transfer learning models of text data showed significantly better performance than those of audio data. Performance gains of the text models may be due to the high similarity between the pre-training text dataset and the CTP text dataset. Our multi-modal transfer learning introduced a slight improvement in accuracy, demonstrating that audio and text data provide limited complementary information. Multi-task transfer learning resulted in limited improvements in classification and a negative impact in regression. By analyzing the meaning behind the Alzheimer's disease (AD)/non-AD labels and Mini-Mental State Examination (MMSE) scores, we observed that the inconsistency between labels and scores could limit the performance of the multi-task learning, especially when the outputs of the single-task models are highly consistent with the corresponding labels/scores. In sum, we conducted a large comparative analysis of varying transfer learning models focusing less on model customization but more on pre-trained models and pre-training datasets. We revealed insightful relations among models, data types, and data labels in this research area.

Item Type: Article
Subjects: STM Library > Computer Science
Depositing User: Managing Editor
Date Deposited: 19 Dec 2022 12:38
Last Modified: 02 May 2024 09:06
URI: http://open.journal4submit.com/id/eprint/293

Actions (login required)

View Item
View Item