Knowledge Resource Center for Ecological Environment in Arid Area
DOI | 10.1016/j.dsp.2023.104229 |
Efficient multimodel method based on transformers and CoAtNet for Alzheimer's diagnosis | |
Kadri, Rahma; Bouaziz, Bassem; Tmar, Mohamed; Gargouri, Faiez | |
通讯作者 | Kadri, R |
来源期刊 | DIGITAL SIGNAL PROCESSING
![]() |
ISSN | 1051-2004 |
EISSN | 1095-4333 |
出版年 | 2023 |
卷号 | 143 |
英文摘要 | Convolutional neural networks (CNNs) have been widely used in medical imaging applications, including brain diseases such as Alzheimer's disease (AD) classification based on neuroimaging data. Researchers extract the potential brain regions related to AD disease using CNN from various imaging modalities due to its architectural inductive bias. The major limitation of the current CNN-based model is that it doesn't capture long-range relationships and long-distance correlation within the image features. Vision transformers (ViT) have proven an astounding performance in encoding long-range relationships with strong modeling capacity and global feature extraction due to the self attention mechanism. However, ViT doesn't model the spatial information or the local features within the image and is hard to train. Researchers have demonstrated that combining CNN and a transformer yields outstanding results. In this study, two new methods are proposed for Alzheimer's disease diagnosis. The first method combines the Swin transformer with an enhanced EfficientNet with multi-head attention and a Depthwise Over-Parameterized Convolutional Layer (DO-Conv). The second method consists of modifying the CoAtNet network with ECA-Net and fused inverted residuals blocks. We evaluated the effectiveness of our proposed methods based on the Open Access Series of Imaging Studies (OASIS) and the Alzheimer's Disease Neuroimaging Initiative (ADNI). Further, we evaluated the proposed methods using the Gradient-based Localization (Grad-CAM) method. The first method achieved 93.23% accuracy of classification on the OASIS dataset. The second method achieved 97.33% accuracy of classification on the OASIS dataset. We applied different multimodal image fusion methods (MRI and PET, MRI and CT) using our proposed method. The experimental results demonstrate that the fusion method based on PET and MRI outperforms the fusion method based on MRI and CT achieving 99.42% accuracy. Our methods outperform some traditional CNN models and the recent methods that are based on transformer for AD classification. |
英文关键词 | CNN ViT CoAtNet Swin transformer Alzheimer's diagnosis Multimodal |
类型 | Article |
语种 | 英语 |
收录类别 | SCI-E |
WOS记录号 | WOS:001094964000001 |
WOS类目 | Engineering, Electrical & Electronic |
WOS研究方向 | Engineering |
资源类型 | 期刊论文 |
条目标识符 | http://119.78.100.177/qdio/handle/2XILL650/395847 |
推荐引用方式 GB/T 7714 | Kadri, Rahma,Bouaziz, Bassem,Tmar, Mohamed,et al. Efficient multimodel method based on transformers and CoAtNet for Alzheimer's diagnosis[J],2023,143. |
APA | Kadri, Rahma,Bouaziz, Bassem,Tmar, Mohamed,&Gargouri, Faiez.(2023).Efficient multimodel method based on transformers and CoAtNet for Alzheimer's diagnosis.DIGITAL SIGNAL PROCESSING,143. |
MLA | Kadri, Rahma,et al."Efficient multimodel method based on transformers and CoAtNet for Alzheimer's diagnosis".DIGITAL SIGNAL PROCESSING 143(2023). |
条目包含的文件 | 条目无相关文件。 |
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。