A growing number of studies suggest that detection of Alzheimer’s disease can be improved by using information derived from distinct neuroimaging modalities. However, so far it remains unresolved how these modalities can be combined within a deep learning model approach. In this study, we proposed a deep-neural-network model GapNet that can work with incomplete dataset including baseline and longitudinal MR, amyloid-PET, and FDG-PET data. We verified the effectiveness of GapNet by comparing it to the conventional Vanilla neural networks and specifically testing their performance in discriminating between healthy controls and individuals with amyloid changes, which is an important early pathological marker in Alzheimer’s Disease. Results showed that, compared to the Vanilla networks, GapNet achieved higher classification accuracy. In sum, our finding suggested that the GapNet model is a promising deep learning approach for detecting Alzheimer’s disease with multi-modal neuroimaging
|