A fine-tuned convolutional neural network model for accurate alzheimer’s disease classification

A fine-tuned convolutional neural network model for accurate alzheimer’s disease classification

Play all audios:

Loading...

ABSTRACT Alzheimer’s disease (AD) is one of the primary causes of dementia in the older population, affecting memories, cognitive levels, and the ability to accomplish simple activities


gradually. Timely intervention and efficient control of the disease prove to be possible through early diagnosis. The conventional machine learning models designed for AD detection work well


only up to a certain point. They usually require a lot of labeled data and do not transfer well to new datasets. Additionally, they incur long periods of retraining. Relatively powerful


models of deep learning, however, also are very demanding in computational resources and data. In light of these, we put forward a new way of diagnosing AD using magnetic resonance imaging


(MRI) scans and transfer learned convolutional neural networks (CNN). Transfer learning makes it easier to reduce the costs involved in training and improves performance because it allows


the use of models which have been trained previously and which generalize very well even when there is very little training data available. In this research, we used three different


pre-trained CNN based architectures (AlexNet, GoogleNet, and MobileNetV2) each implemented with several solvers (e.g. Adam, Stochastic Gradient Descent or SGD, and Root Mean Square


Propagation or RMSprop). Our model achieved impressive classification results of 99.4% on the Kaggle MRI dataset as well as 98.2% on the Open Access Series of Imaging Studies (OASIS)


database. Such results serve to demonstrate how transfer learning is an effective solution to the issues related to conventional models that limits the accuracy of diagnosis of AD, thus


enabling their earlier and more accurate diagnosis. This would in turn benefit the patients by improving the treatment management and providing insights on the disease progression. SIMILAR


CONTENT BEING VIEWED BY OTHERS IDENTIFICATION OF ALZHEIMER'S DISEASE USING A CONVOLUTIONAL NEURAL NETWORK MODEL BASED ON T1-WEIGHTED MAGNETIC RESONANCE IMAGING Article Open access 17


December 2020 THREE-ROUND LEARNING STRATEGY BASED ON 3D DEEP CONVOLUTIONAL GANS FOR ALZHEIMER’S DISEASE STAGING Article Open access 07 April 2023 STRUCTURE FOCUSED NEURODEGENERATION


CONVOLUTIONAL NEURAL NETWORK FOR MODELLING AND CLASSIFICATION OF ALZHEIMER’S DISEASE Article Open access 03 July 2024 INTRODUCTION Alzheimer’s disease (AD) is a type of dementia that


primarily affects older adults, causing significant brain tissue damage and the death of nerve cells. This leads to memory loss and difficulties with daily activities such as reading,


writing, speaking, and recognizing family members1. In advanced stages, patients may struggle to recognize loved ones and experience severe symptoms like heart failure and respiratory


dysfunction, which can be fatal2. The symptoms of AD gradually worsen over time, severely impacting overall health3. By 2050, it is estimated that one in every 85 individuals will suffer


from AD, underscoring its increasing prevalence4. Around 60 to 80% of AD patients exhibit dementia symptoms, making it the world’s second most severe neurological disorder5. The disease


causes the cerebral cortex and hippocampus to shrink while the brain’s ventricles enlarge. The reduction in hippocampal size impairs episodic and spatial memory, leading to problems with


judgment, short-term memory, and planning6. Machine learning and deep learning algorithms have proven highly effective in diagnosing a wide range of diseases, including those affecting the


heart, lungs, brain, retina, breast, and bones7,8,9,10,11,12. These advanced methods play a vital role in medical image classification, a key area of computational intelligence13,14,15,16.


Convolutional neural networks (CNNs) stand out in deep learning due to their exceptional accuracy in medical image classification. A significant advantage of CNNs over traditional machine


learning techniques is their ability to automatically extract the most relevant features and classify the stages of Alzheimer’s disease, eliminating the need for manual feature extraction17.


Nevertheless, several such works still reported accuracy, generalizability, and data efficiency issues when using MRI data for Alzheimer’s classification. The high dimensionality of MRI


data makeup makes the features interdependent and hence there is needed for models which can extract useful features with very little manual effort. There are several studies which utilized


MRI data and different AI techniques for the classification glioma, meningioma, and pituitary tumors18,19,20,21 Several recent studies have examined the reliability of MRI for diagnosing


Alzheimer’s disease. For instance, Zhang et al.22 proposed a hybrid machine learning approach that requires retraining when new data is introduced, which limits its scalability in clinical


applications. Similarly, Liu et al.23 explored the use of auto-encoders for classifying Alzheimer’s stages. While the results were promising, the method required significant data


preprocessing, which can be resource intensive. Additionally, Bilfaqih et al.24 developed a hybrid deep learning approach that achieved 98.5% accuracy using a combination of MRI and other


imaging modalities, but the method was constrained by its binary classification output, limiting its broader applicability in distinguishing between multiple stages of Alzheimer’s disease.


Our study builds upon the advancements of those prior works in that we are able to diagnose fast and accurately thanks to transfer learning using pre-trained CNNs even for datasets that are


relatively small for Alzheimer’s recognition. In this regard, this study is developed as a means of improving the diagnosis of Alzheimer’s disease through the application of transfer


learning to counter the challenges posed by scanty labeled data and excessive computational power required to train appropriate models in the medical imaging field. This study focuses on the


early classification of brain tumors using magnetic resonance imaging (MRI) and artificial intelligence (AI) through deep learning (DL). The researchers developed an efficient method based


on transfer learning, using pre-trained models such as Xception, NasNet Large, DenseNet121, and InceptionResNetV2 to extract deep features from MRI scans. The study utilized two benchmark


datasets, with preprocessing and data augmentation applied to enhance training. The models were trained using three optimizers (ADAM, SGD, RMSprop), and the Xception model with ADAM achieved


the best performance, showing 99.67% accuracy on the larger dataset and 91.94% accuracy on the smaller dataset, outperforming other approaches in the field. Our key contributions include: *


Use of transfer learning for MRI-based Alzheimer’s detection: we utilized three pre-trained deep CNNs—AlexNet, GoogLeNet, and MobileNetV2—applying transfer learning to leverage their


feature extraction capabilities while minimizing the need for large amounts of labeled training data. * Comprehensive performance evaluation: we conducted a detailed performance analysis,


assessing the effects of different solvers (Adam, Stochastic Gradient Descent, RMSprop) on model fine-tuning and classification accuracy. * Comparison of pre-trained architectures: we


compared these models in terms of accuracy, precision, and recall to determine the best-performing architecture for Alzheimer’s diagnosis using MRI datasets. Our analysis offers valuable


insights into the strengths and weaknesses of each model in the context of medical image classification. The rest of the paper is structured as follows: Sect “Materials and methods” details


the dataset we employed and the pre-processing techniques we applied. Sect “Results”offers a brief overview of previous studies. Sect “Limitations” discusses the proposed model and research


methods. Sect “Advantages” presents the experimental results and discussions, followed by the conclusion. RELATED WORKS In the last decade, the medical community has notably advanced imaging


techniques such as Positron Emission Tomography (PET), Diffusion Tensor Imaging (DTI), and particularly MRI, which has become a principal method for predicting AD25. For the classification


and diagnosis of AD, various deep learning techniques, such as support vector machines (SVM) using weighted MRI images and random forest classifiers, have been used26. El-Dahshan et al.27


suggested a hybrid approach involving feature extraction, classification, and dimensionality reduction, three-step classification techniques. This method is advantageous due to its speed,


simplicity, affordability, and non-invasive nature. However, a significant limitation of this hybrid technique is the requirement for new training whenever the dataset is increased. Ahmed et


al.28 introduced a CNN-based and patch-based classifier for classifying AD, reducing computing costs and significantly advancing the field. This approach involves multiple operations on the


dataset, enabling direct disease detection from the input data. In another study, Hong et al.29 developed a specialized recurrent neural network (RNN) based on long short-term memory (LSTM)


for the identification and classification of AD. This model uses current and historical patient data, making it crucial for timely disease identification. Furthermore, it primarily focuses


on predicting disease progression using time series data rather than just classification, making it more effective in anticipating future symptoms and enabling earlier interventions for


Alzheimer’s patients. This method uses the patient’s previous data that are connected to the patient’s present and historical data; this technique is significant for the timely


identification of the disease. The proposed technique primarily uses time series data and describes disease prediction instead of classification. Gupta et al.30 developed a model for


classifying AD using the ADNI dataset. Based on a sparse auto-encoder, the model achieves a maximum accuracy of 95%. Suk et al.31 introduced a new approach for classifying stages of AD and


Mild Cognitive Impairment (MCI) converters using an auto-encoder network, which achieved an accuracy of 95.9% across these MCI stages. Islam et al.32 proposed a CNN-based model for early


disease identification built on the Inception-V4 network. The input and output layers were processed using the inception-A, B, and C components and the reduction-A and B components, with


input and output passing through numerous filter concatenation procedures. The model achieved a total accuracy of 73.75% during the training and testing phases using the open-access imaging


studies (OASIS) database. A feature extraction framework was presented by Guerrero et al.33 that was established using substantial inter-subject variability. The proposed model achieved an


overall accuracy of 71%. However, it has some limitations due to its implementation as a binary classification model. Gorji et al.34 developed a convolutional neural network model with a


94.54% accuracy rate in classifying early and late stages of mild cognitive impairment. In Hosseini-Asl35, deep learning techniques have recently solved the inability of many computer-aided


diagnosis structures to extrapolate the distinguishing characteristics from the raw picture data automatically. There are four essential phases: features are extracted, segments are created,


skulls are stripped, and normalization is required in end-to-end learning to create an accurate prediction of diseases. Additionally, He et al.36 proposed the Deep Residual Network to


counteract the erosion of training accuracy, a common challenge exacerbated by the scarcity of annotated datasets. This limitation makes applying CNN in medical imaging particularly


challenging, as they require substantial data for practical training. Hazarika et al.37 examined several prominent deep-learning models and their effectiveness in classifying AD. Of these,


the DenseNet-121 model delivered the best performance. In another study, Shanmugam et al.38 tested three pre-trained CNN models (including GoogLeNet, AlexNet, and ResNet-18) for the early


detection of AD and related cognitive issues. Among them, ResNet-18 achieved the highest accuracy, recording a rate of 97.51%. Meanwhile, Battineni et al.39 utilized various deep neural


networks to classify MRI brain images, employing standard and fine-tuned features. This approach led to the highest accuracies of 95.7% with the CNN model and 95.5% with the DenseNet model.


Detecting Alzheimer’s disease (AD) through deep learning techniques has shown promise; however, several challenges persist that hinder the effectiveness and generalizability of these models


according to Table 1. Below is an analysis of how our research addresses the gaps in Alzheimer’s detection using MRI images through deep learning: Computational Resources and Efficiency:


Training deep architectures imposes significant computational resources and time overheads. A CNN-based patch classifier designed by Ahmed et al.,28 addresses these limitations with respect


to computational load, but the ensuing operations related to dataset’s creation are still elaborate and costly in resources, warranting the quest for more effective solutions. Our approach


addresses the challenge of computational burden by adopting transfer learning, which precludes the training of any models from scratch especially incurring low computational costs. Through


this strategy, it also becomes possible to make use of optimized/ trained CNN models which is an improvement from the previous work which were resource-intensive and involved complicated


processes. Additionally, we conducted a comparative analysis of three solvers (ADAM, SGD, and RMSprop), providing a comprehensive evaluation of their efficiency in terms of training


convergence and performance. This not only optimized model performance but also highlighted the most resource-efficient approaches, making our model more practical for real-world


applications where computational resources may be limited. Early detection and disease progression: while some approaches are devised that focus solely on identifying and classifying stages


of the disease after ample brain degeneration has taken place, early-stage detection is quite problematic. For the purpose of disease progression prediction, Hong and colleagues29 employed a


long-short-term-memory (LSTM)-based recurrent neural network, stressing this method’s scope was largely dependent on historical patient data made available, which can be a limitation for


early diagnosis. Our research effectively addresses the issue of the early detection of different stages of AD, including Very Mild-Demented and Mild-Demented ones. In other words, the model


can identify the early phase of dementia, which is essential to start intervention and treatment as soon as possible. This early detection is important as the progression of the disease can


be slowed down and the patient’s condition is likely to improve. By focusing on these early stages, our approach fills a key research gap, as identified in the work of Hong et al.,29 and


demonstrates the model’s capacity to support early-stage dementia identification, where interventions are most effective. Model Complexity and Interpretability: Accurate predictions have


been recorded in performing specific tasks using advanced models such as deep residual networks and autoencoders. However, such advanced models may present problems with regards to


interpretation and reproducibility. For example, Hosseini-Asl et al.35 proposed an end‐to‐end deep learning system for AD diagnosis but stressed that many preprocessing steps were necessary


adding to the complexity and potential sources of errors. In this research, we used CNN architectures like AlexNet, MobileNetV2, and GoogleNet that strike a good balance between performance


and interpretability as opposed to using complex models like autoencoders or deep residual networks. There is a degree of explainability which is intrinsic to CNNs as their convolutional


layers are looking at local features like edges or textures of MRI images and make it possible to see which areas of the brain contribute to the model. There is a high importance of this


form of interpretability in medical cases, as doctors need to know how the AI is making decisions. Thus, by applying pre-trained networks using transfer learning, we reduced computational


complexity while ensuring the model is simple and easy to understand. Limited dataset size and quality: deep learning architectures like convolutional neural networks (CNNs) rely on large,


annotated datasets to achieve their intended purpose effectively. Acquisition of such datasets is difficult in medical imaging due to the lack of labeled data as well as differences in


imaging protocols from one institution to the other. For example, He et al.36 pointed out that the availability of annotated data is a major bottleneck in training deep networks for AD


detection, especially so given that CNNs more often require large datasets. In our study, we efficiently incorporate transfer learning by incorporating pre-trained models (AlexNet,


MobileNetV2, and GoogleNet), which relieves the burden of large scale, well-annotated datasets. Datasets that are smaller in size can be used to fine-tune pre-trained models to address the


challenge of dataset scarcity. Moreover, the application of transfer learning in medical image analysis has gained much acceptance to address the challenge of limited dataset size.


Generalization across datasets: a model that is trained on specific data may not work as effectively on other data sets due to its unique distributions, imaging methods, or even different


patients. For instance, Battineni et al.39 achieved accurate results employing CNN and DenseNet architecture for Advanced Dementia diagnosis but indicated that some level of fine-tuning is


required to achieve such results efficiently, may limit generalizability across different MRI datasets. This work displays great results for the classification of AD with MRI images both on


the Kaggle dataset as well as OASIS. Nonetheless, we consider assessment of the model’s generalizability as very important, and we agree with the opinion of other studies, such as the one


conducted by Battineni et al.,39. Therefore, further research intends to incorporate external validation on different datasets to evaluate the performance of the model in various clinical


and imaging conditions. MATERIALS AND METHODS EMPLOYED DATASETS Using two MRI datasets, we tested our proposed model for identifying and classifying AD. The first dataset, available on


Kaggle40 and referred to here as the Kaggle dataset, is a multi-class, multi-label database containing 5000 images divided into four classes, each representing different stages of dementia


severity (Table 2). Given that the original images in the Kaggle dataset were highly imbalanced, which could lead to overfitting issues during deep learning training, we increased the number


of images. Different techniques were used to augment the data and improve model performance. Vertical and horizontal flipping, along with rotation, were applied to enhance the diversity of


the dataset. Initially, images were rotated by 45° to introduce variation. If class imbalance remained, additional rotations of 90 and 180° were performed. These combined techniques ensured


a balanced number of images across all classes, helping the model generalize more effectively. For the augmentation process, the same techniques were applied uniformly across all classes to


maintain consistency. This approach successfully expanded the Kaggle dataset to 10,254 MRI images (Table 2; Fig. 1). Figure 1 shows sample images for each class from the Alzheimer’s disease


dataset. The images were resized to meet the input specifications of the pre-trained models. We also performed the external evaluation of our proposed model using a second dataset of 382 MRI


images sourced from the OASIS database curated by Daniel S. Marcus at the Neuroimaging Informatics Analysis Center (NIAC) at Washington University School of Medicine41 (Table 3). This


dataset, referred to here as the OASIS dataset, consists of 3 to 4 individual T1-weighted MRI scans contributed by each subject, all obtained during a single imaging session. For


classification purposes, we divided the images into four classes according to the (CDR) scores: CDR-0 (non-demented), CDR-0.5 (very mild-demented), CDR-1 (mild-demented), and CDR-2


(moderate-demented). This dataset comprises MRI images of AD patients aged 23 to 86 years. To ensure balanced data and enhance the model’s learning efficacy, we also applied an augmentation


strategy to this dataset. While the original OASIS41 dataset images are 256 × 256 pixels, our proposed CNN model requires images to be resized to 224 × 224 pixels. Therefore, we


appropriately scaled the OASIS dataset images to fit the required dimensions for our model. Finally, for both datasets, we partitioned the data into two subsets: the training set, which


comprises 80% of the images, and the test set, which contains the remaining 20% to provide sufficient training data, ensure better generalization, and balance overfitting and underfitting.


Although the Kaggle dataset is popular, it had problems with class imbalance and a limited number of images, which were fixed during model training. We evaluated the trained model’s


generalizability using the OASIS dataset for external validation, and it performed admirably. Our study is currently restricted to these two datasets, but we intend to test the models on


more datasets in the future to assess their generalizability and performance in further detail. GRAD-CAM VISUALIZATIONS Figure 2 presents the Grad-CAM Visualizations of Correctly MRIs.


Original MR images of four different types alongside their respective Grad-CAM heatmaps. Each pair shows the original image on the left, followed by the Grad-CAM visualization on the right.


The true and predicted classes are displayed for each image. Grad-CAM highlights the regions of the image that the model focused on when making its classification decision. Warmer colors


(red/yellow) indicate regions with a higher contribution to the model’s decision, while cooler colors (blue/green) signify less impact. These visualizations provide insight into the model’s


attention to key features that align with expert-based classification criteria. Figure 3 shows the Grad-CAM Visualizations of Misclassified MRIs. Misclassified MRIs of four Alzheimer’s


classes and their corresponding Grad-CAM heatmaps. For each pair, the left image shows the original MR image, while the right image displays the Grad-CAM heatmap. The true class and


predicted class for each image are labeled, providing insights into the model’s misclassification patterns. Grad-CAM highlights the regions of the image that the model focused on during its


prediction process, with warmer colors (red/yellow) indicating areas that contributed most to the model’s decision, and cooler colors (blue/green) showing less influential regions. These


visualizations reveal potential challenges in differentiating between MR images of Alzheimer’s types. EXPERIMENTAL SETUP This study’s deep neural networks were developed using Matlab version


9.8, 2020a. All experiments were conducted on a machine with an Intel Core (TM) i7-6500 CPU @ 2.50 GHz and 16 GB RAM. In this research, we applied CNN pre-trained neural networks through


transfer learning to classify AD. This research utilizes pre-trained models such as AlexNet, GoogleNet, and MobileNetV2 for classification purposes. We carefully examined how changing


different parameters could help us achieve the best accuracy for our model. We experimented with various epochs and different initial learning rates during training to understand their


impact on performance. After that, we investigated which solver would work best with the parameters we tested. To do this, we tried out several solvers with three different architectures.


Our goal was to fully fine-tune the model, ensuring it performs optimally based on our findings. Table 4 shows the architectures used in the study, the solvers applied, initial learning


rates, and the number of training epochs. AlexNet, GoogleNet, and MobileNetV2 all used SGDM, ADAM, and RMSPROP as solvers, with a default batch size of 128. The initial learning rate is set


to 0.001, and each model was trained for 25 epochs. Since these pre-trained models have already undergone training on extensive databases of images, they require significantly less time and


computing resources for further training than developing new models45. These networks are comprised of thousands of neurons and have millions of parameters. AlexNet, with 62.3 million


parameters, pioneered deep learning with its three fully connected and five convolutional layers, while GoogleNet’s 22 layers, expanded to 27 with pooling layers, revolutionized neural


network architectures. MobileNetV2, with fifty-three deep layers, excels in classifying 100 object categories, leveraging millions of images for training. Although the number of layers in


each model differs, some common layers are found in almost all CNN models, including input, convolution, pooling, fully connected, and output layers42. The solver coordinates the network’s


forward inference and backward gradients to create parameter updates that aim to reduce the loss. This process is known as model optimization. The solver is in charge of managing the


optimization and producing parameter updates. In our study, we used three different solvers to optimize training architectures, as follows43: STOCHASTIC GRADIENT DESCENT WITH MOMENTUM (SGDM)


Stochastic gradient descent with momentum (SGDM) is an optimization algorithm that enhances the basic gradient descent approach by introducing a momentum factor. SGDM updates the network’s


weights and biases to reduce the loss function by incrementally moving toward the negative gradient of the loss function with each iteration. This momentum factor helps reduce oscillation


during parameter updates, leading to smoother convergence towards the optimal solution43. $$\delta \ell +1\,=\,\delta \ell - a\nabla E(\delta \ell ){\text{ }}+g(\delta \ell - \delta \ell -


1)$$ (1) > Where ℓ represents the number of iterations, α > 0 specifies > the learning rate, δ represents the parameter vector, E(δ) denotes > the loss function, and γ indicates 


the contribution of the previous > gradient step to the current iteration. ROOT MEAN SQUARE PROPAGATION (RMSPROP) RMSPROP is one method that stores a moving average using the squares of


the parameter gradients by element43. $$v\ell ={\rm B}2v\ell - 1+(1 - {\rm B}2){\text{ }}[\nabla E(\delta \ell )]2$$ (2) vℓ: Exponentially Weighted Moving Average (EWMA) of the square of


gradients at layer ℓ. B2: Hyperparameter scaling the previous EWMA term. (1−B2): Complement of B2, controlling direct contribution of current gradient’s square. [∇E(δℓ)]^2: Square of the


gradient of the loss function. Multiplying B2 with vℓ scales the previous EWMA term. Multiplying (1−B2) by [∇E(δℓ)]^2 scales the current gradient’s square. Combined, they update parameters


during optimization, often used in algorithms like Adam for stochastic optimization. Where B2 reflects the moving average’s decay rate. The squared gradient’s average length is equal to 1/


(1−B2), with the value of 10, 100, and 1000 parameter updates, respectively43. $$\delta \ell +1\,=\,\delta \ell - a\nabla E(\delta \ell ){\text{ }}/\surd v\ell + \in$$ (3) Here, each element


is divided separately When using RMSProp, the learning rates for parameters with small gradients are effectively increased, while the learning rates for parameters with high gradients are


effectively decreased. To prevent division by zero, a small constant ɛ is included. ADAPTIVE MOMENT ESTIMATION (ADAM) ADAM employs a parameter similar to RMSPROP with the addition of a


momentum term. It maintains track of the parameter gradients and their squared values using a moving average based on elements43. $$m\ell = {\rm B}1m\ell - 1 + (1 - {\rm B}1)\nabla E{\text{


}}(\delta \ell )$$ (4) $$v\ell =B2v\ell - 1+(1 - B2){\text{ }}[\nabla E(\delta \ell )]2$$ (5) Where the decay rates for the training choices squared gradient decay factor and gradient decay


factor are B1 and B2, respectively. To update the network settings, Adam employs moving averages as43: $$\delta \ell +1\,=\,\delta \ell - aml/\surd vl+\smallint$$ (6) Where ɛ with the help


of the epsilon training option. If gradients across numerous iterations are the same, using a moving average of the gradient enables parameter changes to build momentum in one direction.


When noise dominates the gradients, the gradient’s moving average becomes smaller, resulting in smaller parameter updates. The proposed model is shown in Fig. 4. The diagram represents a


comprehensive workflow for proposed model, divided into three key layers: data layer, training layer, and testing layer. Detailed explanation of each layer is given as: DATA LAYER Data


Source: Data is collected from hospitals, consisting of Alzheimer’s patient scans. Data Augmentation & Image Processing: To improve the robustness of the model, data augmentation


techniques were implemented such as flipping, rotation, scaling, etc. This step helps in creating more diverse training data and prevents overfitting. Image processing techniques also


include that the images are converted into a suitable format for the model. Data Division: After preprocessing, the dataset is divided into training, and testing sets to ensure proper


evaluation of the model’s performance. TRAINING LAYER Model selection: Three different pre-trained convolutional neural networks (CNNs) are utilized: AlexNet, GoogleNet, and MobileNetV2.


These models are known for their capability to extract significant image features. Optimizer choice: Three different optimization techniques like SGDM, ADAM, and RMSPROP are used for each


architecture. These techniques adjust the model’s weights during training to make the model perform better. Learning criteria: The process involves a decision block to determine whether the


model has met the predefined performance criteria. If the model does not meet the expected accuracy or other criteria, it loops back for weight retraining, optimizing until satisfactory


results are achieved. TESTING LAYER The testing phase involves using unseen images to evaluate the model’s generalization capability. Classification: The trained model classifies the input


images into one of the four categories: mild demented, moderate demented, very mild demented, and non-demented. This classification assists in detecting the level of Alzheimer’s disease in


the patient based on the provided medical images. The evaluation matrices listed below were employed to assess the performance of the implemented models7: $$Accuracy~=\frac{{\varepsilon +


\propto }}{{\varepsilon +\gamma + \propto +\delta }}$$ (7) $$Specificity~=\frac{ \propto }{{ \propto +\gamma }}$$ (8) $$\operatorname{Re} call~or~Sensitivity=\frac{\varepsilon }{{\varepsilon


+\delta }}$$ (9) $$\Pr ecision~=\frac{\varepsilon }{{\varepsilon +\gamma }}$$ (10) $$F1~score~=\frac{{2*~\Pr ecision~*~\operatorname{Re} call~}}{{~\Pr ecision~+~\operatorname{Re} call~}}$$


(11) $$Misclassification~rate=100~ - ~(\frac{{\varepsilon + \propto }}{{\varepsilon +\gamma + \propto +\delta }})$$ (12) Where ε represents the predicted true positive value, ∝ represents


the predicted true negative value, γ represents the predicted false positive value, and δ represents the predicted false negative value. RESULTS For classification purposes, we divided the


images into four classes according to the CDR scores: CDR-0 (Non-Demented), CDR-0.5 (Very Mild-Demented), CDR-1 (Mild-Demented), and CDR-2 (Moderate-Demented). We implemented three distinct


CNN models (AlexNet, GoogleNet, and MobileNetV2), testing each model with three different solvers (SGDM, ADAM, RMSPROP) to identify which model performs best with each solver. In this


section, we provided detailed results for each model alongside the solver it performed best with, ensuring that the results section is brief. Table 5 illustrates the number of epochs,


initial learning rate, number of iterations, implemented solvers, and accuracy performance of each model. We calculated the true positive, true negative, false positive, and false negative


values for each class using the models that yielded the best results with specific solvers. Based on these values, we performed additional calculations to assess the implemented


architectures’ class-wise precision, misclassification rate, accuracy, recall, specificity, and F1 score. After selecting the optimal hyper-parameters, we evaluated the performance of these


models using all the data from the training and testing subsets CLASS-WISE DETAILED RESULTS OF IMPLEMENTED ARCHITECTURES TRAINED WITH OPTIMAL SOLVERS PERFORMANCE EVALUATION OF ALEXNET MODEL


TRAINED WITH ADAM SOLVER To evaluate the effectiveness of different CNN architectures, we tested the AlexNet model using the ADAM solver on a dataset of MRI images. We assessed the


performance of the AlexNet model trained with the ADAM solver on four classes of MRI images sourced from the Kaggle dataset. Our findings revealed that the model attained an accuracy of


99.4%. We evaluated the class-wise precision and recall of the model (Table 6; Figs. 5 and 6). Our findings revealed that the model achieved 99.4% precision and 100% recall for the


Mild-Demented class, 100% precision and 100% recall for the Moderate-Demented class, 99.6% precision and 98.0% recall for the Non-Demented class, and 98.6% precision and 99.6% recall for the


Very Mild-Demented class. Next, we assessed the number of correctly classified images in each class. Out of the 515 images in the Mild-Demented class, 512 were correctly classified, with


three misclassified. All 514 images in the Moderate-Demented class were correctly classified. Of the 504 images in the Non-Demented class, 502 were correctly classified, with two


misclassified. Lastly, out of the 518 images in the Very Mild-Demented class, 511 were correctly classified, with seven misclassified. Analyzing the performance of the AlexNet model, which


was trained with the ADAM solver, across various stages of Alzheimer’s Disease, namely: mild-demented, moderate-demented, non-demented and very mild-demented disorders has also been done in


Table 6. Performance was quantified in terms of Accuracy, Precision, F1-Score, Recall, Specificity and Misclassification Rate, for the respective cases. While the model performed quite well,


we noticed very few instances of misclassification especially in the Non Demented and Very Mild Demented levels. The model performed exceptionally well for the Mild-Demented and


Moderate-Demented classes, with the Moderate-Demented classification showing 0% misclassification, indicating the accuracy of the model in differentiating moderate dementia from other


levels. In contrast, there were very low (0.00585 and 0.00439, respectively) misclassification rates for the Non-Demented and Very Mild-Demented classes, which indicate that it may be


somewhat difficult to the subject who can carry the two stages. More specifically in the Non-Demented class, there were 10 False Negatives (FN) and 2 False Positives (FP), which means that


the Recall (98.05%) and F1-Score (98.82%) were slightly low than in other classes. This could mean that some Non-Demented individuals have dementia stage features overlapped with them which


makes it difficult for the model to distinguish between very early dementia, and completely healthy individuals, as is the case here. Subtle structural changes in the brain that do not


manifest as clinical symptoms may contribute to these misclassifications. There were two False Negatives (FN), and seven False Positives (FP) in the Very Mild-Demented class, which provided


a Precision of 98.65 and a Recall of 99.61. The higher rate of False Positives in this class indicates that the model at times, misclassified Non-Demented patients as those with Very


Mild-Dementia. This may be attributed to the thin line that exists between normal aging and a very mild form of cognitive impairment that exhibits no structural changes in the brain at this


stage — changes that can easily be confused with variations within normal aging patterns. Possible Reasons for misclassifications could be the feature overlap– Non-Demented and Very mild


Demented individuals tend to have indistinct MRI characteristics in most cases as in the case of atrophy or thinning of the hippocampus. Entering into some of these changes in the brain at


this stage is a feature that may not be easy to appreciate without finer techniques of feature extraction. CNNs are great for facilitating the process of feature extraction. When it comes to


MRI scans, there are usually some finer points that unfortunately this model fails to capture most especially for cases that border on the extremes. Transferring learning might be effective


but there would be need for more tweaks or more complex architectures to enhance the results. All calculated values described in Table 6 are shown graphically in Fig. 7. PERFORMANCE


EVALUATION OF GOOGLENET MODEL TRAINED WITH ADAM We assessed the performance of the GoogleNet model trained with the ADAM solver on four classes of MRI images sourced from the Kaggle dataset.


Our findings revealed that the model attained an accuracy of 98.0%. We evaluated the class-wise precision and recall of the model (Table 7; Figs. 8 and 9). Figure 9 demonstrates that we


attained a precision of 99.8% and a recall of 100% for the Mild-Demented class, 100% precision and recall for the Moderate Demented class, 92.6% precision and 100% recall for the


Non-Demented class, and 100% precision and 91.8% recall for the Very Mild Demented class. Of the 513 images in the Mild-Demented class, 512 were correctly classified, with one misclassified.


All 514 images in the Moderate-Demented class were correctly classified. Of the 533 images in the Non-Demented class, 512 were correctly classified, with 41 misclassified. Lastly, all 471


images in the Very Mild-Demented class were correctly classified. Next, we assessed the class-specific GoogleNet trained with ADAM architecture’s performance according to Accuracy,


Precision, F-1 Score, Recall, Specificity and Misclassification rate (Table 7; Fig. 10). For the Mild-Demented class, the following results were attained: Accuracy of 99.95%, F-1 score of


99.90%, Precision of 99.81%, Specificity of 99.94%, Recall of 100%, and a Misclassification rate of 0.000487567. For the Moderate-Demented class, the model achieved an Accuracy, F-1 score,


Precision, Specificity, and Recall of 100%, with a 0 Misclassification Rate. For the Non-Demented class, the following results were obtained: Accuracy of 98.0%, F-1 score of 96.15%,


Precision of 92.59%, Specificity of 97.34%, Recall of 100%, and a Misclassification rate of 0.019990249. For the Very Mild-Demented class, the results were as follows: Accuracy of 97.95%,


F-1 score of 95.73%, Precision of 100%, Specificity of 100%, Recall of 91.81%, and a Misclassification Rate of 0.020477816. The model demonstrated exceptional performance across different


dementia classes, achieving high accuracy, precision, recall, and specificity. Table 7 illustrates the performance results obtained with the GoogleNet model trained using the ADAM solver for


a four-stage AD classification namely Mildly-Demented, Moderately-Demented, Non-Demented and Very Mildly-Demented too. Although the model performed well in accuracy, precision, recall and


specificity on most of the classes, there are some crucial points of the class misclassification rates, especially in Non-Demented and Very Mild-Demented classes, that should be addressed.


As for the Mild-Demented and Moderate-Demented classes, the model performed almost perfectly, achieving a recall and precision of 1, which implies that the model can easily differentiate the


above stages from the rest. The misclassification rates are close to zero, The Non-Demented category, on the other hand, had 41 False Positives, which affected the precision to 92.59%,


which means that some Non-Demented persons were labeled as demented. Such a misdiagnosis can have implications in healthcare as it can lead to needless interventions or treatments for


patients who are wrongly diagnosed with dementia. Even though there is no dementia case that is in this class that was left out, as indicated by the False Negative rate of 0%, the alarming


False Positive rate (41 images) compromises the precision increasing chances of overdiagnosis. In a real-life situation, this introduces frustration to both the patients and their relatives


as well as unnecessary follow-ups or treatments. The class of very mild-demented had 42 cases (False Negatives) which were wrongly identified bringing the recall at 91.81%. Recall here


suggests that the model failed to determine 42 cases of very mild dementia. When it comes to the precision, it stands at 100%, that is, No Non-Demented patient was misclassified as very


mild-demented. However, the low recall shows that the model was able to identify almost 8% of patients who were very mild-demented. For example, this type of underdiagnosis can postpone


treatment, therefore the patients may not benefit from early treatment which is fundamental to curbing the advancement of AD. The Very Mild-Demented class presents a paradox; this is an


instance where the model manages to detect cases but with some limitations. High precision means that when the model classifies an instance as a case of very mild dementia, it is accurate


100% of the time. Nonetheless, the recall is relatively low (91.81%) and this indicates that there are many actual cases that go undetected which is quite worrisome in a clinical setting


where one aims to detect the onset of the illness as early as possible i.e. AD. The False Negatives within this level illustrate losses in regard to identifying the disease at an appropriate


stage, hence, allowing room for treatments which may not be helpful to the patient. PERFORMANCE EVALUATION OF MOBILENETV2 MODEL TRAINED WITH SGDM SOLVER > We assessed the performance of 


the MobileNetV2 model trained with > the SGDM solver on four classes of MRI images sourced from the > Kaggle dataset. Our findings revealed that the model attained an > accuracy of 


96.5%. We evaluated the class-wise precision and recall of the model (Table 8; Fig. 11, and Fig. 12). Figure 11 illustrates that we achieved 93.43% precision and 100% recall for the


mild-demented class, 99.8% precision and 100% recall for the moderate-demented class, 100% precision and 86.5% recall for the non-demented class, and 93.8% precision and 99.6% recall for the


very mild-demented class. Of the 548 images in the mild-demented class, 512 were correctly classified, with 36 misclassified. Of the 515 images in the moderate-demented class, 514 were


correctly classified, with one misclassified. All 443 images in the very non-demented class were correctly classified. Lastly, out of the 545 images in the very mild-demented class, 511 were


correctly classified, with 34 misclassified. Next, we assessed the class-specific MobileNetV2 trained with SGDM architecture’s performance according to Accuracy, Precision, F1-Score,


Recall, Specificity, and Misclassification rate (Table 8; Fig. 13). For the Mild-Demented class, we achieved 98.24% Accuracy, 93.43% Precision, 96.60% F1 score, 100% Recall, 97.66%


Specificity, and a Misclassification rate of 0.017552413. For the Moderate-Demented class, we achieved 99.95% Accuracy, 99.81% Precision, 99.90% F1 score, 100% Recall, 99.93% Specificity,


and a Misclassification rate of 0.000487567. For the Non-Demented class, we attained an Accuracy of 96.64%, a Specificity of 100%, an F1 score of 92.77%, a Recall of 86.52%, a Precision of


100%, and a Misclassification rate of 0.033642126. Moreover, for the Very Mild-Demented class, we achieved 98.24% Accuracy, 93.76% Precision, 96.60% F1 score, 99.61% Recall, 97.79%


Specificity, and a Misclassification rate of 0.017552413. Table 8 shows how the MobileNetV2 model was evaluated in four stages of dementia: mild-demented, moderate-demented, non-demented,


and very mild-demented. The system performs well at 99.66% and 97.31% accuracy rates in the moderate-demented and non-demented classes respectively. However, several errors of classification


are present in the mild-demented and the very mild-demented classes which need to be evaluated, understood, and resolved. When it comes to the mild-demented class, it shows a 1.76%


misclassification rate with 36 false positives (FP) leading to a Precision of 93.43% and an F1-Score of 96.60%. Likewise, when it comes to the Very Mild-Demented class, it also demonstrates


a 1.76% misclassification rate with 34 False Positives (FP) and 2 False Negatives (FN) giving rise to 93.76% precision and an F1-Score of 96.60%. The aforementioned misclassifications


indicate that the model experiences difficulty differentiating between the early stages of dementia, such as Mild and Very Mild Demented, and other stages. This may be due to the absence of


distinct brain changes since these changes start in milder forms of dementia which may also be present in the elderly or in the more advanced forms of dementia. Characteristics observed via


MRI may include atrophy of the hippocampus or thinning of cortical levels which may not be visibly pronounced; causing the model to confuse Mild-Demented and Very Mild-Demented patients with


those who do not have dementia but are in a more advanced stage. In the classification of dementia, the problem of class overlaps in the early-stage diagnosis is a serious one, as it is


often the case that changes in the brain are subtle or non-specific. Here, the term False Positives point to the tendency of the model to incorrectly label Non-Demented subjects as


Mild-Demented and Very Mild-Demented individuals. In practice, this might lead to an increase in the errors of diagnosis among the population, including the treatment of patients with normal


cognition. The model performs exceptionally well in the Moderate-Demented class recording an Accuracy of 99.95%, Precision of 99.81%, and a mere 0.05% misclassification rate. This implies


that the model is quite effective in classifying more advanced forms of dementia as there are gross changes in the brain that can be easily identified. In contrast, the Non-Demented class


shows a rate of 3.36% of misclassification with 69 False Negatives (FN) recorded which resulted in Recall of 86.52% and F1-Score of 92.77% respectively. The False Negatives in this class may


suggest that some Non-Demented individuals are being. Using the precision and recall metrics it is possible to draw conclusions about the clinical relevance of the model. For example, high


precision means that when the model classifies a case as positive Mild-Demented or Very Mild-Demented, this is most likely to be correct, which is important in avoiding becoming overzealous


in pursuit of a diagnosis where one does not exist that may lead to harmful treatments. The non-demented and mildly demented categories being associated with lower recall means that, in


fact, the model may be misclassifying some early-stage patients, which would lead to some cases being undiagnosed. In real-world clinical practice, it becomes imperative for a physician to


be able to detect dementia among patients since it is known that early intervention would mitigate the course of the disease. The model exhibits low recall in the Non-Demented and Very


Mild-Demented classes, which implies that some cases of early-stage dementia in practice may go undetected, leading to a postponement in treatment. This is particularly true here where


improving recall in these classes is very important to maximize the detection of cognitive impairment at its most early stages for timely intervention. CLASS-WISE PERFORMANCE SUMMARY FOR ALL


ARCHITECTURES As shown in Table 9, the AlexNet, GoogleNet, and MobileNetV2 models were evaluated across four classes: mild demented, moderate demented, non demented, and very mild demented.


For Mild Demented, all models had good results, and all the models were examined to have perfect recall, and minimized misclassification were noticed but precision of MobileNetV2 was not


much better, which was 93.43%. In Moderate Demented, both AlexNet and GoogleNet performed perfectly across all metrics (100%), and MobileNetV2 also showed strong performance but with a


slight misclassification rate. In non Demented, overall results are encouraging but recall results of MobileNetV2 and Precision of GoogleNet are slightly low. In Very Mild Dementia, it was


found that GoogleNet had perfect precision but low recall (91.81%), while AlexNet had excellent precision (98.65%) and high recall (99.61%). EXTERNAL EVALUATION PERFORMANCE OF THE PROPOSED


MODEL ON THE OASIS DATASET Our experiments revealed that the AlexNet model trained with the ADAM solver achieved the highest performance. To validate this finding, we conducted an external


evaluation to assess the performance of this model using four classes of MRI images from the OASIS dataset. We found that the model achieved 98.2% classification accuracy on the OASIS


dataset using the AlexNet architecture and ADAM solver. We evaluated the model’s class-wise precision and recall (Figs. 14 and 15). Our findings revealed that the model achieved 100%


precision and recall for the mild-demented class and 100% precision and recall for the moderate-demented class. Additionally, it achieved 100% precision and 94.0% recall for the Non-Demented


class and 92.4% precision and 100% recall for the very mild-demented class. We evaluated the class-specific proposed model’s performance according to accuracy, precision, F1- score, recall,


specificity, and misclassification rate (Table 10; Fig. 11). For the mild-demented class, we achieved 100.00% Accuracy, 100.00% Precision, 100.00% F1-Score, 100.00% Recall, 100.00%


Specificity, and a Misclassification Rate of 0. For the Moderate-Demented class, we achieved 100.00% Accuracy, 100.00% Precision, 100.00% F1-Score, 100.00% Recall, 100.00% Specificity, and a


Misclassification Rate of 0. For the Non-Demented class, we achieved 98.18% Accuracy, 100.00% Precision, 96.91% F1 Score, 94.01% Recall, 100.00% Specificity, and a Misclassification Rate of


0.018214936. For the Very Mild-Demented class, we achieved 98.24% Accuracy, 92.42% Precision, 96.06% F1-Score, 100.00% Recall, 97.76% Specificity, and a Misclassification Rate of


0.017574692. The class-specific proposed model demonstrated remarkable performance across various dementia classes, achieving high accuracy, precision, recall, and specificity. In Fig. 16,


each bar reflects the corresponding metric for each class, with 100.00% values for most metrics in mild demented and moderate demented classes, and slightly lower values for non-demented and


very mild demented. Specifically, non-demented has a Recall of 94.01% and F1 Score of 96.91%, while Very Mild Demented shows a Precision of 92.42% and F1 Score of 96.06%. PERFORMANCE


COMPARISON OF PROPOSED MODEL ON MULTIPLE DATASETS In this research paper, we proposed a novel approach for early detection of Alzheimer’s disease using MRI scans and transfer learning


techniques with pre-trained CNNs. Our study focused on three distinct CNN architectures, AlexNet, GoogleNet, and MobileNetV2, trained using three different solvers: SGDM, ADAM, and RMSPROP.


We tested the models on two datasets: the Kaggle dataset and the OASIS dataset. The choice of solver significantly influenced the performance of the models, as the ADAM solver consistently


yielded the highest accuracy across all architectures. SGDM and RMSPROP also produced competitive results but generally lagged behind ADAM. The AlexNet model achieved the highest accuracy of


99.4% on the Kaggle dataset when trained with the ADAM solver as shown in Table 10, followed by the GoogleNet model with an accuracy of 98.2% and the MobileNetV2 model with an accuracy of


96.5%. The class-wise evaluation of the models revealed their ability to classify different stages of Alzheimer’s disease accurately. In our research, we identified several constraints,


including the dataset size, and the model’s generalizability to different datasets. To enhance our evaluation, we utilized an external dataset called OASIS, which allowed us to assess how


well our proposed model performs on unknown MRI images. The results from this new dataset were encouraging. We compared our findings with the previous studies, which utilized the same


datasets from Kaggle and OASIS as shown in Tables 11 and 12. Our results were indeed encouraging. The results align closely with the expected outcomes, showcasing the model’s effectiveness.


The performance can be significantly enhanced by meticulously fine-tuning the model, adjusting parameters such as the number of epochs and initial learning rate, and, most importantly,


identifying the optimal solver. In the context of Alzheimer’s disease detection, the ADAM optimizer stands out due to its ability to adaptively adjust the learning rate for each parameter.


This feature enhances the model’s ability to learn from complex data, such as MRI images, by efficiently handling sparse gradients that often occur in medical imaging tasks. The proposed


model is designed to assist doctors in automatic disease detection. These findings can enhance clinical Alzheimer’s disease detection by providing more accurate and reliable models for


identifying the disease in patients. Improved detection methods can lead to earlier diagnosis and better treatment strategies. LIMITATIONS While our proposed approach for Alzheimer’s disease


detection using transfer learning and pre-trained CNN models has shown promising results, several limitations must be acknowledged. Our models rely on pre-trained networks, which were


initially trained on large-scale datasets unrelated to medical images. This may result in a suboptimal feature extraction for MRI data compared to training CNNs from scratch on medical


datasets, although the latter would require much larger labeled datasets. Although the model performed well on the Kaggle and OASIS datasets, the generalizability of the model to other MRI


datasets remains untested. MRI data from different sources might vary due to differences in imaging protocols, patient demographics, and scanner settings, which could affect model


performance. Despite using transfer learning to reduce the computational cost, the fine-tuning process still requires significant computational power and time for training, especially when


testing different solvers and models. Although the OASIS dataset is a useful tool for detecting Alzheimer’s disease, it contains built-in restrictions that could affect how broadly


applicable, our findings are: * 1. 1. Imaging techniques: The standard techniques used to capture images in this database may differ from those used in other imaging datasets or in clinical


practice. When evaluated on fresh and varied datasets, models may perform differently due to variations in the SUBJECTS involved, such as scanners, resolution, and imaging sequences. * 2. 2.


Demographic variability: When using a resource like OASIS, researchers are limited to using data taken from a group of people who fall into particular age and geographic categories. This


could introduce bias into the model, making it potentially inaccurate. To address these limitations and improve the robustness of the model, we have taken the following steps: * 3. 1. Use of


multiple datasets: yo provide preliminary external validation, we tested the model’s performance across various imaging methods and demographics using two datasets (Kaggle and OASIS). * 4.


2. Transfer learning with pretrained models: to take use of a variety of features and optimization strategies, we used transfer learning with three pretrained models (AlexNet, GoogleNet,


MobileNetv2) and solvers (ADAM, SGDM, and RMSprop). ADVANTAGES Our research provides several key advantages that distinguish it from previous work in Alzheimer’s disease detection. Our use


of transfer learning with fine-tuned CNN architectures resulted in a high classification accuracy, with AlexNet achieving 99.4% on the Kaggle dataset, outperforming several state-of-the-art


models. By utilizing pre-trained models, we significantly reduced the need for extensive computational resources and time, compared to training models from scratch. This method is quicker to


put into action and is more geared towards researchers without any available resources. We compared the effectiveness of different solvers (ADAM, SGDM, RMSprop) with three different CNN


architectures and showed how solver selection helps in improving the performance of the model. This method could be modified to suit other medical image classification tasks apart from the


detection of Alzheimer’s disease, hence making it quite useful in medical diagnostics. BROADER IMPACT STATEMENT The proposed model has significant implications for real-world clinical


applications, particularly in under-resourced settings: * 1. Adoption in resource-limited environments: the suggested model’s computational simplicity, which is attained by combining


transfer learning with lightweight architectures like MobileNetv2, makes it ideal for deployment in environments with constrained computational resources. Utilizing pretrained models reduces


the requirement for intensive local training, which further reduces computational expenses. * 2. Scalability and accessibility: the model can be used in a variety of clinical settings,


including ones with varying imaging methods and demographic characteristics, as evidenced by its flexibility with regard to different datasets. This scalability guarantees that a variety of


healthcare facilities can use the system. * 3. Reducing barriers to adoption: the model’s open-source framework and low computational needs enable its adoption in resource-constrained


environments, allowing healthcare providers to implement it without major infrastructure upgrades. * 4. Future directions for impact: in order to facilitate clinical acceptance, future


research endeavors may concentrate on developing intuitive user interfaces and incorporating the model into current diagnostic procedures. Furthermore, working together with physicians in


contexts with limited resources may yield insightful input that helps improve the system for real-world use. These considerations highlight the potential of the model to make a meaningful


impact in real-world healthcare. FUTURE WORK The work presented here may be developed in several ways in order to make it more solid and applicable to other patient groups. Future studies


should proceed to the verification of the models proposed on other MRI datasets from other centers and countries in order to evaluate the efficacy and applicability of the model in various


imaging settings. In addition to MRI, imaging modalities such as PET or CT can be combined in a muli-modal deep learning framework which will help increase the sensitivity of diagnosing


Alzheimer’s disease by adding more information through other imaging techniques. Implementation of more modern deep learning architectures such as EfficientNet or Vision Transformers (ViTs)


are also possible to enhance the accuracy level and further cut down the training period since they are efficient and highly scale up for images classification. Future work could focus on


optimizing the model for real-time applications in clinical settings, where time efficiency and interpretability are critical for providing timely diagnoses. CONCLUSION This research


presented a new concept in classifying Alzheimer’s disease (AD) in four classes, namely mild-demented, moderate-demented, non-demented, and very mild-demented. Employing transfer learning


and adjusting three pre-trained convolutional neural networks AlexNet, MobileNetV2, and GoogleNet, the classification accuracy was very high with the pre-trained AlexNet model with ADAM


solver achieving the highest accuracy. This high accuracy suggests that our approach could facilitate earlier and more accurate diagnosis of Alzheimer’s disease, which is critical for timely


intervention. Earlier detection of AD allows for prompt treatment and management strategies, potentially improving patient outcomes by slowing disease progression. The outcomes of this


research have substantial clinical implications. For example, very high classification accuracy in the earlier stages of dementia, such as Very-Mild-Demented and Mild-Demented, implies that


such a model may be operationalized in the clinical settings to assist medical practitioners in making an accurate and timely diagnosis of Alzheimer’s disease prior to what is currently


possible, thereby improving the treatment interventions offered. Whereas AD is at its most challenging to diagnose at incipient stages due to various factors, the present approach provides a


consistent and dependable approach that minimizes the chances of misdiagnosis and helps in preventing and improving the treatment of various disorders, hence increasing the well-being of


patients. Our methodology, in addition to Alzheimer’s disease, seems to have a good purpose for additional medical conditions, mostly due to the characteristics of medical imaging. This is


because transfer learning can easily incorporate any pre-trained models, which makes it possible for use even in other degenerative conditions likely Parkinson’s disease or even various


other imaging techniques such as CT or PET scans. Further fine-tuning would be needed to adapt the models to the specifics of each condition and imaging modality, but the general methodology


remains applicable across various healthcare diagnostics. In relation to the current methodologies in the field, our model boasts remarkable benefits in terms of both precision and


processing time. Because we employed transfer learning technique rather than training models afresh, we cut down operational expenses and yet comparable accuracy was retained. This not only


guarantees less time consumption when making the models but also makes the incorporation of artificial intelligence in medicine less challenging to the physicians who have little computing


power. Given the superior performance coupled with ease of use, it is conceivable that our model will perform better than the conventional machine learning models, which depend on feature


engineering and large data sets. As we improve our approach, the next steps in the work will concentrate on cross-dataset generalization of the models, enhancing potential analysis of the


models through explainable AI, and implementing real-time clinical deployment to increase diagnostic capabilities and enhance patient care. DATA AVAILABILITY The dataset and simulation files


used during the current study are available from the corresponding author upon request. CODE AVAILABILITY Current code version: V1.0, Permanent link to code/repository used for this code


version: https://github.com/KainatSaher/Alzheimer/blob/main/task.m, Software code languages, tools and services used: Matlab, Compilation requirements, operating environments, and


dependencies: Core i7-9850H Precision 5540, 32GB RAM, Microsoft Windows, Support email for questions: [email protected] and [email protected]. REFERENCES * Mehmood, A. et al. A


transfer learning approach for early diagnosis of Alzheimer’s disease on MRI images. _Neuroscience_ 460, 43–52 (2021). Article  CAS  PubMed  MATH  Google Scholar  * Association, A.


Alzheimer’s disease facts, and figures include a special report on Alzheimer’s detection in the primary care setting: connecting patients and physicians. _Alzheimer’s Dement._ 15, 321–387


(2019). MATH  Google Scholar  * Picón, E. et al. Does empirically derived classification of individuals with subjective cognitive complaints predict dementia. _Brain Sci._ 9, 314 (2019).


Article  PubMed  PubMed Central  MATH  Google Scholar  * Afzal, S., Maqsood, M., Nazir, F., Khan, U. & Song, O. A data augmentation-based framework to handle class imbalance problem for


Alzheimer’s stage detection. _IEEE Access._ 7, 115528–115539 (2019). Article  MATH  Google Scholar  * Schlaggar, C. N. L., Rosario, O. L. D., Morris, J. C., Ances, B. M. & Schlaggar, B.


L. Adaptation of the clinical dementia rating scale for adults with down syndrome. _J. Neurodev. Disord_. 11, 1–10 (2019). MATH  Google Scholar  * Umbach, G. et al. Time cells in the human


hippocampus and entorhinal cortex support episodic memory. _Proc. Natl. Acad. Sci._ 117, 28463–28474 (2020). Article  ADS  CAS  PubMed  PubMed Central  MATH  Google Scholar  * Mehmood, S. et


al. Malignancy detection in lung and Colon histopathology images using transfer learning with Class Selective Image Processing. _IEEE Access._ 10, 25657–25668 (2022). Article  MATH  Google


Scholar  * Taleb, N. et al. Ovary Cancer Diagnosing Empowered with Machine Learning. International Conference on Business Analytics for Technology and Security (ICBATS) 2022. 1–6. (2022). *


Nasir, M. U. et al. IoMT-Based Osteosarcoma Cancer detection in histopathology images using transfer learning empowered with Blockchain, Fog Computing, and Edge Computing. _Sensors_ 22, 5444


(2022). Article  ADS  PubMed  PubMed Central  MATH  Google Scholar  * Rahman, A. et al. Advance genome disorder prediction Model Empowered with Deep Learning. _IEEE Access._ 10, 70317–70328


(2022). Article  Google Scholar  * Rahman, A. et al. IoMT based Mitochondrial and multifactorial genetic inheritance disorder prediction using machine learning. _Comput. Intell._ 650742.


(2022). * Naseer, A. et al. Refining Parkinson’s neurological disorder identification through deep transfer learning. _Neural Comput._ 32, 839–854 (2019). Article  MATH  Google Scholar  *


Naz, S. et al. Urdu Nastaliq recognition using convolutional recursive deep learning. _Neurocomputing_ 243, 80–87 (2017). Article  MATH  Google Scholar  * Razzak, M. I. Malarial parasite


classification using recurrent neural network. _Int. J. Image Process._ 9, 69 (2015). MATH  Google Scholar  * Razzak, M. I., Naz, S. & Zaib, A. Deep learning for medical image


processing: overview, challenges, and the future. _Classif. BioApps_. 4, 323–350 (2018). Article  MATH  Google Scholar  * Razzak, M. I. & Naz, S. Microscopic blood smear segmentation and


classification using deep contour aware CNN and extreme machine learning. IEEE Conference on Computer Vision and Pattern Recognition Workshops 49–55. (2017). * Baron, J. C. et al. In vivo


mapping of gray matter loss with Voxel-based morphometry in mild Alzheimer’s disease. _Neuroimage_ 14, 298–309 (2001). Article  CAS  PubMed  MATH  Google Scholar  * Asif, S. et al. Improving


Effectiveness of Different Deep Transfer Learning-Based Models for Detecting Brain Tumors From MR Images. _IEEE Access_. 10, 34716–34730. (2022). https://doi.org/10.1109/ACCESS.2022.3153306


* Asif, S., Zhao, M., Tang, F. & Zhu, Y. An enhanced deep learning method for multi-class brain tumor classification using deep transfer learning. _Multimed. Tools Appl._ 82 (20),


31709–31736 (2023). Article  Google Scholar  * Khan, S. U., Rehman, M., Zhao, S., Asif & Chen, X. Hybrid-NET: a fusion of DenseNet169 and advanced machine learning classifiers for


enhanced brain tumor diagnosis. _Int. J. Imaging Syst. Technol._ 34 (1), e22975 (2024). Article  Google Scholar  * Asif, S., Zhao, M., Chen, X. & Zhu, Y. BMRI-NET: a deep stacked


ensemble model for multi-class brain tumor classification from MRI images. _Interdiscipl. Sci. Comput. Life Sci._ 15 (3), 499–514 (2023). Google Scholar  * Zhang, F. et al. Accelerating


hyperparameter tuning in machine learning for Alzheimer’s disease with high performance computing. _Front. Artif. Intell._ 4, 798962 (2021). Article  PubMed  PubMed Central  Google Scholar 


* Hedayati, R., Khedmati, M. & Taghipour-Gorjikolaie, M. Deep feature extraction method based on ensemble of convolutional auto encoders: application to Alzheimer’s disease diagnosis.


_Biomed. Signal Process. Control_. 66, 102397 (2021). Article  Google Scholar  * Balaji, P., Chaurasia, M. A. & Bilfaqih, S. M. Anandhavalli Muniasamy, and Linda Elzubir Gasm Alsid.


Hybridized deep learning approach for detecting Alzheimer’s disease. _Biomedicines_ 11 (1), 149. (2023). * Plant, C. Automated detection of brain atrophy patterns based on MRI for the


prediction of Alzheimer’s disease. _Neuroimage_ 50, 162–174 (2010). Article  PubMed  MATH  Google Scholar  * Gray, R., Aljabar, P., Heckemann, R. A., Hammers, A. & Rueckert, D. Random


forest-based similarity measures for multi-modal classification of Alzheimer’s disease. _Neuroimage_ 65, 167–175 (2013). Article  PubMed  Google Scholar  * Dahshan, E. S. A. E., Hosny, T.


& Salem, A. B. M. Hybrid intelligent techniques for MRI brain images classification. _Digit. Signal. Process. Rev. J._ 20, 433–441 (2010). Article  ADS  MATH  Google Scholar  * Ahmed, S.


Ensembles of patch-based classifiers for diagnosis of Alzheimer diseases. _IEEE Access._ 7, 73373–73383 (2019). Article  Google Scholar  * Hong, X. Predicting Alzheimer’s disease using


LSTM. _IEEE Access._ 7, 80893–80901 (2019). Article  Google Scholar  * Gupta, A., Ayhan, M. S. & Maida, A. S. Natural image bases to represent neuroimaging data. _Int. Conf. Mach.


Learn._ 28, 2024–2031 (2013). MATH  Google Scholar  * Suk, H., Lee, S. & Shen, D. Deep ensemble learning of sparse regression models for brain disease diagnosis. _Med. Image Anal._ 37,


101–113 (2017). Article  PubMed  PubMed Central  MATH  Google Scholar  * Islam, J. & Zhang, Y. A novel deep learning based multi-class classification method for Alzheimer’s disease


detection using brain MRI data. _Lect. Notes Comput. Sci._ 10654, 213–222 (2017). Article  MATH  Google Scholar  * Guerrero, R., Wolz, R., Rao, A. W. & Rueckert, D. Manifold population


modeling as a neuro-imaging biomarker: application to adni and adni-go. _Neuroimage_ 94, 275–286 (2014). Article  CAS  PubMed  Google Scholar  * Gorji, H. T. & Kaabouch, N. A deep


learning approach for diagnosis of mild cognitive impairment based on MRI images. _Brain Sci._ 9, 217 (2019). Article  PubMed  MATH  Google Scholar  * Hosseini-Asl, E., Keynton, R. &


El-Baz, A. Alzheimer’s disease diagnostics by adaptation of 3d convolutional network._ IEEE International Conference on Image Processing (ICIP)_ 126–130. (2016). * He, K., Zhang, X., Ren, S.


& Sun, J. Deep residual learning for image recognition. In_ Proceedings of the IEEE conference on computer vision and pattern recognition_ 770–778. (2016). * Hazarika, R. A., Kandar, D.


& Maji, A. K. An Experimental Analysis of Different deep Learning Based Models for Alzheimer’s Disease Classification Using Brain Magnetic Resonance Images. _J. King Saud Univ. Comput.


Inf. Sci._ (2021). * Shanmugam, J. V., Duraisamy, B., Simon, B. C. & Bhaskaran, P. Alzheimer’s disease classification using pre-trained deep networks. _Biomed. Signal Process. Control_.


71, 103217 (2022). Article  Google Scholar  * Battineni, G., Hossain, M. A., Chintalapudi, N. & Amenta, F. _Alzheimer’s Disease Classification Using Feed Forwarded Deep Neural Networks


for Brain MRI Images_. Predictive Analytics of Psychological Disorders in Healthcare. Springer, 269–283. (2022). * Sarvesh Dubey. Alzheimer’s Dataset.


https://www.kaggle.com/datasets/tourist55/alzheimers-dataset-4-class-of-images Accessed 10 June 2022. (2022). * Marcus, D. S., Fotenos, A. F., Csernansky, J. G., Morris, J. C. & Buckner,


R. L. Open access series of imaging studies: longitudinal MRI data in nondemented and demented older adults. _J. Cogn. Neurosci._ 22, 2677–2684 (2010). Article  PubMed  PubMed Central 


Google Scholar  * Krishna, S. K. & Kalluri, H. K. Deep learning and transfer learning approaches for image classification. _Int. J. Recent. Technol. Eng._ 7, 427–432 (2019). MATH  Google


Scholar  * MathWorks. https://www.mathworks.com/help/deeplearning/ref/trainingoptions.html Accessed 01 Aug 2022. (2022). * Sethi, M., Rani, S., Singh, A. & Mazón, J. L. V. A CAD system


for alzheimer’s disease classification using neuroimaging MRI 2D slices._ Comput. Math. Methods Med._, 2022. (2022). * Mggdadi, E., Al-Aiad, A., Al-Ayyad, M. S. & Darabseh, A. Prediction


Alzheimer’s disease from MRI images using deep learning. In_ 2021 12th International Conference on Information and Communication Systems (ICICS)_. IEEE, 120-125. (2021) * Suganthe, R. C. et


al. _Multiclass Classification of Alzheimer’s Disease Using Hybrid deep Convolutional Neural Network__ NVEO Nat. Volatiles Essent. Oils J._ 145–153. (2021). * Murugan, S. et al. DEMNET: a


deep learning model for early diagnosis of Alzheimer diseases and dementia from MR images. _Ieee Access._ 9, 90319–90329 (2021). Article  MATH  Google Scholar  * Battineni, G., Chintalapudi,


N., Amenta, F. & Traini, E. Deep Learning Type Convolution Neural Network Architecture for Multiclass Classification of Alzheimer’s Disease._ Bioimaging_. 209–215. (2021). * Yildirim,


M. & Cinar, A. Classification of Alzheimer’s Disease MRI Images with CNN Based Hybrid Method._ Ingénierie Syst. Inf._ 25(4), 413–418. (2020). Download references FUNDING The Publication


of this study was funded by the Weill Cornell Medicine- Qatar Health Sciences Library. AUTHOR INFORMATION AUTHORS AND AFFILIATIONS * Faculty of Information Technology and Computer Science,


University of Central Punjab, Lahore, 54000, Pakistan Muhammad Zahid Hussain * Department of Computer Engineering, COMSATS University Islamabad, Sahiwal Campus, Sahiwal, 57000, Pakistan


Tariq Shahzad * Department of Computer Science, Bahria University Lahore Campus, Lahore, 54000, Pakistan Shahid Mehmood * Department of Computer Science, University of Engineering and


Technology, Lahore, 54000, Pakistan Kainat Akram * Department of Software, Faculty of Artificial Intelligence and Software, Gachon University, Seongnam-si, 13120, Republic of Korea Muhammad


Adnan Khan * Abu Dhabi University, Abu Dhabi, UAE Muhammad Usman Tariq * AI Center for Precision Health, Weill Cornell Medicine-Qatar, Doha, Qatar Arfan Ahmed Authors * Muhammad Zahid


Hussain View author publications You can also search for this author inPubMed Google Scholar * Tariq Shahzad View author publications You can also search for this author inPubMed Google


Scholar * Shahid Mehmood View author publications You can also search for this author inPubMed Google Scholar * Kainat Akram View author publications You can also search for this author


inPubMed Google Scholar * Muhammad Adnan Khan View author publications You can also search for this author inPubMed Google Scholar * Muhammad Usman Tariq View author publications You can


also search for this author inPubMed Google Scholar * Arfan Ahmed View author publications You can also search for this author inPubMed Google Scholar CONTRIBUTIONS Study conception and


design: M.Z. Hussain, T. Shahzad, S. Mehmood and K. Akram; data collection: M.Z. Hussain, T. Shahzad and K. Akram; analysis and interpretation of results: T. Shahzad, M.A. Khan, M.U. Tariq,


and A. Ahmed; draft manuscript preparation: M.Z. Hussain, M.A. Khan, M.U. Tariq and A. Ahmed. All authors reviewed the results and approved the final version of the manuscript. CORRESPONDING


AUTHORS Correspondence to Muhammad Adnan Khan or Arfan Ahmed. ETHICS DECLARATIONS COMPETING INTERESTS The authors declare no competing interests. ADDITIONAL INFORMATION PUBLISHER’S NOTE


Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. ELECTRONIC SUPPLEMENTARY MATERIAL Below is the link to the electronic


supplementary material. SUPPLEMENTARY MATERIAL 1 SUPPLEMENTARY MATERIAL 2 SUPPLEMENTARY MATERIAL 3 RIGHTS AND PERMISSIONS OPEN ACCESS This article is licensed under a Creative Commons


Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original


author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the


article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use


is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit


http://creativecommons.org/licenses/by/4.0/. Reprints and permissions ABOUT THIS ARTICLE CITE THIS ARTICLE Hussain, M.Z., Shahzad, T., Mehmood, S. _et al._ A fine-tuned convolutional neural


network model for accurate Alzheimer’s disease classification. _Sci Rep_ 15, 11616 (2025). https://doi.org/10.1038/s41598-025-86635-2 Download citation * Received: 05 June 2024 * Accepted:


13 January 2025 * Published: 04 April 2025 * DOI: https://doi.org/10.1038/s41598-025-86635-2 SHARE THIS ARTICLE Anyone you share the following link with will be able to read this content:


Get shareable link Sorry, a shareable link is not currently available for this article. Copy to clipboard Provided by the Springer Nature SharedIt content-sharing initiative KEYWORDS *


Classification * Convolutional neural networks (CNN) * Solvers/optimizers * Alzheimer’s disease detection * Transfer learning