Abstract
Sporadic Alzheimer’s disease (AD) is a complex neurological disorder characterized by many risk loci with potential associations with different traits and diseases. AD, characterized by a progressive loss of neuronal functions, manifests with different symptoms such as decline in memory, movement, coordination, and speech. The mechanisms underlying the onset of AD are not always fully understood, but involve a multiplicity of factors. Early diagnosis of AD plays a central role as it can offer the possibility of early treatment, which can slow disease progression. Currently, the methods of diagnosis are cognitive testing, neuroimaging, or cerebrospinal fluid analysis that can be time-consuming, expensive, invasive, and not always accurate. In the present study, we performed a genetic correlation analysis using genome-wide association statistics from a large study of AD and UK Biobank, to examine the association of AD with other human traits and disorders. In addition, since hippocampus, a part of cerebral cortex could play a central role in several traits that are associated with AD; we analyzed the gene expression profiles of hippocampus of AD patients applying 4 different artificial neural network models. We found 65 traits correlated with AD grouped into 9 clusters: medical conditions, fluid intelligence, education, anthropometric measures, employment status, activity, diet, lifestyle, and sexuality. The comparison of different 4 neural network models along with feature selection methods on 5 Alzheimer’s gene expression datasets showed that the simple basic neural network model obtains a better performance (66% of accuracy) than other more complex methods with dropout and weight regularization of the network.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Introduction
Alzheimer’s disease (AD), the most common form of dementia with 60–70% of total cases, has an onset over 65 years of age. AD is a progressive, incurable neurodegenerative disease characterized by a gradual decline in cognition, memory, and thinking (Kumar et al. 2022).
Currently, the therapeutic approaches offer limited results in symptoms and progression of the disease, but there is not a definitive treatment (Lane et al. 2018; Brookmeyer et al. 1998). Thus, a great challenge is to develop novel methods for early detection, in order to decrease or prevent disease progression.
Sporadic cases of AD originate from complex genetic architecture that involves many risk loci with small single influences (Tesi et al. 2021). Indeed, AD-associated SNPs were common with other medical conditions and human traits (Tesi et al. 2021). Genetic correlation analysis based on phenome-wide screening generates novel hypotheses related to risk conditions and comorbid events of AD (Liu and Crawford 2022). Rapid rising of high-throughput technologies (e.g., microarray and next generation sequencing) over the past years has resulted in a significant recent increase of novel computational methods of many diseases including AD (van IJzendoorn et al. 2019).
Genome-wide association studies (GWAS) are essential tools to address this complexity and opening up new therapeutic challenges. Previous GWAS in AD demonstrated the association between immune system and lipid metabolism and identified several genes and genetic variants related to lipid metabolism (Baloni et al. 2022; Kunkle et al. 2019). In addition, recent studies demonstrated that cardiovascular and life style factors could increase the risk of AD development as well as diabetes, obesity, and hypertension (Broce et al. 2019; Desikan et al. 2015). Adewuyi et al. showed an association between gut and brain, suggesting a potential genetic susceptibility of gastrointestinal disorders with AD’s risk (Adewuyi et al. 2022). Another group with AD genetically associated traits is related to the dietary habits (Squitti et al. 2014). For example, a lower incidence of AD has been reported in subjects following a Mediterranean diet (Gardener et al. 2012). The lack of micronutrients in the diet such as vitamins B1, C, and folate has been related to cognitive decline in elderly people (Solfrizzi et al. 2011). However, genetic correlation is not a diagnostic tool, but a method to establish the genetic similarity between complex traits (van IJzendoorn et al. 2019).
Most machine learning algorithms proposed for AD classification are based on phenotypic data such as imaging, and few studies use genetic data (Lee et al. 2018). Since 2012, deep learning, a branch of machine learning, has been shown good performance in several areas outside biological problems (Abiodun et al. 2018). Lately, with this assumption, several studies have demonstrated the potential of deep learning to address also biological questions as diagnostic tools (Zhu et al. 2020a; Rukhsar et al. 2022). Deep learning indicates machine learning algorithms that are composed of deep neural networks. Different studies are based on the biological application of neural networks and few studies are focused on how the network architectures could improve the performance of models (Bellot et al. 2018; Yu et al. 2019; Wilentzik Müller and Gat-Viks 2020). There are different challenges in obtaining the optimal model of neuronal network in a classification problem (Rukhsar et al. 2022). Mainly, the performance of neural network models can be influenced by the amount of data that could generate overfitting problems (Esteva et al. 2019). To resolve this problem, researchers have developed several techniques such as regularization methods, dropout, class balanced, and feature selection (Esteva et al. 2019). However, there is not a best general method because it is difficult to obtain the performance of each network architecture using a same dataset (Nusrat and Jang 2018; Moolayil 2019). In addition, unlike imaging or text data, classifiers based on neural networks are still novel in gene expression analysis (Hanczar et al. 2022).
In this work, using genome-wide associations statistics from public datasets, we explored a genetic correlation between AD and many other human traits. In addition, we considered and compared different methods to reduce the overfitting using gene expression profiles of Alzheimer patients derived by 5 published datasets. The dimensionality of the gene expression profiles was reduced with principal component analysis (PCA) as it transforms the features into a lower dimensional space considering the relationships between the features. Furthermore, after a procedure to avoid unbalanced classes, we evaluated 5 network models considering the performance of the classifier with accuracy, sensitivity, and specificity.
The aims of our study are (i) to identify the mechanisms of AD genetic liability that could be connected to different human traits, (ii) to explore artificial neural network models for AD diagnosis, (iii) comparison of different models based on artificial neural network using gene expression profiles of AD patients. In particular, the present findings could (i) suggest future studies to assess the impact of several traits with AD risk and (ii) open new potential frontiers in the study of AD.
Materials and methods
Genome-wide association studies
GWAS summary statistics of AD were downloaded from GWAS Atlas (Jansen et al. 2019). The dataset consists of European cohort and have participants of both sexes. We used the summary statistics of 71,880 AD cases and 383,378 controls (Jansen et al. 2019).
We also considered genome-wide association statistics of other phenotypes and diseases, derived from the UK Biobank (UKB) (Bycroft et al. 2018). This dataset was downloaded from (http://www.nealelab.is/uk-biobank/, accessed on 4 February 2022).
The UK Biobank enrolled approximately 500,000 participants aged 40–69 years, of both sexes from the UK (Bycroft et al. 2018). UKB participants were analyzed for a wide range of phenotypic information such as diet, educational status, cognitive function, social activities, health status, and other phenotypes.
Data quality control was performed separately for each data set (AD and UKB).
In particular, we calculated SNP-heritability for AD and UKB phenotypes and considered for further analyses only phenotypes with SNP-heritability z > 4.
In addition, AD and UKB genome-wide association statistics were processed by removing SNP with a minor allele frequency (MAF) < 1%. More detailed descriptions of quality control step are available at https://github.com/Nealelab/UK_Biobank_GWAS.
Genetic correlation analysis
We estimated the genetic correlation analysis between the AD phenotypes and the other phenotypes included in UKB. To perform the genetic correlation, we used the package linkage disequilibrium score regression (LDSC) (Accessed on May 2022; LD Score: https://github.com/bulik/ldsc) (). This is a method that performs the linkage disequilibrium (LD) mechanism to calculate the distribution of effect sizes for each SNP, thus assigning the score and the type of correlation between phenotypes.
We used SNPs present in the HapMap 3 reference panel, and as reference data, the individuals of European ancestry from the 1000 Genomes Project. We performed the genetic correlation analysis between AD phenotype with UKB traits with SNP-based heritability z score > 4, in line with the guidelines of the LDSC developers (). We considered statistically significant genetic correlations as those that had an FDR less than 0.05.
Gene expression data
Five publicly available datasets of gene expression profiles of Alzheimer patients (GSE1297, GSE5281, GSE36980, GSE29378, and GSE48350) were downloaded from the Gene Expression Omnibus (GEO). These datasets contain the gene expression profiles of hippocampus of Alzheimer patients as this brain area is involved in the early stages of disease (Quarato et al. 2022). We considered hippocampus because we suppose that this part of brain plays a fundamental role in different traits associated with AD. Table 1 shows the number of samples for Alzheimer patients and controls of each dataset.
Training and testing sets
We split each GEO dataset into two sets: training and testing sets. Neural network was trained using the training set and the testing set to test the model: 70% of the original dataset for the training and 30% for the testing.
To avoid unbalanced datasets, namely a number of samples in a class (e.g., Alzheimer) is greater than another class (e.g., control), we performed a random oversampling. This approach balances the minority class with majority class.
In addition, we standardized each dataset separately converting the data distribution per feature to a normal distribution.
Feature selection
The presence of unrelated features in the dataset can decrease the accuracy of the models. During the feature selection step, we selected a subset of features that contribute to reduce overfitting (Moolayil 2019). PCA was used to decrease the dimensionality of datasets and to identify the key components on the standardized training data (Moolayil 2019). The number of basic components according to the training data was defined considering 95% of the variance of the data. We considered the same components in both the training and testing data.
Neural network models
Similar to other machine learning methods, neural network model is composed of (i) the training step where the parameters of the network are estimated from a given training dataset and (ii) the testing step that applies the trained network to predict the classes of new input data.
The neurons in our models are organized in 4 layers where all nodes in a specific layer must be connected to all the nodes in the next layer.
In all models, the four layers were defined as follows: the input (first) layer consists of a number of neurons equal to the number of features (i.e., key components derived by PCA). The first hidden layer characterized by 17 neurons and the second hidden layer of 8 neurons. The output layer returns the predicted class.
Each neuron calculates a weighted sum of its inputs and applies an activation function. We used a model rectified linear unit (ReLU) at each node of the network for all models (Glorot et al. 2011). It is the most common activation function used and it generates 0 as output for negative inputs, following the formula:
A sigmoid activation function is used for the output layer to identify the class to be predicted for all model:
where x is a feature vector.
The Adam stochastic gradient descent optimization is used as optimizer algorithm in all our models to train the network (Kingma and Ba 2014). It assigns the parameters that decreases the loss function. Gradient descent uses the first derivate of the activation function to modify the parameters of the model. Specifically, Adam changes the weights of the model in the training set iteratively to maximize a particular class (Kingma and Ba 2014).
Table 2 shows the parameters considered for all 4 models. We evaluated the parameters as setted in Izadkhah 2022.
We tested 4 different neural network architectures for the binary classification problem which differ in loss function, metrics, dropout, and weight regularization.
The loss functions to minimize during the training that we used are binary cross-entropy or mean squared logarithmic error. This function is used to evaluate the classifier through the model error and to quantify how the model fits (Rengasamy et al. 2020).
It is a common strategy to reduce the overfitting of neural network to add dropout or introduce a penalty (weight regularization).
A dropout can be used to decrease the overfitting of the model. The dropout consists in removing a random subset of nodes (Srivastava et al. 2014).
Table 3 shows the 4 different models of neural networks used.
Summarizing, the different 4 models are organized as follows:
-
First model: The first model consists of binary cross-entropy as loss function, adam as optimization algorithm, and the binary accuracy as metric. Binary cross-entropy determines the cross-entropy loss between the predicted classes and the true labels.
-
Second model: The second model consists of mean squared logarithmic error as loss function, adam as optimization algorithm, and the accuracy as metric. Mean squared logarithmic error is calculated between the true classes and predicted classes.
-
Third model: The third model consists of mean squared logarithmic error as loss function, adam as optimization algorithm, and as metric the accuracy. Dropout is applied between the second and third layer to reduce overfitting and the dropout rate is set to 0.5. Dropout consists of a random selection of a small number of nodes instead of all nodes changing by regularly the nodes in the training process (Kingma and Ba 2014).
-
Fourth model: The fourth model consists of mean squared logarithmic error as loss function, adam as optimization algorithm, and as metric the accuracy. Weight regularization is used to reduce overfitting and regularization hyper-parameter value is set to 0.001. Weight regularization is a method to reduce the overfitting by regulating the weight distribution adding a regularization expression to the cost function (loss function) (Maki 2019). Weight regularization to reduce the error is based on criterion in our study: it adds “summed squared weights” as penalty term to the loss function (Maki 2019).
In all models, we introduced an “early stopping function” presents in the package Keras (https://keras.io/callbacks/#earlystopping) that regularly checks loss values of testing data and stops the training process when there is not a significant improvement in the loss values of the testing data. The quantification of acceptable improvement is set to 0.005 and if there are not improvement of the loss values in at the last 5 interactions the training process will terminate. To reduce the time and memory activity, the model was trained with a batch size = 8 and run for a maximum of 200 epochs.
Finally, we compared the performance of the 4 models (sensitivity, specificity, and accuracy) in the testing set for each GEO dataset. It must be noted that neural networks are based on stochastic algorithms and so, the performance on the same data with same model can slightly differ. In order to obtain more realistic results, we calculated the average sensitivity, average specificity, and average accuracy running 10 times the same model.
The neural network model code was implemented in Python using the keras package (version 2.10).
Results
Genetic correlation
After quality control step, the number of SNPs in GWAS of AD is reduced from 13,367,299 to 9,736,043 SNPs. Out of 4000 phenotypes of UKB, only 957 passed the quality control.
Genetic correlation analysis can demonstrate if AD is influenced by external factors. We found 65 traits correlated with AD (Table 4).
Neural network models
As first aim of our work, we investigated if neural networks can be used as tool for Alzheimer diagnosis (i.e., classification Alzheimer vs control) considering 5 gene expression datasets. For each of these datasets, we explored different neural network models.
All neural network models consist of three layers. Input nodes equal to the number of input feature (i.e., key components derived by PCA). We used as hidden layer a number of 50% of input nodes. Being a binary classification, the models require an only output node. Figure 1 shows the described neural network.
We investigated 4 neural network models. As the most basic neural network structure, we examined a neural network that uses binary cross-entropy as loss function and binary accuracy as metric to evaluate the model in the training. The classification model was demonstrated to be more accurate in GSE5281 (accuracy 0.78, sensitivity 0.68, and specificity 0.88) achieving an overall average good performance in all datasets (accuracy 0.66, sensitivity 0.62, and specificity 0.712).
We then tested the classification using a model neural network based on mean squared logarithmic error as loss function and the accuracy as metric. The average performance of all GEO datasets showed a dramatic decrease: accuracy 0.546, sensitivity 0.524, and specificity 0.566. Similar results were obtained with the third model that used dropout: accuracy 0.554, sensitivity 0.492, and specificity 0.628.
A slight improvement was achieved with the weight regularization in the fourth model: accuracy 0.582, sensitivity 0.6, and specificity 0.566.
Table 5 shows the performance of the classifier for each GEO dataset.
Overall, the best performances were achieved with the first and fourth model (Fig. 2).
Discussion
Sporadic AD is the most common form of dementia. It is due to the effects of many risk loci with small single consequences. In the present study, we performed a genetic correlation analysis between genome-wide association statistics of AD derived by GWAS Atlas and human traits from UK Biobank.
We observed that AD was mainly associated with fluid intelligence score, medical conditions, diet, and activities. Regarding the diet, AD is positively associated with cereal and salt intake and inversely correlated with dried fruit and alcohol intake. Further studies should be performed to understand the potential beneficial effect of alcohol consumption and negative effect of salt intake.
Another macro-area with multiple AD genetically correlated phenotypes is anthropometric measurements: positively associated with standing height and inversely correlated with leg fat percentage (left), high light scatter reticulocyte count, leg fat percentage (right), and body mass index (BMI).
As hippocampus, a part of cerebral cortex, plays a central role in several traits that we found to be associated with AD, we applied to gene expression profiles of hippocampus of AD patients and artificial neural models.
Regarding the development of diagnostic tools for AD, we explored the role of artificial neural network based on gene expression of hippocampus of Alzheimer patients.
Artificial neural network, an emergent field of machine learning, is a computational model involving interconnected nodes inspired by neurons in the human brain to solve complex problems. It uses one or more hidden layers, an activation function and hyper-parameters to elaborate the input and generate the output.
Recent studies in bioinformatics have proposed the use of neural networks in molecular classification of diseases by gene expression and multi-omics data (Qiu et al. 2022; Shao et al. 2021). Many studies were focused on the comparison between artificial neuronal network and other machine learning methods, demonstrating that artificial neural networks are more flexible and work on different types of data (e.g., discrete or continuous data) (Esteva et al. 2019; Biganzoli et al. 1998; Zhu et al. 2020b).
However, few studies have been performed to evaluate different procedures to avoid overfitting and improve the performance of the artificial neuronal network considering gene expression datasets (Hanczar et al. 2022; Zhu et al. 2020b; Chen et al. 2016). This could be explained by the great number of hyper-parameters to test.
Our study compared 4 neural network models applied to gene expression datasets of Alzheimer, showing that the simple basic neural network model achieves a better performance than other more complex methods with dropout and weight regularization (accuracy 0.66, sensitivity 0.62, and specificity 0.712). However, increasing the size of the samples in the datasets, the model could further improve the performance and confirm these results. Indeed, the dataset size is a critical aspect that could influence the performance of models. Typically, larger datasets could lead to better performance and small datasets may generate overfitting (Prusa et al. 2015). Supervised machine learning methods also depend on the diversity and quality of the dataset to achieve good performances in generalization step (Leguy et al. 2021).
In line with our results, a previous study found that simple neural network models have obtained similar performance compared to other complex methods (Zhu et al. 2020b). Although the values of hyper-parameters used in this study are closely associated with our data, we can suggest the use of simple basic neural network for gene expression classification. In addition, loss function with binary cross-entropy seems to work with better performance than mean squared logarithmic error. Note, regularization methods seem to reduce the overfitting and work better than dropout procedures.
Conclusion
In conclusion, the present study with genetic correlation analysis suggested several mechanisms of AD that could be associated with different human traits. It can be grouped into 9 clusters: medical conditions, fluid intelligence, education, anthropometric measures, employment status, activity, diet, lifestyle, and sexuality. However, correlation analysis does not necessarily imply causation, namely the cause-and-effect relationship between two variables. In order to establish causality, it is necessary to conduct further studies that can identify cause-effect relationships more reliably. In addition, further studies should be conducted to fully understand the impact of SNPs on these relationships.
Related to neural network models in our study, we compared the most suitable schemes for artificial neuronal network applied to gene expression datasets of patients with Alzheimer. Our results showed that the simple basic neural network model achieved a better performance (66% of accuracy). To our knowledge, in literature, there was not similar research, and more studies are needed to completely define standard procedures to achieve more efficient results. It could be also interesting to explore more sophisticated deep neural networks also increasing the size of the datasets.
Data availability
All the summary statistics used in this study are available from the United Kingdom Biobank database (http://www.nealelab.is/uk-biobank). Gene expression profiles are available from Gene Expression Omnibus (Home—GEO—NCBI (nih.gov)).
References
Abiodun OI, Jantan A, Omolara AE, Dada KV, Mohamed NA, Arshad H (2018) State-of-the-art in artificial neural network applications: A survey. Heliyon. 4(11):e00938. https://doi.org/10.1016/j.heliyon.2018.e00938
Adewuyi EO, O’Brien EK, Nyholt DR, Porter T, Laws SM (2022) A large-scale genome-wide cross-trait analysis reveals shared genetic architecture between Alzheimer’s disease and gastrointestinal tract disorders. Commun Biol 5(1):691. https://doi.org/10.1038/s42003-022-03607-2
Baloni P, Arnold M, Buitrago L, Nho K, Moreno H, Huynh K, Brauner B, Louie G, Kueider-Paisley A, Suhre K, Saykin AJ, Ekroos K, Meikle PJ, Hood L, Price ND, Alzheimer’s Disease Metabolomics Consortium, Doraiswamy PM, Funk CC, Hernández AI, Kastenmüller G, Baillie R, Han X, Kaddurah-Daouk R (2022) Multi-Omic analyses characterize the ceramide/sphingomyelin pathway as a therapeutic target in Alzheimer’s disease. Commun Biol. 5(1):1074. https://doi.org/10.1038/s42003-022-04011-6
Bellot P, de Los CG, Pérez-Enciso M (2018) Can deep learning improve genomic prediction of complex human traits? Genetics 210(3):809–819. https://doi.org/10.1534/genetics.118.301298
Biganzoli E, Boracchi P, Mariani L, Marubini E (1998) Feed forward neural networks for the analysis of censored survival data: a partial logistic regression approach. Stat Med 17(10):1169–1186. https://doi.org/10.1002/(sici)1097-0258(19980530)17:10%3c1169::aid-sim796%3e3.0.co;2-d
Broce IJ, Tan CH, Fan CC, Jansen I, Savage JE, Witoelar A, Wen N, Hess CP, Dillon WP, Glastonbury CM, Glymour M, Yokoyama JS, Elahi FM, Rabinovici GD, Miller BL, Mormino EC, Sperling RA, Bennett DA, McEvoy LK, Brewer JB, Feldman HH, Hyman BT, Pericak-Vance M, Haines JL, Farrer LA, Mayeux R, Schellenberg GD, Yaffe K, Sugrue LP, Dale AM, Posthuma D, Andreassen OA, Karch CM, Desikan RS (2019) Dissecting the genetic relationship between cardiovascular risk factors and Alzheimer’s disease. Acta Neuropathol 137(2):209–226. https://doi.org/10.1007/s00401-018-1928-6
Brookmeyer R, Gray S, Kawas C (1998) Projections of Alzheimer’s disease in the United States and the public health impact of delaying disease onset. Am J Public Health 88(9):1337–1342. https://doi.org/10.2105/ajph.88.9.1337
Bulik-Sullivan B, Finucane HK, Anttila V, Gusev A, Day FR, Loh PR, ReproGen Consortium; Psychiatric Genomics Consortium; Genetic Consortium for Anorexia Nervosa of the Wellcome Trust Case Control Consortium 3, Duncan L, Perry JR, Patterson N, Robinson EB, Daly MJ, Price AL, Neale BM (2015) An atlas of genetic correlations across human diseases and traits. Nat Genet. 47(11):1236–41. https://doi.org/10.1038/ng.3406
Bycroft C, Freeman C, Petkova D, Band G, Elliott LT, Sharp K, Motyer A, Vukcevic D, Delaneau O, O’Connell J, Cortes A, Welsh S, Young A, Effingham M, McVean G, Leslie S, Allen N, Donnelly P, Marchini J (2018) The UK Biobank resource with deep phenotyping and genomic data. Nature 562(7726):203–209. https://doi.org/10.1038/s41586-018-0579-z
Chen Y, Li Y, Narayan R, Subramanian A, Xie X (2016) Gene expression inference with deep learning. Bioinformatics 32(12):1832–1839. https://doi.org/10.1093/bioinformatics/btw074
Desikan RS, Schork AJ, Wang Y, Thompson WK, Dehghan A, Ridker PM, Chasman DI, McEvoy LK, Holland D, Chen CH, Karow DS, Brewer JB, Hess CP, Williams J, Sims R, O’Donovan MC, Choi SH, Bis JC, Ikram MA, Gudnason V, DeStefano AL, van der Lee SJ, Psaty BM, van Duijn CM, Launer L, Seshadri S, Pericak-Vance MA, Mayeux R, Haines JL, Farrer LA, Hardy J, Ulstein ID, Aarsland D, Fladby T, White LR, Sando SB, Rongve A, Witoelar A, Djurovic S, Hyman BT, Snaedal J, Steinberg S, Stefansson H, Stefansson K, Schellenberg GD, Andreassen OA, Dale AM, Inflammation working group, IGAP and DemGene Investigators (2015) Polygenic Overlap Between C-Reactive Protein, Plasma Lipids, and Alzheimer Disease. Circulation. 131(23):2061–2069. https://doi.org/10.1161/CIRCULATIONAHA.115.015489
Esteva A, Robicquet A, Ramsundar B, Kuleshov V, DePristo M, Chou K, Cui C, Corrado G, Thrun S, Dean J (2019) A guide to deep learning in healthcare. Nat Med 25(1):24–29. https://doi.org/10.1038/s41591-018-0316-z
Gardener S, Gu Y, Rainey-Smith SR, Keogh JB, Clifton PM, Mathieson SL, Taddei K, Mondal A, Ward VK, Scarmeas N, Barnes M, Ellis KA, Head R, Masters CL, Ames D, Macaulay SL, Rowe CC, Szoeke C, Martins RN, AIBL Research Group (2012) Adherence to a Mediterranean diet and Alzheimer’s disease risk in an Australian population. Transl Psychiatry 2(10):e164. https://doi.org/10.1038/tp.2012.91
Glorot X, Bordes A, Bengio Y (2011) Deep sparse rectifier neural networks. In: Proceedings of the fourteenth international conference on artificial intelligence and statistics, PMLR, vol 15. JMLR Workshop and Conference Proceedings, pp 315–323
Hanczar B, Bourgeais V, Zehraoui F (2022) Assessment of deep learning and transfer learning for cancer prediction based on gene expression data. BMC Bioinformatics 23(1):262. https://doi.org/10.1186/s12859-022-04807-7
Izadkhah H (2022) Deep learning in bioinformatics: techniques and applications in practice. Elsevier Science
Jansen IE, Savage JE, Watanabe K, Bryois J, Williams DM, Steinberg S, Sealock J, Karlsson IK, Hägg S, Athanasiu L, Voyle N, Proitsi P, Witoelar A, Stringer S, Aarsland D, Almdahl IS, Andersen F, Bergh S, Bettella F, Bjornsson S, Brækhus A, Bråthen G, de Leeuw C, Desikan RS, Djurovic S, Dumitrescu L, Fladby T, Hohman TJ, Jonsson PV, Kiddle SJ, Rongve A, Saltvedt I, Sando SB, Selbæk G, Shoai M, Skene NG, Snaedal J, Stordal E, Ulstein ID, Wang Y, White LR, Hardy J, Hjerling-Leffler J, Sullivan PF, van der Flier WM, Dobson R, Davis LK, Stefansson H, Stefansson K, Pedersen NL, Ripke S, Andreassen OA, Posthuma D (2019) Genome-wide meta-analysis identifies new loci and functional pathways influencing Alzheimer’s disease risk. Nat Genet 51(3):404–413. https://doi.org/10.1038/s41588-018-0311-9
Kingma DP, Ba J (2015) ADAM: a method for stochastic optimization. In: Proceedings of the 3rd International Conference for Learning Representations—ICLR, 2015, San Diego
Kumar A, Sidhu J, Goyal A, Tsao JW (2022) Alzheimer disease. In: StatPearls [Internet]. StatPearls publishing, Treasure Island (FL)
Kunkle BW, Grenier-Boley B, Sims R, Bis JC, Damotte V, Naj AC, Boland A, Vronskaya M, van der Lee SJ, Amlie-Wolf A et al (2019) Genetic meta-analysis of diagnosed Alzheimer’s disease identifies new risk loci and implicates Aβ, tau, immunity and lipid processing. Nat Genet 51(3):414–430. https://doi.org/10.1038/s41588-019-0358-2
Lane CA, Hardy J, Schott JM (2018) Alzheimer’s disease. Eur J Neurol 25(1):59–70. https://doi.org/10.1111/ene.13439
Lee JS, Kim C, Shin JH, Cho H, Shin DS, Kim N, Kim HJ, Kim Y, Lockhart SN, Na DL, Seo SW, Seong JK (2018) Machine learning-based individual assessment of cortical atrophy pattern in Alzheimer’s disease spectrum: development of the classifier and longitudinal evaluation. Sci Rep 8(1):4161. https://doi.org/10.1038/s41598-018-22277-x
Leguy J, Glavatskikh M, Cauchy T, Da Mota B (2021) Scalable estimator of the diversity for de novo molecular generation resulting in a more robust QM dataset (OD9) and a more efficient molecular optimization. J Cheminform 13(1):76. https://doi.org/10.1186/s13321-021-00554-8
Liu S, Crawford DC (2022) Maturation and application of phenome-wide association studies. Trends Genet 38(4):353–363. https://doi.org/10.1016/j.tig.2021.12.002
Maki A (2019) Toward principled regularization of deep networks-From weight decay to feature contraction. Sci Robot 4(30):eaaw1329. https://doi.org/10.1126/scirobotics.aaw1329
Moolayil J (2019) Learn keras for deep neural networks. A fast-track approach to modern deep learning with Python. Apress, Berkeley. https://doi.org/10.1007/978-1-4842-4240-7
Nusrat I, Jang S-B (2018) A comparison of regularization techniques in deep neural networks. Symmetry 10(11):648. https://doi.org/10.3390/sym10110648
Prusa J, Khoshgoftaar TM, Seliya N (2015) The effect of dataset size on training tweet sentiment classifiers. In: Proceedings of the 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 9–11. 96–102. https://api.semanticscholar.org/CorpusID:1291234
Qiu WR, Qi BB, Lin WZ, Zhang SH, Yu WK, Huang SF (2022) Predicting the lung adenocarcinoma and its biomarkers by integrating gene expression and DNA methylation data. Front Genet. 13:926927. https://doi.org/10.3389/fgene.2022.926927
Quarato V, D’Antona S, Battista P, Zupo R, Sardone R, Castiglioni I, Porro D, Frasca M, Cava C (2022) Transcriptional profiling of hippocampus identifies network alterations in Alzheimer’s disease. Appl Sci 12(10):5035. https://doi.org/10.3390/app12105035
Rengasamy D, Jafari M, Rothwell B, Chen X, Figueredo GP (2020) Deep learning with dynamically weighted loss function for sensor-based prognostics and health management. Sensors (basel) 20(3):723. https://doi.org/10.3390/s20030723
Rukhsar L, Bangyal WH, Ali Khan MS, Ag Ibrahim AA, Nisar K, Rawat DB (2022) Analyzing RNA-Seq gene expression data using deep learning approaches for cancer classification. Appl Sci 12(4):1850. https://doi.org/10.3390/app12041850
Shao X, Yang H, Zhuang X, Liao J, Yang P, Cheng J, Lu X, Chen H, Fan X (2021) scDeepSort: a pre-trained cell-type annotation method for single-cell transcriptomics using deep learning with a weighted graph neural network. Nucleic Acids Res. 49(21):e122. https://doi.org/10.1093/nar/gkab775
Solfrizzi V, Panza F, Frisardi V, Seripa D, Logroscino G, Imbimbo BP, Pilotto A (2011) Diet and Alzheimer’s disease risk factors or prevention: the current evidence. Expert Rev Neurother 11(5):677–708. https://doi.org/10.1586/ern.11.56
Squitti R, Siotto M, Polimanti R (2014) Low-copper diet as a preventive strategy for Alzheimer’s disease. Neurobiol Aging 35(Suppl 2):S40-50. https://doi.org/10.1016/j.neurobiolaging.2014.02.031
Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15(1):1929–1958
Tesi N, Hulsman M, van der Lee SJ, Jansen IE, Stringa N, van Schoor NM, Scheltens P, van der Flier WM, Huisman M, Reinders MJT, Holstege H (2021) The effect of Alzheimer’s disease-associated genetic variants on longevity. Front Genet. 12:748781. https://doi.org/10.3389/fgene.2021.748781
van IJzendoorn DGP, Szuhai K, Briaire-de Bruijn IH, Kostine M, Kuijjer ML, Bovée JVMG (2019) Machine learning analysis of gene expression data reveals novel diagnostic and prognostic biomarkers and identifies therapeutic targets for soft tissue sarcomas. Plos Comput Biol 15(2):e1006826. https://doi.org/10.1371/journal.pcbi.1006826
Wilentzik Müller R, Gat-Viks I (2020) Exploring neural networks and related visualization techniques in gene expression data. Front Genet 11:402. https://doi.org/10.3389/fgene.2020.00402
Yu H, Samuels DC, Zhao YY, Guo Y (2019) Architectures and accuracy of artificial neural network for disease classification from omics data. BMC Genomics 20(1):167. https://doi.org/10.1186/s12864-019-5546-z
Zhu W, Xie L, Han J, Guo X (2020) The application of deep learning in cancer prognosis prediction. Cancers (Basel) 12(3):603. https://doi.org/10.3390/cancers12030603
Zhu W, Xie L, Han J, Guo X (2020b) The application of deep learning in cancer prognosis prediction. Cancers (basel) 12(3):603. https://doi.org/10.3390/cancers12030603
Funding
Open access funding provided by IBFM - SEGRATE within the CRUI-CARE Agreement. Financial support was provided by National Recovery and Resilience Plan- National Biodiversity Future Center -NBFC: CN00000033.
Author information
Authors and Affiliations
Contributions
Conceptualization: C. C.; methodology: C. C., S. D., F. M.; formal analysis: C. C., S. D., F. M. writing—original draft preparation: C. C. and F.M..; supervision: I. C., D. P. All authors have read and agreed to the published version of the manuscript.
Corresponding author
Ethics declarations
Ethical approval
Ethical permission was not applied because we used gene expression profiles from publica database.
Competing interests
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Cava, C., D’Antona, S., Maselli, F. et al. From genetic correlations of Alzheimer’s disease to classification with artificial neural network models. Funct Integr Genomics 23, 293 (2023). https://doi.org/10.1007/s10142-023-01228-4
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10142-023-01228-4