Transfer learning improves outcome predictions for ASD from gene expression in blood

Research output: Other contribution

Abstract

AbstractBackground<jats:p/>Predicting outcomes on human genetic studies is difficult because the number of variables (genes) is often much larger than the number of observations (human subject tissue samples). We investigated means for improving model performance on the types of under-constrained problems that are typical in human genetics, where the number of strongly correlated genes (features) may exceed 10,000, and the number of study participants (observations) may be limited to under 1,000.Methods<jats:p/>We created ‘train’, ‘validate’ and ‘test’ datasets from 240 microarray observations from 127 subjects diagnosed with autism spectrum disorder (ASD) and 113 ‘typically developing’ (TD) subjects. We trained a neural network model (a.k.a., the ‘naive’ model) on 10,422 genes using the ‘train’ dataset, composed of 70 ASD and 65 TD subjects, and we restricted the model to one, fully-connected hidden layer to minimize the number of trainable parameters, including a dropout layer to help prevent overfitting. We experimented with alternative network architectures and tuned the hyperparameters using the ‘validate’ dataset, and performed a single, final evaluation using the holdout ‘test’ dataset. Next, we trained a neural network model using the identical architecture and identical genes to predict tissue type in GTEx data. We transferred that learning by replacing the top layer of the GTEx model with a layer to predict ASD outcome and we retrained the new layer on the ASD dataset, again using the identical 10,422 genes.Findings<jats:p/>The ‘naive’ neural network model had AUROC=0.58 for the task of predicting ASD outcomes, which saw a statistically significant 7.8% improvement from transfer learning.Interpretation<jats:p/>We demonstrated that neural network learning could be transferred from models trained on large RNA-Seq gene expression to a model trained on a small, microarray gene expression dataset with clinical utility for mitigating over-training on small sample sizes. Incidentally, we built a highly accurate classifier of tissue type with which to perform the transfer learning.Funding<jats:p/>This work was supported in part by NIMH R01-MH110558 (E.C., N.E.L.)Author Summary<jats:p/>Image recognition and natural language processing have enjoyed great success in reusing the computational efforts and data sources to overcome the problem of over-training a neural network on a limited dataset. Other domains using deep learning, including genomics and clinical applications, have been slower to benefit from transfer learning. Here we demonstrate data preparation and modeling techniques that allow genomics researchers to take advantage of transfer learning in order to increase the utility of limited clinical datasets. We show that a non-pre-trained, ‘naive’ model performance can be improved by 7.8% by transferring learning from a highly performant model trained on GTEx data to solve a similar problem.
Original languageUndefined
DOIs
StatePublished - Jun 29 2021

Cite this