Diet Networks: Thin Parameters for Fat Genomics

Abstract

Learning tasks such as those involving genomic data often poses a serious challenge: the number of input features can be orders of magnitude larger than the number of training examples, making it difficult to avoid overfitting, even when using the known regularization techniques. We focus here on tasks in which the input is a description of the genetic variation specific to a patient, the single nucleotide polymorphisms (SNPs), yielding millions of ternary inputs. Improving the ability of deep learning to handle such datasets could have an important impact in medical research, more specifically in precision medicine, where high-dimensional data regarding a particular patient is used to make predictions of interest. Even though the amount of data for such tasks is increasing, this mismatch between the number of examples and the number of inputs remains a concern. Naive implementations of classifier neural networks involve a huge number of free parameters in their first layer (number of input features times number of hidden units): each input feature is associated with as many parameters as there are hidden units. We propose a novel neural network parametrization which considerably reduces the number of free parameters. It is based on the idea that we can first learn or provide a distributed representation for each input feature (e.g. for each position in the genome where variations are observed in data), and then learn (with another neural network called the parameter prediction network) how to map a feature's distributed representation (based on the feature's identity not its value) to the vector of parameters specific to that feature in the classifier neural network (the weights which link the value of the feature to each of the hidden units). This approach views the problem of producing the parameters associated with each feature as a multi-task learning problem. We show experimentally on a population stratification task of interest to medical studies that the proposed approach can significantly reduce both the number of parameters and the error rate of the classifier.

Cite

Text

Romero et al. "Diet Networks: Thin Parameters for Fat Genomics." International Conference on Learning Representations, 2017.

Markdown

[Romero et al. "Diet Networks: Thin Parameters for Fat Genomics." International Conference on Learning Representations, 2017.](https://mlanthology.org/iclr/2017/romero2017iclr-diet/)

BibTeX

@inproceedings{romero2017iclr-diet,
  title     = {{Diet Networks: Thin Parameters for Fat Genomics}},
  author    = {Romero, Adriana and Carrier, Pierre Luc and Erraqabi, Akram and Sylvain, Tristan and Auvolat, Alex and Dejoie, Etienne and Legault, Marc-André and Dubé, Marie-Pierre and Hussin, Julie G. and Bengio, Yoshua},
  booktitle = {International Conference on Learning Representations},
  year      = {2017},
  url       = {https://mlanthology.org/iclr/2017/romero2017iclr-diet/}
}