4.8 Article

Integrating convolution and self-attention improves language model of human genome for interpreting non-coding regions at base-resolution

Journal

NUCLEIC ACIDS RESEARCH
Volume 50, Issue 14, Pages -

Publisher

OXFORD UNIV PRESS
DOI: 10.1093/nar/gkac326

Keywords

-

Funding

  1. Guangdong Provincial Academician Workstation of BGI Synthetic Genomics [2017B090904014]
  2. Program of Shanghai Academic Research Leader [20XD1401100]
  3. Program for Outstanding Medical Academic Leader [2019LJ01]

Ask authors/readers for more resources

LOGO is a self-attention based pre-trained language model that learns bi-directional representations of the unlabelled human reference genome. It achieves good performance in interpreting non-coding regions and shows advantages in tasks such as promoter identification and enhancer-promoter interaction prediction.
Interpretation of non-coding genome remains an unsolved challenge in human genetics due to impracticality of exhaustively annotating biochemically active elements in all conditions. Deep learning based computational approaches emerge recently to help interpret non-coding regions. Here, we present LOGO (Language of Genome), a self-attention based contextualized pre-trained language model containing only two self-attention layers with 1 million parameters as a substantially light architecture that applies self-supervision techniques to learn bidirectional representations of the unlabelled human reference genome. LOGO is then fine-tuned for sequence labelling task, and further extended to variant prioritization task via a special input encoding scheme of alternative alleles followed by adding a convolutional module. Experiments show that LOGO achieves 15% absolute improvement for promoter identification and up to 4.5% absolute improvement for enhancer-promoter interaction prediction. LOGO exhibits state-of-the-art multi-task predictive power on thousands of chromatin features with only 3% parameterization benchmarking against the fully supervised model, DeepSEA and 1% parameterization against a recent BERT-based DNA language model. For allelic-effect prediction, locality introduced by one dimensional convolution shows improved sensitivity and specificity for prioritizing non-coding variants associated with human diseases. In addition, we apply LOGO to interpret type 2 diabetes (T2D) GWAS signals and infer underlying regulatory mechanisms. We make a conceptual analogy between natural language and human genome and demonstrate LOGO is an accurate, fast, scalable, and robust framework to interpret non-coding regions for global sequence labeling as well as for variant prioritization at base-resolution.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available