4.8 Article

ProtTrans: Toward Understanding the Language of Life Through Self-Supervised Learning

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2021.3095381

关键词

Proteins; Training; Amino acids; Task analysis; Databases; Computational modeling; Three-dimensional displays; Computational biology; high performance computing; machine learning; language modeling; deep learning

资金

  1. Software Campus 2.0 (TUM) through the German Ministry for Research and Education (BMBF)
  2. Alexander von Humboldt foundation through the German Ministry for Research and Education (BMBF)
  3. Deutsche Forschungsgemeinschaft [DFG-GZ: RO1320/4-1]
  4. NVIDIA
  5. National Research Foundation of Korea [2019R1A6A1A10073437, NRF-2020M3A9G7103933]
  6. SeoulNational University
  7. Google Cloud
  8. Google Cloud Research Credits Program under Covid19 HPC Consortium grant
  9. DOE Office of Science User Facility [DEAC05-00OR22725]
  10. TPU pods under TensorFlow Research Cloud grant
  11. National Research Foundation of Korea [2019R1A6A1A10073437] Funding Source: Korea Institute of Science & Technology Information (KISTI), National Science & Technology Information Service (NTIS)

向作者/读者索取更多资源

Computational biology and bioinformatics provide valuable data for the development of language models in natural language processing. In this study, six different models were trained on protein sequence data and the resulting embeddings were used for various protein structure prediction tasks, demonstrating their advantages over traditional methods.
Computational biology and bioinformatics provide vast data gold-mines from protein sequences, ideal for Language Models (LMs) taken from Natural Language Processing (NLP). These LMs reach for new prediction frontiers at low inference costs. Here, we trained two auto-regressive models (Transformer-XL, XLNet) and four auto-encoder models (BERT, Albert, Electra, T5) on data from UniRef and BFD containing up to 393 billion amino acids. The protein LMs (pLMs) were trained on the Summit supercomputer using 5616 GPUs and TPU Pod up-to 1024 cores. Dimensionality reduction revealed that the raw pLM-embeddings from unlabeled data captured some biophysical features of protein sequences. We validated the advantage of using the embeddings as exclusive input for several subsequent tasks: (1) a per-residue (per-token) prediction of protein secondary structure (3-state accuracy Q3=81%-87%); (2) per-protein (pooling) predictions of protein sub-cellular location (ten-state accuracy: Q10=81%) and membrane versus water-soluble (2-state accuracy Q2=91%). For secondary structure, the most informative embeddings (ProtT5) for the first time outperformed the state-of-the-art without multiple sequence alignments (MSAs) or evolutionary information thereby bypassing expensive database searches. Taken together, the results implied that pLMs learned some of the grammar of the language of life. All our models are available through https://github.com/agemagician/ProtTrans.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.8
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据