4.3 Article

Language models for the prediction of SARS-CoV-2 inhibitors

Journal

Publisher

SAGE PUBLICATIONS LTD
DOI: 10.1177/10943420221121804

Keywords

COVID-19; drug design; machine learning; language model; pre-training; fine-tuning; genetic algorithm

Funding

  1. Exascale Computing Project [17SC-20-SC]
  2. DOE CARES through the Advanced Scientific Computing Research (ASCR) program

Ask authors/readers for more resources

The article introduces the research results of using deep learning language models to generate and score drug candidates, reducing pre-training time and increasing dataset size. By fine-tuning the language model, inhibitors for important protein targets of the novel coronavirus were successfully identified. Genetic algorithm was used to find optimal candidates.
The COVID-19 pandemic highlights the need for computational tools to automate and accelerate drug design for novel protein targets. We leverage deep learning language models to generate and score drug candidates based on predicted protein binding affinity. We pre-trained a deep learning language model (BERT) on similar to 9.6 billion molecules and achieved peak performance of 603 petaflops in mixed precision. Our work reduces pre-training time from days to hours, compared to previous efforts with this architecture, while also increasing the dataset size by nearly an order of magnitude. For scoring, we fine-tuned the language model using an assembled set of thousands of protein targets with binding affinity data and searched for inhibitors of specific protein targets, SARS-CoV-2 Mpro and PLpro. We utilized a genetic algorithm approach for finding optimal candidates using the generation and scoring capabilities of the language model. Our generalizable models accelerate the identification of inhibitors for emerging therapeutic targets.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.3
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available