Your browser doesn't support javascript.
loading
Relative molecule self-attention transformer.
Maziarka, Lukasz; Majchrowski, Dawid; Danel, Tomasz; Gainski, Piotr; Tabor, Jacek; Podolak, Igor; Morkisz, Pawel; Jastrzebski, Stanislaw.
Afiliación
  • Maziarka L; Faculty of Mathematics and Computer Science, Jagiellonian University, Lojasiewicza 6, 30-348, Cracow, Poland. lukasz.maziarka@ii.uj.edu.pl.
  • Majchrowski D; NVIDIA, 2788 San Tomas Expy, Santa Clara, CA, 95051, USA.
  • Danel T; Faculty of Mathematics and Computer Science, Jagiellonian University, Lojasiewicza 6, 30-348, Cracow, Poland.
  • Gainski P; Faculty of Mathematics and Computer Science, Jagiellonian University, Lojasiewicza 6, 30-348, Cracow, Poland.
  • Tabor J; Ardigen, Podole 76, 30-394, Cracow, Poland.
  • Podolak I; Faculty of Mathematics and Computer Science, Jagiellonian University, Lojasiewicza 6, 30-348, Cracow, Poland.
  • Morkisz P; Faculty of Mathematics and Computer Science, Jagiellonian University, Lojasiewicza 6, 30-348, Cracow, Poland.
  • Jastrzebski S; NVIDIA, 2788 San Tomas Expy, Santa Clara, CA, 95051, USA.
J Cheminform ; 16(1): 3, 2024 Jan 03.
Article en En | MEDLINE | ID: mdl-38173009
ABSTRACT
The prediction of molecular properties is a crucial aspect in drug discovery that can save a lot of money and time during the drug design process. The use of machine learning methods to predict molecular properties has become increasingly popular in recent years. Despite advancements in the field, several challenges remain that need to be addressed, like finding an optimal pre-training procedure to improve performance on small datasets, which are common in drug discovery. In our paper, we tackle these problems by introducing Relative Molecule Self-Attention Transformer for molecular representation learning. It is a novel architecture that uses relative self-attention and 3D molecular representation to capture the interactions between atoms and bonds that enrich the backbone model with domain-specific inductive biases. Furthermore, our two-step pretraining procedure allows us to tune only a few hyperparameter values to achieve good performance comparable with state-of-the-art models on a wide selection of downstream tasks.
Palabras clave

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Tipo de estudio: Prognostic_studies Idioma: En Revista: J Cheminform Año: 2024 Tipo del documento: Article País de afiliación: Polonia Pais de publicación: Reino Unido

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Tipo de estudio: Prognostic_studies Idioma: En Revista: J Cheminform Año: 2024 Tipo del documento: Article País de afiliación: Polonia Pais de publicación: Reino Unido