Probabilistic Transformer: Modelling Ambiguities and Distributions for RNA Folding and Molecule Design

05/27/2022
by   Jörg K. H. Franke, et al.
0

Our world is ambiguous and this is reflected in the data we use to train our algorithms. This is especially true when we try to model natural processes where collected data is affected by noisy measurements and differences in measurement techniques. Sometimes, the process itself can be ambiguous, such as in the case of RNA folding, where a single nucleotide sequence can fold into multiple structures. This ambiguity suggests that a predictive model should have similar probabilistic characteristics to match the data it models. Therefore, we propose a hierarchical latent distribution to enhance one of the most successful deep learning models, the Transformer, to accommodate ambiguities and data distributions. We show the benefits of our approach on a synthetic task, with state-of-the-art results in RNA folding, and demonstrate its generative capabilities on property-based molecule design, outperforming existing work.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/24/2021

Fold2Seq: A Joint Sequence(1D)-Fold(3D) Embedding-based Generative Model for Protein Design

Designing novel protein sequences for a desired 3D topological fold is a...
research
11/16/2022

GAMMT: Generative Ambiguity Modeling Using Multiple Transformers

We introduce a new model based on sets of probabilities for sequential d...
research
07/25/2023

Towards Resolving Word Ambiguity with Word Embeddings

Ambiguity is ubiquitous in natural language. Resolving ambiguous meaning...
research
03/14/2023

Explanation Shift: Investigating Interactions between Models and Shifting Data Distributions

As input data distributions evolve, the predictive performance of machin...
research
09/09/2022

Gluformer: Transformer-Based Personalized Glucose Forecasting with Uncertainty Quantification

Deep learning models achieve state-of-the art results in predicting bloo...
research
09/13/2023

A Cancellation Law for Probabilistic Processes

We show a cancellation property for probabilistic choice. If distributio...
research
02/06/2023

Probabilistic Contrastive Learning Recovers the Correct Aleatoric Uncertainty of Ambiguous Inputs

Contrastively trained encoders have recently been proven to invert the d...

Please sign up or login with your details

Forgot password? Click here to reset