Insights Into the Inner Workings of Transformer Models for Protein Function Prediction

09/07/2023
by   Markus Wenzel, et al.
0

Motivation: We explored how explainable AI (XAI) can help to shed light into the inner workings of neural networks for protein function prediction, by extending the widely used XAI method of integrated gradients such that latent representations inside of transformer models, which were finetuned to Gene Ontology term and Enzyme Commission number prediction, can be inspected too. Results: The approach enabled us to identify amino acids in the sequences that the transformers pay particular attention to, and to show that these relevant sequence parts reflect expectations from biology and chemistry, both in the embedding layer and inside of the model, where we identified transformer heads with a statistically significant correspondence of attribution maps with ground truth sequence annotations (e.g., transmembrane regions, active sites) across many proteins. Availability and Implementation: Source code can be accessed at https://github.com/markuswenzel/xai-proteins .

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/17/2022

Transformer Neural Networks Attending to Both Sequence and Structure for Protein Prediction Tasks

The increasing number of protein sequences decoded from genomes is openi...
research
06/26/2020

BERTology Meets Biology: Interpreting Attention in Protein Language Models

Transformer architectures have proven to learn useful representations fo...
research
06/14/2022

Exploring evolution-based -free protein language models as protein function predictors

Large-scale Protein Language Models (PLMs) have improved performance in ...
research
03/02/2022

FastFold: Reducing AlphaFold Training Time from 11 Days to 67 Hours

Protein structure prediction is an important method for understanding ge...
research
04/29/2018

OPA2Vec: combining formal and informal content of biomedical ontologies to improve similarity-based prediction

Motivation: Ontologies are widely used in biology for data annotation, i...
research
01/12/2023

Tracr: Compiled Transformers as a Laboratory for Interpretability

Interpretability research aims to build tools for understanding machine ...
research
02/18/2021

Gifsplanation via Latent Shift: A Simple Autoencoder Approach to Progressive Exaggeration on Chest X-rays

Motivation: Traditional image attribution methods struggle to satisfacto...

Please sign up or login with your details

Forgot password? Click here to reset