LMExplainer: a Knowledge-Enhanced Explainer for Language Models

03/29/2023
by   Zichen Chen, et al.
0

Large language models (LMs) such as GPT-4 are very powerful and can process different kinds of natural language processing (NLP) tasks. However, it can be difficult to interpret the results due to the multi-layer nonlinear model structure and millions of parameters. Lack of understanding of how the model works can make the model unreliable and dangerous for everyday users in real-world scenarios. Most recent works exploit the weights of attention to provide explanations for model predictions. However, pure attention-based explanation is unable to support the growing complexity of the models, and cannot reason about their decision-making processes. Thus, we propose LMExplainer, a knowledge-enhanced interpretation module for language models that can provide human-understandable explanations. We use a knowledge graph (KG) and a graph attention neural network to extract the key decision signals of the LM. We further explore whether interpretation can also help AI understand the task better. Our experimental results show that LMExplainer outperforms existing LM+KG methods on CommonsenseQA and OpenBookQA. We also compare the explanation results with generated explanation methods and human-annotated results. The comparison shows our method can provide more comprehensive and clearer explanations. LMExplainer demonstrates the potential to enhance model performance and furnish explanations for the reasoning processes of models in natural language.

READ FULL TEXT
research
06/01/2023

TopEx: Topic-based Explanations for Model Comparison

Meaningfully comparing language models is challenging with current expla...
research
10/13/2022

Explanations from Large Language Models Make Small Reasoners Better

Integrating free-text explanations to in-context learning of large langu...
research
05/04/2023

Are Human Explanations Always Helpful? Towards Objective Evaluation of Human Natural Language Explanations

Human-annotated labels and explanations are critical for training explai...
research
06/24/2023

On the Uses of Large Language Models to Interpret Ambiguous Cyberattack Descriptions

The volume, variety, and velocity of change in vulnerabilities and explo...
research
08/23/2023

Diagnosing Infeasible Optimization Problems Using Large Language Models

Decision-making problems can be represented as mathematical optimization...
research
01/28/2021

Explaining Natural Language Processing Classifiers with Occlusion and Language Modeling

Deep neural networks are powerful statistical learners. However, their p...
research
04/30/2020

WT5?! Training Text-to-Text Models to Explain their Predictions

Neural networks have recently achieved human-level performance on variou...

Please sign up or login with your details

Forgot password? Click here to reset