NAR-Former V2: Rethinking Transformer for Universal Neural Network Representation Learning

06/19/2023
by   Yun Yi, et al.
0

As more deep learning models are being applied in real-world applications, there is a growing need for modeling and learning the representations of neural networks themselves. An efficient representation can be used to predict target attributes of networks without the need for actual training and deployment procedures, facilitating efficient network deployment and design. Recently, inspired by the success of Transformer, some Transformer-based representation learning frameworks have been proposed and achieved promising performance in handling cell-structured models. However, graph neural network (GNN) based approaches still dominate the field of learning representation for the entire network. In this paper, we revisit Transformer and compare it with GNN to analyse their different architecture characteristics. We then propose a modified Transformer-based universal neural network representation learning model NAR-Former V2. It can learn efficient representations from both cell-structured networks and entire networks. Specifically, we first take the network as a graph and design a straightforward tokenizer to encode the network into a sequence. Then, we incorporate the inductive representation learning capability of GNN into Transformer, enabling Transformer to generalize better when encountering unseen architecture. Additionally, we introduce a series of simple yet effective modifications to enhance the ability of the Transformer in learning representation from graph structures. Our proposed method surpasses the GNN-based method NNLP by a significant margin in latency estimation on the NNLQP dataset. Furthermore, regarding accuracy prediction on the NASBench101 and NASBench201 datasets, our method achieves highly comparable performance to other state-of-the-art methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/28/2023

Transformer and Snowball Graph Convolution Learning for Biomedical Graph Classification

Graph or network has been widely used for describing and modeling comple...
research
11/15/2022

NAR-Former: Neural Architecture Representation Learning towards Holistic Attributes Prediction

With the wide and deep adoption of deep learning models in real applicat...
research
05/16/2023

Inductive Graph Neural Networks for Moving Object Segmentation

Moving Object Segmentation (MOS) is a challenging problem in computer vi...
research
05/25/2021

Graph Neural Network Based VC Investment Success Prediction

Predicting the start-ups that will eventually succeed is essentially imp...
research
06/09/2021

Do Transformers Really Perform Bad for Graph Representation?

The Transformer architecture has become a dominant choice in many domain...
research
01/23/2022

Investigating Expressiveness of Transformer in Spectral Domain for Graphs

Transformers have been proven to be inadequate for graph representation ...
research
11/05/2022

Inductive Graph Transformer for Delivery Time Estimation

Providing accurate estimated time of package delivery on users' purchasi...

Please sign up or login with your details

Forgot password? Click here to reset