official code for our KDD21 paper "Adaptive Transfer Learning on Graph Neural Networks"
Graph neural networks (GNNs) is widely used to learn a powerful representation of graph-structured data. Recent work demonstrates that transferring knowledge from self-supervised tasks to downstream tasks could further improve graph representation. However, there is an inherent gap between self-supervised tasks and downstream tasks in terms of optimization objective and training data. Conventional pre-training methods may be not effective enough on knowledge transfer since they do not make any adaptation for downstream tasks. To solve such problems, we propose a new transfer learning paradigm on GNNs which could effectively leverage self-supervised tasks as auxiliary tasks to help the target task. Our methods would adaptively select and combine different auxiliary tasks with the target task in the fine-tuning stage. We design an adaptive auxiliary loss weighting model to learn the weights of auxiliary tasks by quantifying the consistency between auxiliary tasks and the target task. In addition, we learn the weighting model through meta-learning. Our methods can be applied to various transfer learning approaches, it performs well not only in multi-task learning but also in pre-training and fine-tuning. Comprehensive experiments on multiple downstream tasks demonstrate that the proposed methods can effectively combine auxiliary tasks with the target task and significantly improve the performance compared to state-of-the-art methods.READ FULL TEXT VIEW PDF
In recent years, graph neural networks (GNNs) have been widely adopted i...
Recent works found that fine-tuning and joint training—two popular
With the development of deep networks on various large-scale datasets, a...
While deep learning has been very beneficial in data-rich settings, task...
Transfer of pre-trained representations can improve sample efficiency an...
Recently, self-supervised learning methods have achieved remarkable succ...
With the growth of the academic engines, the mining and analysis acquisi...
official code for our KDD21 paper "Adaptive Transfer Learning on Graph Neural Networks"
Graph neural networks (GNNs) (bruna2013spectral; gcn) attract a lot of attention in representation learning for graph-structured data. Recent work demonstrates various GNN architectures (gcn; gat; graphsage; xu2018powerful) achieve state-of-the-art performances in many graph-based tasks, such as node classification (gcn), link prediction (zhang2018link), and graph classification (DiffPool). These GNNs have been proven effective to learn powerful representations in a variety of scenarios, including recommendation system (DBLP:conf/kdd/YingHCEHL18) and social network (fan2019graph).
GNNs are often trained in an end-to-end manner with supervision which usually requires abundant labeled data, but task-specific labels can be scarce for many graph datasets. Some recent works (hu2020gpt) study GNN pre-training and employing auxiliary tasks (hwang2020self) to solve such problems. Their goal is to transfer the learned knowledge from self-supervised tasks to downstream tasks, and these methods prove making use of abundant unlabeled data could further improve graph representation. However, due to the difference of optimization objectives and data distribution between self-supervised tasks and downstream tasks, there is an inherent gap between pre-training and fine-tuning. In addition, conventional pre-training methods are not aware of which self-supervised tasks benefit more for downstream tasks and do not make any adaptation for them. This leads to knowledge transfer not effective enough for downstream tasks. There is little exploration in improving the effectiveness of transfer learning on GNNs, fully leveraging information in self-supervision, and enabling sufficient adaptation for downstream tasks.
In this work, we explore effective knowledge transfer between self-supervised tasks and downstream tasks, which is an under-investigated problem in GNNs compared to natural language processing (NLP). Optimizing the effectiveness of knowledge transfer and alleviating the gap between the pre-trained model and the target task have been the focus of recent studies in NLP(howard2018universal; chronopoulou2019embarrassingly; gururangan2020don). (chronopoulou2019embarrassingly) addresses this by using LM as an auxiliary task to prevent catastrophic forgetting.
Inspired by these works, we propose a new generic transfer learning paradigm that adaptively selects and combines various auxiliary tasks with the target task in fine-tuning stage to achieve a better adaptation for downstream tasks. In other words, we combine the optimization objective from the target task with multiple objectives from auxiliary tasks. This preserves sufficient knowledge captured by self-supervised tasks while improving the effectiveness of transfer learning on GNNs.
Challenges. Transferring knowledge from graph-based self-supervised tasks is a powerful tool for learning a better graph representation. Many self-supervision methods on graphs, such as DGI (velickovic2019deep), edge generation (hu2020gpt), attribute masking (pretrain) and meta-path prediction (hwang2020self) have been studied and their effectiveness has been proved. The basic principle, based on which additional auxiliary tasks can benefit the target task, is that auxiliary tasks share some unknown information with the target task, and to some extent, help learn features that are beneficial to the target task. However, the correlation between auxiliary tasks and the target task can be large or small and may change along with training. An auxiliary task may be helpful to the target task at the beginning of training, but may no longer be useful in the later stages of training (lin2019adaptive). If an auxiliary task is assigned with an inappropriate weight, it may even hurt the target task, which means negative transfer (pan2009survey) occurs.
Graph preserves diverse information and downstream tasks are also diverse, single auxiliary task could not fully explore all organic signals. Previous work involving multiple self-supervised tasks in graph learning usually manually sets the weight for each task using prior intuition, e.g. assuming that all auxiliary tasks matter equally or considering the magnitude of the loss for each task (hu2020gpt). We analyze the effect of different combinations of auxiliary tasks. Figure 1 shows that the weights of tasks are critical, different combinations of weights lead to different results on the target task and bad combinations lead to negative transfer. We also find the same set of weights behave differently in two transfer schemes, multi-task learning from scratch and pre-training and fine-tuning (in this scheme, we add auxiliary tasks in fine-tuning stage). Therefore, the challenge of using auxiliary tasks lies in the difficulty in selecting and combining different auxiliary tasks especially when the number of tasks grows. It is still an open question that has been less explored in graph transfer learning, especially in adding auxiliary tasks in fine-tune stage after pre-train.
Some related works try to adaptively tune the weights of each task during the training process (chen2018gradnorm; liu2019end; kendall2018multi; du2018adapting; lin2019adaptive; hwang2020self). We compared some of these methods in two schemes, one is multi-task learning and the other is adding auxiliary tasks in fine-tune stage after pre-train. We find that such weighting strategies of auxiliary tasks may be effective in multi-task learning, but may fail in pre-training and fine-tuning scheme. We argue that it’s because the importance of auxiliary tasks to the target task changes in different learning stages. In multi-tasking learning, both the target task and auxiliary tasks are trained from scratch; whereas in pre-training and fine-tuning, models are already learned by abundant pre-train data.
Contributions. To summarize, this work makes the following major contributions.
We propose a new generic transfer learning paradigm on GNNs which adaptively selects and combines various auxiliary tasks on graphs. It can better adapt to target tasks and improve the effectiveness of knowledge transfer. We quantify the consistency between an auxiliary task and the target task by measuring the cosine similarity between the gradients of the auxiliary task and the target task, and we design an adaptive auxiliary loss weighting model to learn the weights of auxiliary tasks by considering this signal. Besides, the weighting model could be optimized through meta-learning.
We conduct comprehensive experiments on both multi-task learning and pre-training and fine-tuning across multiple datasets. Our methods outperform other methods and achieve state-of-the-art results. It is worth mentioning that our methods perform well in pre-training and fine-tuning while other methods may be not effective. Our approach is not limited to GNNs scenarios but is a generic transfer learning paradigm that could be flexibly applied to various transfer learning scenarios such as in NLP or computer vision (CV).
(iii) We further conduct a thorough analysis to study the advantages of our methods compared with other baseline methods.
Pre-Training (DBLP:journals/corr/abs-1810-04805) is a common and effective way to transfer knowledge learned from related tasks to a target task to improve generalization. A general way is to pre-train a language model (LM) on a large corpus of unlabeled text and then fine-tune the model on supervised data. Recently, researchers have explored pre-training for GNNs to enable more effective learning on graphs from abundant unlabeled graph data (pretrain; hu2020gpt). Multi-task learning (hwang2020self) is another common method to achieve shared representations by simultaneously training a joint objective of the supervised task and the self-supervised tasks. (hwang2020self) designs a group of auxiliary tasks to learn the heterogeneity of graphs and try to transfer such knowledge from auxiliary tasks to a target task. (you2020does) tries to compare three schemes of self-supervision on GCNs.
However, knowledge transfer for GNN is challenging due to the differences between self-supervised tasks and downstream tasks on optimization objectives and data distribution. Optimizing the effectiveness of knowledge transfer and alleviating the gap between the pre-trained model and the target task has attracted a lot of attention in recent NLP research (howard2018universal; chronopoulou2019embarrassingly; gururangan2020don). (howard2018universal) shows continued pre-train of an LM on the unlabeled data of a given task is beneficial for the downstream task. Towards the same direction, (chronopoulou2019embarrassingly) addresses this by using LM as an auxiliary task to prevent catastrophic forgetting. Besides, (gururangan2020don) show the importance of further adaptation with large corpora of in-domain data and task’s unlabeled data. Inspired by these works, we argue that during the transfer process, knowledge from self-supervised tasks is required for both pre-training and fine-tuning.
Self-supervision is a promising direction to learn more transferable, generalized and robust representations. Research of self-supervised tasks on graphs attracts a lot of attention recently (kipf2016variational; graphsage; infograph; velickovic2019deep; sun2020multi). Graph-structured data provides rich information that enables us to design self-supervised tasks from various perspectives, e.g. node attributes (pretrain; hu2020gpt), underlying graph structure (graphsage; peng2020self; sun2020multi) or heterogeneity of graph (hwang2020self), etc. For example, edge generation (hu2020gpt) and context prediction (pretrain) map nodes appearing in local contexts to nearby embeddings. K-hop context prediction (peng2020self) tries to preserve the global topology of graph to better characterize the similarity and differentiation between nodes. There are also methods (DGI (velickovic2019deep), InfoGraph (infograph), etc.) introduce Graph Informax which maximizes the mutual information between local and global representations. Besides, attribute generation (pretrain) ensure certain aspects of node attributes are encoded in the node embeddings.
Some recent work utilizes multiple self-supervised tasks together on graphs. (hu2020gpt) proposes to use attribute generation and edge generation tasks to capture the inherent dependency between node attributes and graph structure during the generative process. (hwang2020self) transfers the knowledge of heterogeneity of graph to the target task via adding auxiliary tasks of predicting meta-paths. However how to effectively select and combine different self-supervised tasks are less explored in GNNs.
Some other works (jin2020self; lu2021learning) have studied that the same auxiliary task behaves differently on different downstream tasks, and sometimes it may even cause negative-transfer. We argue that multiple kinds of information on graphs could be captured by different auxiliary tasks and their interactions and combinations provide big opportunities for us to learn a more powerful representation.
Past work has proven that, multi-task learning or auxiliary learning which simultaneously learns multiple related tasks could achieve better generalization by sharing a representation (liu2019end; hwang2020self; liu2019multi). The goal of such methods includes achieving best performance across all tasks and only for the target task. In this paper, we only discuss the later one. The success of these approaches depends on how well aligned the auxiliary tasks are with the target task and how to combine these losses with the target task. However, the usefulness of auxiliary tasks is non-trivial to know and may change through the training.
The weight for each of the auxiliary task is usually manually set using prior intuition or through hyper-parameter tuning in above mentioned GNN self-supervised works. It becomes much harder to determine the weights as the number of auxiliary tasks grows.
Recent works study various weighting strategies to balance tasks. Some of them treat all tasks equally and adapt the weights based on the gradient norm (chen2018gradnorm) or task uncertainty (kendall2018multi). Some other works evaluate the usefulness of auxiliary tasks to the target task and adapt the weights accordingly so that the more useful ones receive higher weights (du2018adapting; lin2019adaptive). (du2018adapting) proposes to use gradient similarity between an auxiliary task and the target task on shared parameters to determine the usefulness of auxiliary tasks. Some recent works (maxl; hwang2020self) use meta-learning in auxiliary task selection. (maxl) proposes an elegant solution to generate new auxiliary tasks by collapsing existing classes. (hwang2020self) use meta-learning to automatically select and balance tasks and transfer the knowledge from auxiliary tasks to the target task. In this paper, we use meta-learning to learn a weigh function and the consistency of auxiliary tasks to the target task is considered in this function.
In the following, we first explain the existence of a gap between pre-training and fine-tuning, and to address the challenges, we propose our Generic Auxiliary loss Transfer Learning Framework. The goal of this framework is to effectively leverage self-supervised tasks as auxiliary tasks to help the target task. Later we introduce our Adaptive Auxiliary Loss Weighting Algorithm using Task Similarity, in which, we design an adaptive auxiliary loss weighting model to learn the weights of auxiliary tasks by quantifying task similarity and we optimize the model through meta-learning. Finally, we discuss the generality and effectiveness of our methods.
We use to denote a graph, where represents the node set with nodes, stands for the edge set with edges. Denoting as the node feature matrix, and as the adjacency matrix or related variant. A GNN model encodes node attributes and contexts into a embedding by iteratively aggregating messages from node’s neighbors.
A classic paradigm of pre-training on graphs is to first pre-train a general GNN model by self-supervised tasks on abundant unlabeled data to achieve a good initialization for downstream tasks. The network is trained with the objective as follows:
where denotes the losses of pre-training tasks, denotes the pseudo labels of self-supervised data and
represents task-specific parameters for pre-train tasks, e.g. a classifier to predict pseudo labels.
Then in the fine-tuning process, shared GNN model is initialized by pre-trained model and fine-tuned on downstream tasks by supervised data. The objective of fine-tuning is to optimize the following:
where is the loss of fine-tuning on the downstream task, denotes labels of supervised data, and represents task-specific parameters for the target task.
The goal of pre-training is to learn some general and critical features that can be transferred to downstream tasks; whereas, there is a gap between training objectives of self-supervised tasks and supervised tasks and there is not enough adaptation for downstream tasks in this paradigm of pre-training, which leaves the knowledge transfer not effective enough. As to our knowledge, there is only one work (lu2021learning) studies this problem on graphs, it proposes a method to structure the pre-training stage to simulate the fine-tuning process on downstream tasks, which directly optimizes the pre-trained model’s quick adaptability to downstream tasks. We propose a new transfer learning paradigm on GNNs which would adaptively select and combine different auxiliary tasks with the target task in fine-tuning stage to achieve a better adaptation for downstream tasks and improve the effectiveness of knowledge transfer. Our method can automatically select the most useful tasks for downstream tasks from a task set and balances their weights. It’s a principled and generic approach that can be applied to all kinds of auxiliary tasks.
Figure 2 illustrates our proposed Generic Auxiliary Loss Transfer Learning Framework. In our approach, we first pre-train a GNN model by self-supervised tasks. We then transfer the parameters of the pre-trained model and fine-tune them with a joint loss. The joint loss is the weighted sum of the target task’s loss and multiple auxiliary losses. We also add multiple task-specific layers for such auxiliary tasks which are randomly initialized. Adding auxiliary losses in fine-tuning can avoid catastrophic forgetting and better adapt for the target task, resulting in more powerful representations. The new joint loss for the fine-tuning process can be formulated as:
where is the loss of the target task that we want to complete and represents the loss of auxiliary task and . The losses of both the target task and auxiliary tasks are calculated on shared model parameters and their task-specific layers (omitted in the above formula) at each training step . is the weight for task which is generated by the weighting model, especially means the weight for the target task. We will introduce more details about the weighting model in Section 3.2. Assuming that we use gradient descent to update the model, the optimization of shared parameters on this joint loss is as follows:
If a large number of auxiliary tasks are used, some auxiliary tasks may play a more important role than others in the feature representation learning for the target task; therefore, the weight of each auxiliary task needs to be selected carefully.
In previous work, the process of selecting auxiliary tasks on graphs seems to be too heuristic, and there is not any patterns to follow, which may raise some issues. First, computation will be much expensive when the number of auxiliary tasks K increases. Besides, the usefulness of each auxiliary task may change during the training; using a fixed value as weight may limit or even hurt the performance.
We propose Adaptive Auxiliary Loss Weighting using Task Similarity (AUX-TS), which quantifies the similarity of each auxiliary task to the target task and uses the information obtained after applying the auxiliary task to learn weights during training.
(1) Weighting model quantifies the consistency of each auxiliary task to the target task
Our approach designs a weighting model to learn weights for different auxiliary tasks that benefit the target task most. The weighting model is a 2-layer MLP network, which takes signals that related to current data and task as input and outputs the weight.
(du2018adapting) propose to use the cosine similarity of gradients between tasks to represent the relatedness between tasks and prove the effectiveness of the method. However, this method only uses the signal of cosine similarity in a rule-based way to determine whether adopt the gradients from an auxiliary task. The weights in this method cannot be optimized. Inspired by this work (du2018adapting), our method follows the principle that a useful auxiliary task should provide gradient directions which helps to decrease the loss of the target task.
Specifically, we quantify the consistency of an auxiliary task with the target task by measuring the cosine similarity between the gradients of the auxiliary task and the target task. And we design an adaptive auxiliary loss weighting model to learn the weights of auxiliary tasks by considering this signal. We use to denote the weighting model ( is used in the following formulas as a simplified expression), where represents all the signals that are taken as the input to weighting model, it’s the concatenation of the gradient similarity, loss value and type embedding of the current task. represents the parameters of the weighting model. Since the target task is of our interest, the weighting model is optimized by minimizing the loss of the target task at each training step .
(2) Optimize weighting model via Meta-learning
Some recent works (maxl; hwang2020self) propose the idea of using meta-learning to learn meta-parameters, which themselves are used in another learning procedure. In our method, we treat the weighting model as meta-parameters and optimize it according to the target task during the training. Specifically, we train two networks in two stages. The first stage is to optimize the weighting model which generates weights for all tasks and the second stage is to optimize the shared GNN networks on the combination of the target and auxiliary tasks in a multi-task learning scheme.
In the first stage, the weighting model is optimized by the loss of the target task, encouraging to generate a set of weights, so that if the joint loss weighted by these weights is used to train GNN, the performance of the target task could be maximized.
Gradient similarity between tasks is a strong indicator of the consistency of tasks. We denote the weighting model as which tasks gradients similarity between an auxiliary task and the target task as input:
is the gradient vector evaluated atby the loss of the target task and is the gradient vector evaluated by the loss of auxiliary task . Using to generate weight for each task, the joint loss can be represented as:
After one update of GNN by above total loss, we leverage the loss of the target task on the updated GNN to optimize the weighting model. Therefore, the weighting model is updated as follows:
denotes the updated parameters of the GNN after one gradient update using the joint loss (defined in Equation 6 )
where is the learning rate for GNN. We do not really update the parameters , but plug it in to optimize .
In the second stage, GNN is trained by joint losses of both the target and auxiliary tasks, given the weights of different tasks are determined by the weighting model, we define the multi-task objective of GNN as follows:
We train both networks in an iterative manner until convergence. Assuming that we use gradient descent to update both networks, our proposed method is illustrated in Algorithm 1.
First, our proposed Generic Auxiliary Loss Transfer Learning Framework could alleviate the gap between pre-training and fine-tuning and improve the effectiveness and adaptability of transferring knowledge to downstream tasks significantly. In addition, it has good generality and can be applied to all kinds of self-supervised tasks set and all kinds of GNN models.
Secondly, our proposed AUX-TS algorithm dynamically adjusts the weights for each auxiliary task during the training via meta-learning. We verify our algorithm in two transfer learning schemes. One is multi-task learning, in which GNN model is initialized randomly. The other is pre-training and fine-tuning, in which GNN model is initialized by relatively well pre-trained parameters. Considering the difference between these two learning schemes, the usefulness of auxiliary tasks in different learning stages also changes. In multi-task learning, some auxiliary tasks may contribute more to the joint loss. After the knowledge of the pre-trained model is transferred, the target task may become more important and some of the auxiliary tasks may become less important along with training. Our method uses gradient similarity between an auxiliary task and the target task as an informative signal to learn the usefulness of an auxiliary task. So that it could adaptively generate weights and perform well in both schemes.
We evaluate our proposed methods on four datasets to prove the effectiveness of our Generic Auxiliary loss Transfer Learning Framework and adaptive auxiliary task weighting algorithm (AUX-TS). To evaluate the generality of our methods, we conduct experiments on two auxiliary tasks, multiple GNN models, and two transfer learning schemes. At last, we conduct a comprehensive analysis to study the advantages of our methods compared with other baseline methods.
Datasets We utilize four public datasets from different domains to evaluate the effectiveness of our methods. The detailed statistics are summarized in Table 1.
Open Academic Graph (OAG) (tang2008arnetminer; DBLP:conf/kdd/ZhangLTDYZGWSLW19) is the largest public academic dataset to date which contains more than 178 million nodes and 2.236 billion edges. Each paper in OAG is associated with fields (e.g. Computer Science, Medicine) and publication date ranges from 1900 to 2019. We use Paper-Field (L2) prediction on the Computer Science dataset (CS) as the downstream task and evaluate the performance by MRR and NDCG, where MRR is the default one.
is a social network constructed with Reddit posts in different topical communities. We use the node classification task as the downstream task, which is to classify Reddit posts to different communities and use micro F1-score as the evaluation metric.
Book-Crossing (wang2018ripplenet; wang2019knowledge) contains book ratings from 0 to 10 in the Book-Crossing community. The link prediction of user-item is used as the downstream task and is measured by AUC.
Last-FM (wang2018ripplenet; wang2019knowledge) contains musician listening information from a set of 2 thousand users from Last-FM online music system. The link prediction task is used as the downstream task and is measured by AUC.
|Datasets||# Nodes||# Edges|
Baseline We compare the following approaches:
No Auxiliary Losses denotes our base learning algorithm, which updates the model by the target task only instead of any auxiliary tasks.
Fixed Weights updates the model by both the target task and auxiliary tasks, with a group of pre-defined fixed weights. We assign all the tasks the same weight of 1.
SELAR learns weight functions that consider the loss value to select auxiliary tasks and balance them with the target task via meta-learning.
Cosine Similarity (unweighted) uses gradient similarity to determine whether adopt the gradients from an auxiliary task (du2018adapting). Specifically, it combines the gradients from an auxiliary task and the target task to update the shared parameters only if .
AUX-TS, as described in Algorithm 1. We have two variants, AUX-TS (one sim) corresponds to adding current task’s similarity in weighting model, and AUX-TS (all sims) corresponds to adding all tasks’ similarities in weighting model.
Our proposed auxiliary weighting methods can be applied to two transfer learning schemes, one is multi-task learning (MTL), the other is pre-training and fine-tuning (P&F). We evaluate our methods with baseline methods on these two schemes. In all the experiments, we report the mean of test performance based on three independent runs.
|Metric||NDCG||MRR||NDCG||MRR||F1 score||F1 score|
|No Auxiliary Losses||0.43870.0054||0.49410.0091||0.45840.0039||0.52840.0065||0.84740.0072||0.86050.0073|
|AUX-TS (all sims)||0.4518.0033||0.52150.0076||0.46340.0033||0.54120.0074||0.85910.0116||0.87440.0092|
|AUX-TS (one sim)||0.45240.0033||0.52180.0073||0.46340.0033||0.54060.0073||0.85950.0092||0.87380.0090|
Transfer learning with edge generation and attribute generation. GCN (gcn) is chosen as the base model for OAG_CS and Reddit. We use edge generation and attribute generation as auxiliary tasks and follow the hyper-parameter settings in GPT-GNN (hu2020gpt)
, and the full labeled data is used in our experiments. For multi-task learning, we train the model from scratch for 500/500 epochs, while for pre-training and fine-tuning, we first pre-train a GNN model for 20/50 epochs and then fine-tune it on downstream tasks for 500/500 epochs on OAG_CS/Reddit. We adopt the Time Transfer settings(hu2020gpt) for the pre-training and fine-tuning on OAG_CS, specifically, we use data from different time spans for pre-training (before 2014) and fine-tuning (since 2014). We randomly split Reddit, 70% of which is used for pre-training, and the other 30% is used for fine-tuning (Train/Valid/Test Data each account for 10%). Table 2 summarizes the performance of different methods on OAG_CS and Reddit.
Transfer learning with meta-path prediction . We evaluate our methods on Book-Crossing and Last-FM with multiple GNN architectures: GCN (gcn), GAT (gat) and GIN (xu2018powerful). We use five auxiliary tasks of meta-path prediction and follow most hyper-parameter settings in SELAR (hwang2020self), hyper-parameters such as learning rate and batch size are tuned using validation sets for all models. For multi-task learning, we train the model from scratch for 100/50 epochs, while for pre-training and fine-tuning, we first pre-train a GNN model for 100/50 epochs and then fine-tune it on downstream tasks for 100/50 epochs on Last-FM/Book-Crossing. Since the sizes of these two datasets are relatively small, we pre-train and fine-tune on the same graph. Results on Book-Crossing and Last-FM are summarized in Table 3. Details of implementation and auxiliary tasks can be found in Appendix A/B.
Overall, the proposed methods significantly enhance the performance for all downstream tasks on all datasets. Compared with No Auxiliary Losses baseline, our new transfer learning framework which adaptively selects and combines various auxiliary tasks with the target task effectively improves target tasks. Especially under the pre-training and fine-tuning, it alleviates the gap between pre-training and fine-tuning and achieves a better adaptation for target tasks. On average, our methods achieve (relative) performance gains of 5.61%/2.42% and 1.43%/1.61% over No Auxiliary Losses baseline under MTL/P&F respectively on OAG_CS and Reddit, and 3.55%/1.31% and 6.25%/4.31% on Last-FM and Book-Crossing. These numbers prove the effectiveness of our framework. Moreover, AUX-TS consistently outperforms other auxiliary task weighting methods on all datasets.
Comprehensive experiments evaluate the generality of our methods.
Different transfer schemes: considering the difference between multi-task learning and pre-training and fine-tuning, we verify our methods on both transfer schemes. As observed from Table 2, Fixed Weights and Cosine Similarity are not as effective in pre-training and fine-tuning as in multi-task learning on OAG, while AUX-TS consistently perform well under both schemes.
Diff auxiliary tasks sets: we evaluate on two sets of auxiliary tasks and find AUX-TS greatly improves the results, which proves that AUX-TS can be flexibly applied to all kinds of combinations of tasks.
Different GNN model: Table 3 shows that AUX-TS consistently improves the performance of target tasks for all three GNNs compared with other baseline methods.
The effectiveness and generality of our approach are proved above. We conduct further analysis to understand why our methods work and to study the advantages of our methods compared to other baseline methods. The following analysis is based on our best-performed experiments.
AUX-TS avoids negative transfer. We compare the performance (MRR) curve of our methods with other baseline methods on GCN (gcn) and HGT (hgt) in Figure 3. HGT is a heterogeneous model and is much more complex than GCN. AUX-TS performs better than Fixed Weights and Cosine Similarity on both models, and these two methods hurt the performance on HGT. We can observe that 1) auxiliary tasks are not always helpful to the target task; 2) negative transfer is more likely to occur on the more complex model; and 3) AUX-TS can successfully avoid negative transfer in both models.
AUX-TS adaptively assigns appropriate weights for different tasks during the training. From the loss and weight curves of AUX-TS under pre-training and fine-tuning in Figure 4, we observe that the weights of different tasks change during the training. The weight of the target task increases with the training, while the weights of the two auxiliary tasks increase first and then decrease and AUX-TS assigns higher weights to the attribute generation task. Meanwhile, both the losses of the target task and the attribute generation task decrease continually while the loss of the edge generation task starts to increase at a certain stage instead. Although the loss of the edge generation task increases, we have achieved the best results in the target task. This proves that AUX-TS could learn the usefulness of different auxiliary tasks and can adaptively select auxiliary tasks with appropriate weights for the target task.
Gradient similarity is a good indicator to describe the consistency between tasks, and it changes with learning. The curve of gradient similarity for Fixed Weights under two transfer schemes (as shown in Figure 5) reflects that, in multi-task learning, the gradient similarities between auxiliary tasks and the target task rise rapidly in the early training stage, and then decrease with the progress of training. In fine-tuning stage, after the knowledge of the pre-trained model is transferred to the downstream task, the gradient similarities drop to a much smaller value. Fixed weights or inappropriate weights of auxiliary tasks may not be helpful or even cause negative transfer, especially in fine-tuning stage. This further explains why some other weighting methods are not effective on OAG_CS when the model is pre-trained on a huge graph. Whereas AUX-TS can perform well in both schemes since it considers the consistency of an auxiliary task to the target task.
Cosine Similarity method also uses the gradient similarity to adjust the weight. However, the gradient similarity fluctuates greatly, and there is not a simple positive correlation between the usefulness of auxiliary tasks and gradient similarity. Not only data with large gradient similarity can help the target task, but small ones can also be beneficial. Cosine similarity uses only those whose gradient similarity is greater than zero. So this rule-based method is not effective enough and far from optimal to utilize this important signal to learn weights. Whereas in AUX-TS it’s learnable via meta-learning and the advantage of AUX-TS is that it can dynamically update the weighting model to adapt to the latest training situation. We find that in multi-task learning, the gradient similarity is relatively large (-0.3 0.9), the weight change curve and the gradient similarity curve show a certain positive correlation, but in pre-training and fine-tuning, the similarity is relatively small (-0.2 0.4), this correlation becomes weak or even the opposite.
|Loss Scales||No Auxiliary Losses||SELAR||AUX-TS (all sims)|
AUX-TS performs well regardless of the loss scale. To further prove advantages of AUX-TS over SELAR which only uses loss in the weighting model without gradient similarity, we adjust loss scales of the two auxiliary tasks. Table 4 shows that SELAR is very sensitive to the loss scale and performs poorly on readjusted losses, while AUX-TS consistently performs well in various loss scale settings.
As shown in Figure 6, by comparing the weight curves under different loss scales, it is proved that AUX-TS could adaptively balance weights of different tasks under different loss scales. It assigns lower weights to tasks with high loss scales, which is different from SELAR. We observe a certain consistency between weight and loss in SELAR. SELAR tends to give high weights to hard examples with a large loss more than easy ones. However, the loss which represents the difficulty of a task is insufficient to characterize the importance of the task. The reason why AUX-TS could outperform SELAR significantly is that it not only consider the difficulty of a task but also consider the consistency with the target task.
In this work, we propose a new generic transfer learning paradigm on GNNs which would adaptively add multiple types of auxiliary tasks in the fine-tuning stage to continually incorporate various priors . It can improve the effectiveness of knowledge transfer and narrow the gap between pre-training and fine-tuning. Furthermore, we design an adaptive auxiliary loss weighting model to learn the weights of auxiliary tasks by quantifying the consistency between auxiliary tasks and the target task. The model is learnable through meta-learning. Comprehensive experiments on various downstream tasks and transfer learning schemes prove the generality and superiority of our methods compared with other methods. The effectiveness of our approach is demonstrated on GNNs but not limited to GNNs scenarios. It’s a generic transfer learning paradigm that could be flexibly extended to various transfer learning scenarios such as in NLP or computer vision (CV).
For OAG_CS and Reddit datasets, we run experiments with three seeds and evaluate 40 batches for each run to get more stable results.
For hyper-parameters used in OAG_CS and Reddit, we set the batch size to 256, the learning rate for GNNs to 1e-3/5e-4 respectively for pre-training and fine-tuning, and set the learning rate for the weighting model to 1e-5. We tried sigmoid and softplus as the activation function for the last layer of the weighting model. In most cases, softplus can achieve better performance. We do not use cross validation for experiments on OAG_CS and Reddit, but use 3-fold cross validation on Last-FM and Book-Crossing datasets.
We use five meta-path prediction tasks as auxiliary tasks for Last-FM and Book-Crossing datasets, and use the meta-paths label data published by SELAR. According to SELAR, these auxiliary tasks are listed in 5.
|Meta-paths (Last-FM)||Meta-paths (Book-Crossing)|