1 Introduction
Deep learning has recently received increasing attention from researchers and has been successfully applied to numerous realworld applications. Deep learning algorithms attempt to learn highlevel features from mass data, which make deep learning beyond traditional machine learning. It can automatic extract data features by unsupervised or semisupervised feature learning algorithm and hierarchical feature extraction. In contrast, traditional machine learning methods need to design features manually that seriously increases the burden on users. It can be said that deep learning is an representation learning algorithm based on largescale data in machine learning.
Data dependence is one of the most serious problem in deep learning. Deep learning has a very strong dependence on massive training data compared to traditional machine learning methods, because it need a large amount of data to understand the latent patterns of data. An interesting phenomenon can be found that the scale of the model and the size of the required amount of data has a almost linear relationship. An acceptable explanation is that for a particular problem, the expressive space of the model must be large enough to discover the patterns under the data . The preorder layers in the model can identify highlevel features of training data, and the subsequent layers can identify the information needed to help make the final decision.
Insufficient training data is a inescapable problem in some special domains. The collection of data is complex and expensive that make it is extremely difficult to build a largescale, highquality annotated dataset. For example, each sample in bioinformatics dataset often demonstration a clinical trial or a painful patient. In addition, even we obtain training dataset by paid an expensive price, it is very easy to get out of date and thus cannot be effectively applied in the new tasks.
Transfer learning relaxes the hypothesis that the training data must be independent and identically distributed (i.i.d.) with the test data, which motivates us to use transfer learning to against the problem of insufficient training data. In transfer learning, the training data and test data are not required to be i.i.d., and the model in target domain is not need to trained from scratch, which can significantly reduce the demand of training data and training time in the target domain.
In the past, most studies of transfer learning were conducted in traditional machine learning methods. Due to the dominance position of deep learning in modern machine learning methods, a survey on deep transfer learning and its applications is particularly important. The contributions of this survey paper are as follows:

We define the deep transfer learning and categorizing it into four categories for the first time.

We reviewing the current research works on each category of deep transfer learning, and given a standardized description and sketch map of every category.
2 Deep Transfer Learning
Transfer learning is an important tool in machine learning to solve the basic problem of insufficient training data. It try to transfer the knowledge from the source domain to the target domain by relaxing the assumption that the training data and the test data must be i.i.d. This will leads to a great positive effect on many domains that are difficult to improve because of insufficient training data. The learning process of transfer learning illustrated in the Fig. 1.
Some notations used in this survey need to be clearly defined. First of all, we give the definitions of a domain and a task respectively: A domain can be represented by , which contains two parts: the feature space
and the edge probability distribution
where . A task can be represented by . It consists of two parts: label space and target prediction function . can also be regarded as a conditional probability function . Then, the transfer learning can be formal defined as follows:Definition 1
(Transfer Learning). Given a learning task based on , and we can get the help from for the learning task . Transfer learning aims to improve the performance of predictive function for learning task by discover and transfer latent knowledge from and , where and/or . In addition, in the most case, the size of is much larger than the size of , .
Surveys [19] and [25] divide the transfer learning methods into three major categories with the relationship between the source domain and the target domain, which has been widely accepted. These suverys are good summary of the past works on transfer learning, which introduced a number of classic transfer learning methods. Further more, many newer and better methods have been proposed recently. In recent years, transfer learning research community are mainly focused on the following two aspects: domain adaption and multisource domains transfer.
Nowadays, deep learning has achieved dominating situation in many research fields in recent years. It is important to find how to effectively transfer knowledge by deep neural network, which called deep transfer learning that defined as follows:
Definition 2
(Deep Transfer Learning). Given a transfer learning task defined by . It is a deep transfer learning task where is a nonlinear function that reflected a deep neural network.
3 Categories
Deep transfer learning studies how to utilize knowledge from other fields by deep neural networks. Since deep neural networks have become popular in various fields, a considerable amount of deep transfer learning methods have been proposed that it is very important to classify and summarize them. Based on the techniques used in deep transfer learning, this paper classifies deep transfer learning into four categories: instancesbased deep transfer learning, mappingbased deep transfer learning, networkbased deep transfer learning, and adversarialbased deep transfer learning, which are shown in Table
1.Approach category  Brief description  Some related works 

Instancesbased  Utilize instances in source domain by appropriate weight.  [4], [27], [20], [24], [10], [26], [11] 
Mappingbased  Mapping instances from two domains into a new data space with better similarity.  [23], [12], [8], [14], [2] 
Networkbased  Reuse the partial of network pretrained in the source domain.  [9], [17], [15], [30], [3], [6], [28] 
Adversarialbased  Use adversarial technology to find transferable features that both suitable for two domains.  [1], [5], [21], [22], [13], [16] 
3.1 Instancesbased deep transfer learning
Instancesbased deep transfer learning refers to use a specific weight adjustment strategy, select partial instances from the source domain as supplements to the training set in the target domain by assigning appropriate weight values to these selected instances. It is based on the assumption that ”Although there are different between two domains, partial instances in the source domain can be utilized by the target domain with appropriate weights.”. The sketch map of instancesbased deep transfer learning are shown in Fig. 2.
TrAdaBoost proposed by [4] use AdaBoostbased technology to filter out instances that are dissimilar to the target domain in source domains. Reweighted instances in source domain to compose a distribution similar to target domain. Finally, training model by using the reweighted instances from source domain and origin instances from target domain. It can reduce the weighted training error on different distribution domains that preserving the properties of AdaBoost. TaskTrAdaBoost proposed by [27] is a fast algorithm promote rapid retraining over new targets. Unlike TrAdaBoost is designed for classification problems, ExpBoost.R2 and TrAdaBoost.R2 were proposed by [20] to cover the regression problem. Biweighting domain adaptation (BIW) proposed [24] can aligns the feature spaces of two domains into the common coordinate system, and then assign an appropriate weight of the instances from source domain. [10] propose a enhanced TrAdaBoost to handle the problem of interregional sandstone microscopic image classification. [26] propose a metric transfer learning framework to learn instance weights and a distance of two different domains in a parallel framework to make knowledge transfer across domains more effective. [11] introduce an ensemble transfer learning to deep neural network that can utilize instances from source domain.
3.2 Mappingbased deep transfer learning
Mappingbased deep transfer learning refers to mapping instances from the source domain and target domain into a new data space. In this new data space, instances from two domains are similarly and suitable for a union deep neural network. It is based on the assumption that ”Although there are different between two origin domains, they can be more similarly in an elaborate new data space.”. The sketch map of instancesbased deep transfer learning are shown in Fig. 3.
Transfer component analysis (TCA) introduced by [18] and TCAbased methods [29] had been widely used in many applications of traditional transfer learning. A natural idea is extend the TCA method to deep neural network. [23] extend MMD to comparing distributions in a deep neural network, by introduces an adaptation layer and an additional domain confusion loss to learn a representation that is both semantically meaningful and domain invariant. The MMD distance used in this work is defined as
(1) 
and the loss function is defined as
(2) 
[12] improved previous work by replace MMD distance with multiple kernel variant MMD (MKMMD) distance proposed by [8]
. The hidden layer related with the learning task in the convolutional neural networks (CNN) is mapped into the reproducing kernel Hilbert space (RKHS), and the distance between different domains is minimized by the multicore optimization method.
[14]propose joint maximum mean discrepancy (JMMD) to measurement the relationship of joint distribution. JMMD was used to generalize the transfer learning ability of the deep neural networks (DNN) to adapt the data distribution in different domain and improved the previous works. Wasserstein distance proposed by
[2] can be used as a new distance measurement of domains to find better mapping.3.3 Networkbased deep transfer learning
Networkbased deep transfer learning refers to the reuse the partial network that pretrained in the source domain, including its network structure and connection parameters, transfer it to be a part of deep neural network which used in target domain. It is based on the assumption that ”Neural network is similar to the processing mechanism of the human brain, and it is an iterative and continuous abstraction process. The frontlayers of the network can be treated as a feature extractor, and the extracted features are versatile.”. The sketch map of networkbased deep transfer learning are shown in Fig. 4.
[9] divide the network into two parts, the former part is the languageindependent feature transform and the last layer is the languagerelative classifier. The languageindependent feature transform can be transfer between multi languages. [17]
reuse frontlayers trained by CNN on the ImageNet dataset to compute intermediate image representation for images in other datasets, CNN are trained to learning image representations that can be efficiently transferred to other visual recognition tasks with limited amount of training data.
[15] proposed a approach to jointly learn adaptive classifiers and transferable features from labeled data in the source domain and unlabeled data in the target domain, which explicitly learn the residual function with reference to the target classifier by plugging several layers into deep network. [30] learning domain adaptation and deep hash features simultaneously in a DNN. [3] proposed a novel multiscale convolutional sparse coding method. This method can automatically learns filter banks at different scales in a joint fashion with enforced scalespecificity of learned patterns, and provides an unsupervised solution for learning transferable base knowledge and finetuning it towards target tasks. [6] apply deep transfer learning to transfer knowledge from realworld object recognition tasks to glitch classifier for the detector of multiple gravitational wave signals. It demonstrate that DNN can be used as excellent feature extractors for unsupervised clustering methods to identify new classes based on their morphology, without any labeled examples.Another very noteworthy result is that [28] point out the relationship between network structure and transferability. It demonstrated that some modules may not influence indomain accuracy but influence the transferability. It point out what features are transferable in deep networks and which type of networks are more suitable for transfer. Given an conclusion that LeNet, AlexNet, VGG, Inception, ResNet are good chooses in networkbased deep transfer learning.
3.4 Adversarialbased deep transfer learning
Adversarialbased deep transfer learning refers to introduce adversarial technology inspired by generative adversarial nets (GAN) [7] to find transferable representations that is applicable to both the source domain and the target domain. It is based on the assumption that ”For effective transfer, good representation should be discriminative for the main learning task and indiscriminate between the source domain and target domain.” The sketch map of adversarialbased deep transfer learning are shown in Fig. 5.
The adversarialbased deep transfer learning has obtained the flourishing development in recent years due to its good effect and strong practicality. [1] introduce adversarial technology to transfer learning for domain adaption, by using a domain adaptation regularization term in the loss function. [5] proposed an adversarial training method that suitable for most any feedforward neural model by augmenting it with few standard layers and a simple new gradient reversal layer. [21] proposed a approach transfer knowledge crossdomain and crosstask simultaneity for sparsely labeled target domain data. A special joint loss function was used in this work to force CNN to optimize both the distance between domains which defined as , where is classification loss, is domain adversarial loss. Because the two losses stand in direct opposition to one another, an iterative optimize algorithm are introduced to update one loss when fixed another. [22] proposed a new GAN loss and combine with discriminative modeling to a new domain adaptation method. [13] proposed a randomized multilinear adversarial networks to exploit multiple feature layers and the classifier layer based on a randomized multilinear adversary to enable both deep and discriminative adversarial adaptation. [16] utilize a domain adversarial loss, and generalizes the embedding to novel task using a metric learningbased approach to find more tractable features in deep transfer learning.
4 Conclusion
In this survey paper, we have review and category current researches of deep transfer learning. Deep transfer learning is classified into four categories for the first time: instancesbased deep transfer learning, mappingbased deep transfer learning, networkbased deep transfer learning, and adversarialbased deep transfer learning. In most practical applications, the above multiple technologies are often used in combination to achieve better results. Most current researches focuses on supervised learning, how to transfer knowledge in unsupervised or semisupervised learning by deep neural network may attract more and more attention in the future. Negative transfer and transferability measures are important issues in traditional transfer learning. The impact of these two issues in deep transfer learning also requires us to conduct further research. In addition, a very attractive research area is to find a stronger physical support for transfer knowledge in deep neural network, which requires the cooperation of physicists, neuroscientists and computer scientists. It can be predicted that deep transfer learning will be widely applied to solve many challenging problems with the development of deep neural network.
References
 [1] Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M.: Domainadversarial neural networks. arXiv preprint arXiv:1412.4446 (2014)
 [2] Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein gan. arXiv preprint arXiv:1701.07875 (2017)
 [3] Chang, H., Han, J., Zhong, C., Snijders, A., Mao, J.H.: Unsupervised transfer learning via multiscale convolutional sparse coding for biomedical applications. IEEE transactions on pattern analysis and machine intelligence (2017)
 [4] Dai, W., Yang, Q., Xue, G.R., Yu, Y.: Boosting for transfer learning. In: Proceedings of the 24th international conference on Machine learning. pp. 193–200. ACM (2007)
 [5] Ganin, Y., Lempitsky, V.: Unsupervised domain adaptation by backpropagation. arXiv preprint arXiv:1409.7495 (2014)
 [6] George, D., Shen, H., Huerta, E.: Deep transfer learning: A new deep learning glitch classification method for advanced ligo. arXiv preprint arXiv:1706.07446 (2017)
 [7] Goodfellow, I., PougetAbadie, J., Mirza, M., Xu, B., WardeFarley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: Advances in neural information processing systems. pp. 2672–2680 (2014)
 [8] Gretton, A., Sejdinovic, D., Strathmann, H., Balakrishnan, S., Pontil, M., Fukumizu, K., Sriperumbudur, B.K.: Optimal kernel choice for largescale twosample tests. In: Advances in neural information processing systems. pp. 1205–1213 (2012)
 [9] Huang, J.T., Li, J., Yu, D., Deng, L., Gong, Y.: Crosslanguage knowledge transfer using multilingual deep neural network with shared hidden layers. In: Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on. pp. 7304–7308. IEEE (2013)
 [10] Li, N., Hao, H., Gu, Q., Wang, D., Hu, X.: A transfer learning method for automatic identification of sandstone microscopic images. Computers & Geosciences 103, 111–121 (2017)
 [11] Liu, X., Liu, Z., Wang, G., Cai, Z., Zhang, H.: Ensemble transfer learning algorithm. IEEE Access 6, 2389–2396 (2018)
 [12] Long, M., Cao, Y., Wang, J., Jordan, M.: Learning transferable features with deep adaptation networks. In: International Conference on Machine Learning. pp. 97–105 (2015)
 [13] Long, M., Cao, Z., Wang, J., Jordan, M.I.: Domain adaptation with randomized multilinear adversarial networks. arXiv preprint arXiv:1705.10667 (2017)
 [14] Long, M., Wang, J., Jordan, M.I.: Deep transfer learning with joint adaptation networks. arXiv preprint arXiv:1605.06636 (2016)
 [15] Long, M., Zhu, H., Wang, J., Jordan, M.I.: Unsupervised domain adaptation with residual transfer networks. In: Advances in Neural Information Processing Systems. pp. 136–144 (2016)
 [16] Luo, Z., Zou, Y., Hoffman, J., FeiFei, L.F.: Label efficient learning of transferable representations acrosss domains and tasks. In: Advances in Neural Information Processing Systems. pp. 164–176 (2017)

[17]
Oquab, M., Bottou, L., Laptev, I., Sivic, J.: Learning and transferring midlevel image representations using convolutional neural networks. In: Computer Vision and Pattern Recognition (CVPR), 2014 IEEE Conference on. pp. 1717–1724. IEEE (2014)
 [18] Pan, S.J., Tsang, I.W., Kwok, J.T., Yang, Q.: Domain adaptation via transfer component analysis. IEEE Transactions on Neural Networks 22(2), 199–210 (2011)
 [19] Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Transactions on knowledge and data engineering 22(10), 1345–1359 (2010)
 [20] Pardoe, D., Stone, P.: Boosting for regression transfer. In: Proceedings of the 27th International Conference on International Conference on Machine Learning. pp. 863–870. Omnipress (2010)
 [21] Tzeng, E., Hoffman, J., Darrell, T., Saenko, K.: Simultaneous deep transfer across domains and tasks. In: Computer Vision (ICCV), 2015 IEEE International Conference on. pp. 4068–4076. IEEE (2015)
 [22] Tzeng, E., Hoffman, J., Saenko, K., Darrell, T.: Adversarial discriminative domain adaptation. In: Computer Vision and Pattern Recognition (CVPR). vol. 1, p. 4 (2017)
 [23] Tzeng, E., Hoffman, J., Zhang, N., Saenko, K., Darrell, T.: Deep domain confusion: Maximizing for domain invariance. arXiv preprint arXiv:1412.3474 (2014)

[24]
Wan, C., Pan, R., Li, J.: Biweighting domain adaptation for crosslanguage text classification. In: IJCAI ProceedingsInternational Joint Conference on Artificial Intelligence. vol. 22, p. 1535 (2011)
 [25] Weiss, K., Khoshgoftaar, T.M., Wang, D.: A survey of transfer learning. Journal of Big Data 3(1), 9 (2016)
 [26] Xu, Y., Pan, S.J., Xiong, H., Wu, Q., Luo, R., Min, H., Song, H.: A unified framework for metric transfer learning. IEEE Transactions on Knowledge and Data Engineering 29(6), 1158–1171 (2017)
 [27] Yao, Y., Doretto, G.: Boosting for transfer learning with multiple sources. In: Computer vision and pattern recognition (CVPR), 2010 IEEE conference on. pp. 1855–1862. IEEE (2010)
 [28] Yosinski, J., Clune, J., Bengio, Y., Lipson, H.: How transferable are features in deep neural networks? In: Advances in neural information processing systems. pp. 3320–3328 (2014)
 [29] Zhang, J., Li, W., Ogunbona, P.: Joint geometrical and statistical alignment for visual domain adaptation. In: CVPR (2017)
 [30] Zhu, H., Long, M., Wang, J., Cao, Y.: Deep hashing network for efficient similarity retrieval. In: AAAI. pp. 2415–2421 (2016)
Comments
There are no comments yet.