Missing-Class-Robust Domain Adaptation by Unilateral Alignment for Fault Diagnosis

01/07/2020 ∙ by Qin Wang, et al. ∙ ETH Zurich 0

Domain adaptation aims at improving model performance by leveraging the learned knowledge in the source domain and transferring it to the target domain. Recently, domain adversarial methods have been particularly successful in alleviating the distribution shift between the source and the target domains. However, these methods assume an identical label space between the two domains. This assumption imposes a significant limitation for real applications since the target training set may not contain the complete set of classes. We demonstrate in this paper that the performance of domain adversarial methods can be vulnerable to an incomplete target label space during training. To overcome this issue, we propose a two-stage unilateral alignment approach. The proposed methodology makes use of the inter-class relationships of the source domain and aligns unilaterally the target to the source domain. The benefits of the proposed methodology are first evaluated on the MNIST→MNIST-M adaptation task. The proposed methodology is also evaluated on a fault diagnosis task, where the problem of missing fault types in the target training dataset is common in practice. Both experiments demonstrate the effectiveness of the proposed methodology.



There are no comments yet.


page 1

Code Repositories

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

In recent years, deep learning methods 

[aizenberg2001multi, lecun2015deep, goodfellow2016deep] have achieved some remarkable results on various tasks [he2016identity, devlin2019bert, girshick2015fast]. However, the methods require not only large training datasets, but also labels to learn the relevant patterns in the data. This data-intensive nature of deep learning methods and particularly the requirement of labels, which can be expensive or even impossible to acquire, has limited their utilization in practical applications. In addition, the trained models usually don’t generalize well if a distribution shift is encountered between training and test data.

Unsupervised domain adaptation techniques [fernando2013unsupervised, ganin2014unsupervised, long2015learning, sun2015subspace] provide a promising solution to alleviate both challenges: missing labels and domain shift. Domain adaptation aims at leveraging unlabeled target data to improve the model’s generalization ability in the target domain. It allows knowledge transfer from a source domain to a different but related target domain [pan2011domain]. Recently, adversarial domain adaptation approaches [ganin2014unsupervised, tzeng2015simultaneous, ganin2016domain, tzeng2017adversarial, luo2017label, long2018conditional] have significantly improved domain adaption performance by aligning source and target data in an adversarial way and enforcing domain-invariant features in the latent space.

Fig. 1: Effect of different numbers of missing classes in the target training set on adversarial domain adaptation approach on the MNIST MNIST-M task: Baseline model trained using only source data (Blue). Model trained using traditional adversarial domain adaptation (Red). The proposed unilateral adaptation that mitigates the misalignment risk caused by the missing classes (Green) .

These unsupervised domain adaptation methods relax the need of labels on the target domain by transferring knowledge across domains. However, they typically assume that the label space of source and target data is identical [saito2018open]. This assumption imposes a significant limitation for real applications since the training dataset may not contain the complete set of classes. For example, in the industrial fault diagnosis, it would be beneficial to be able to transfer the source knowledge without having to wait for the occurrence of every possible fault in the target domain. :

Direct adversarial domain alignment results in a large uncertainty on the model performance in case of missing classes in the target domain. The domain alignment is, in this case, performed between a source domain with a complete set of labels and a target training set containing only a subset of labels. The alignment results in a negative effect on the model performance on the missing classes in the target training set. One additional drawback of the direct domain alignment is the negative effect on the inter-class relationship. When domain adaptation techniques are applied on the non-identical label space, the alignment effect is different among the classes that are present in the target domain during training and the missing classes. This means that the inter-class relationship for the aligned domains is likely to be distorted in an unpredictable way from the original inter-class relationship in the source domain. It is expected that, this misalignment effect is larger when there are more missing classes in the target domain. We demonstrate an example of this phenomenon using the MNIST MNIST-M adaptation task in Figure 1 with a varying number of missing classes.

In order to mitigate this negative alignment effect and better transfer the learned inter-class relationships from source to target domain, we propose to unilaterally align target domain towards the source domain instead of aligning both to an unknown intermediate space. The ultimate goal of the proposed approach is to make adversarial domain adaptation robust against missing classes in the target domain.

The intuition of the approach is to make use of the discriminative information learned from source domain in a better way. Since we don’t have access to samples of some classes in the target domain, we can only infer them by using the inter-class relationships we learned from the source domain. We argue that this approach will provide a more robust representation for the target domain.

The proposed method is a two-stage process: 1) We first train an anchor model based on source data only and extract pre-trained source features. 2) By minimizing the distance between the source features and the pre-trained features, while performing source and target feature alignment, the target distribution is unilaterally transformed to match the source distribution.

We show the improvement on the MNIST MNIST-M adaptation task in Figure 1 achieved by the proposed method with a varying number of missing classes in the target dataset.

To summarize, we propose a solution for the domain adaptation problem with missing classes in the target training data, while the model performance is still evaluated on all classes in the target domain. Similar to previous methods, we align the features in an adversarial way. However, different from previous methods, we align them unilaterally towards the source.

The main contribution of this paper is the proposed methodology to preserve the inter-class relationships of the source domain by aligning the target domain distribution unilaterally towards the source domain instead of jointly transforming the two domains to a common distorted latent space. The proposed methodology is particularly beneficial in problem setups with an incomplete set of class labels in the target training dataset.

We validate the effectiveness of the proposed methods first on the MNISTMNIST-M transfer task. For this alignment task, the proposed methodology demonstrates its robustness to the incompleteness of the target label space and achieves a similar level of performance in case of missing classes as with the complete set of classes.

To demonstrate the applicability of the proposed methodology in practical applications, we additionally evaluate our method on the task of transferring the learned knowledge between two different operating conditions for the fault diagnosis tasks on a bearing dataset. By applying the proposed unilateral alignment methodology, we are able to improve diagnosis performance on the bearing dataset.

Ii Related Work

Ii-a Domain Adaptation

Domain adaptation has been successfully applied in fields such as computer vision and natural language understanding 

[patel2015visual, csurka2017domain, pan2011domain, long2015learning, li2016revisiting, ganin2014unsupervised, saito2018maximum]. One common idea underlying different domain adaptation methods is the alleviation of the distribution discrepancy between source and target data, or in other words, they aim at aligning the source and target distributions. Different approaches have been proposed to address this task. [pan2011domain] proposed to use transfer components analysis across domains. [long2015learning]

proposed to align the source and target distributions by minimizing the Maximum Mean Discrepancy (MMD). Driven by a similar motivation, Adaptive Batch Normalization (AdaBN) 

[li2016revisiting] and AutoDial [carlucci2017autodial]

align the distributions via modified batch normalization layers. Domain Adversarial Neural Networks (DANN) 

[ganin2014unsupervised] aim at aligning the distributions by using a domain discriminator and train the model adversarially in order to make the feature space indistinguishable for the different domains.

Ii-B Missing Class in Partial Domain Adaptation

Another related topic is partial domain adaptation [cao2018partialcvpr]. In this setup, algorithms aim at solving the problem, where the target classes are a subset of source classes. [zhang2018deep] proposed to use importance weighted adversarial networks to focus on shared classes. [cao2018partial]

alleviates negative transfer by down-weighing the data of outlier source classes.

[cao2019learning] proposed to learn domain-invariant representations across domains and a progressive weighting scheme.

In partial domain adaptation setup, the missing target classes are not evaluated during test time. This is the key difference with the setup used in our paper.

Ii-C Domain Adaptation in Fault Diagnosis

Missing class problem is especially severe when adapting domains or operating conditions for fault diagnosis problems. Without considering missing classes, domain adaptation methods [zhang2017new, zhang2018adversarial, wang2019domain, li2019multi, michau2019, yang2019intelligent, 8643085] have recently been introduced to the fault diagnosis problems. Several approaches [lu2016deep, li2018cross] have been proposed to deal with missing-classes in the the context of fault diagnosis. However, they assume that the target training dataset contains exactly one class (the healthy condition). The main difference with our paper is that our proposed method is able to deal with different number of missing classes, meaning that the proposed method is more general as we don’t assume the target training data all from healthy condition.

Iii Problem Description

Formally, we mainly consider the following unsupervised domain adaptation task with missing classes.

  • Training data from source domain with all classes

  • Unlabeled training data from target domain with missing classes

  • Test data from target domain with all classes

where is the complete set of classes and the labels of target training set only contains a subset of it. Note that for the target training set, it is unknown which classes belong to . The only assumption is . Therefore, these samples are also part of the test set since their correct classification also needs to be evaluated. This evaluation of the classification accuracy of samples that were used for the alignment is in fact similar to that used by [li2018cross] and is common in transductive [gammerman1998learning, arnold2007comparative] domain adaptation problems.

Iv Proposed Methodology

The missing classes in the target training set makes the direct application of standard domain adaptation methods difficult. The shared idea behind most methods is to transform both source and target into a shared feature space. Such an alignment requires sufficient support, from both source and target, and also from all classes. The lack of information on the target domain can lead to an unexpected behavior of the alignment. One of the potential issues is that the alignment is explicitly changing the distribution of the given classes in the target domain, while no supervised domain guidance is given for the missing classes in the target domain. This unbalanced alignment behavior between present and missing classes may distort the well-learnt source inter-class relationships, and thus make the alignment sub-optimal or even deteriorate the model performance.

In order to fully leverage the limited healthy data from target domain and improve model performance on all classes, we propose a two-stage framework. We first learn a classification model in the source domain, and extract relevant features for source data. In the second stage, we apply domain adversarial adaptation techniques and align the source and target data. We strengthen the alignment by making sure the alignment is unilateral, that is, forcing the aligned features to be as close as possible to those learned in the first step.

We visualize the proposed method in Figure 2.

Fig. 2: The proposed two-stage unilateral alignment method. (Stage 1) We extract source features using a separate network. (Stage 2) We unilaterally align the distributions by adding the consistency loss. The loss is computed using current calculated source features and its corresponding pre-trained features.

Iv-a Stage 1: Extract Inter-class Relationship via Source-only Learning

Since the missing classes are unavailable for the target domain, it is only possible to learn inter-class relationships from source data. We thus propose to pre-train a separate neural network to extract this relationship. We assume that we have the same backbone architecture as the main network used in the second stage: a feature extractor parameterized by

, and a classifier

parameterized by

. We, thus, apply a standard supervised training procedure to learn the relationship. Formally, we train this separate network by using the following loss function:

where is the softmax cross entropy loss [bishop2006pattern, goodfellow2016deep] function, which is widely used in supervised classification problems. The network is trained only source data . This stage-1 network is frozen after this stage of training.

After a successful training on the network in stage 1, we can extract the pre-trained source features for each source training sample . These features contain meaningful inter-class relationships because a simple classifier is able to make strong predictions for source data. These features are then used as a reference for the next training step.

Iv-B Stage 2: Unilateral Adversarial Domain Adaptation

Iv-B1 Off-the-shelf Adversarial Domain Alignment

Motivated by the successful applications of DANN in computer vision [ganin2014unsupervised, ganin2016domain] and its later adoption in industrial application [wang2019domain]. We propose to make use of this off-the-shelf adaption technique in our adaptation task with missing classes.

As shown in Figure 5, our main architecture has three components: the feature extractor , the classifier , and an additional discriminator . The alignment is achieved by introducing the discriminator to distinguish between healthy features coming from the source and those from the target. Meanwhile, the feature extractor is encouraged to fool the discriminator so that the features are unbiased towards their origin. Formally, this is equivalent to the following mini-max problem [ganin2014unsupervised]:

where is the source data and is the target data, is again the softmax cross entropy loss function. is the cross entropy loss for the domain classification subtask. The objective function is similar to that of Generative Adversarial Networks (GAN) [goodfellow2014generative]

. It includes two parts: a classification loss for supervised learning and a domain adversarial loss for alignment.

We minimize the classification loss, w.r.t. parameters of the feature extractor and classifier. In addition to this supervised loss, we maximize the adversarial alignment loss w.r.t. parameters of the feature extractor in order to achieve domain-invariant features. We further minimize the adversarial alignment loss w.r.t. the discriminator, and thus, train the discriminator to provide a precise prediction of the origin of features.

It has been shown in [ganin2014unsupervised] that by reversing feature extractor’s gradients before sending them to the domain classifier, we can reformulate the problem and alleviate the -Divergence between source and target distributions [ben2010theory, cortes2011domain]. We adopted this optimization technique for our domain adaptaion problem with missing classes. Using the gradient reverse layer (GRL) from DANN [ganin2014unsupervised, ganin2016domain], the above loss function can be rewritten as:

Thus, it is able to be trained as an end-to-end learning problem.

Input : Source and target training samples.
  1. Stage 1: Train a source feature extractor and classifier using only source data, then freeze this pre-trained network. Extract reference source features from the pre-trained network.

  2. Stage 2: Unilateral Adversarial Domain Adaptation

    1. Initialize a new feature extractor , a new classifier, and a domain discriminator.

    2. Calculate the classification loss using source data.

    3. Calculate the consistency loss using source data and pre-trained features.

    4. Reverse gradients for healthy features from both source and target, then calculate the adversarial alignment loss

    5. Calculate the overall loss and gradients, update accordingly.

    6. Go back to b) for next iteration training until convergence.

  3. Evaluate the learnt model on the target test set with all classes.

Algorithm 1 Training procedure for the proposed unilateral alignment method.

Iv-B2 Unilateralism as an Additional Loss Term

The DANN method is directly aligning the complete source data with a target data that has missing classes. Thus a significant misalignment is expected. In order to avoid the potential negative effect of the above alignment and preserve the inter-class relationships while applying domain adaptation techniques, we propose to unilaterally align the target distribution to the corresponding part of the source distribution, instead of aligning both to a shared new space. We consider the pre-trained source features as a good representation of all classes, since class separability could be achieved. To transfer this good representation, in stage 2, an additional constraints is applied to force the aligned source features to be as close as possible to the pre-trained one. If the alignment is successful, then the target features should also be aligned with the pre-trained source features.

In order to preserve the inter-class relationships while applying the partial domain-adversarial alignment, we make use of the pre-trained source features , and force the aligned source features to be close to the pre-trained ones:

where K is the number of features in the feature space. We add this additional constraint to the loss function described in the previous paragraph. The overall loss function becomes thereby:

This additional loss is inspired by the consistency loss introduced in [chen2018domain] where a similar distance is used to improve the cross-domain robustness of the bounding box predictor for object detection tasks. However, it is used for a different purpose here since we are trying to encourage the alignment in one direction and preserve the inter-class relationships. We tested the additional loss using both and loss and found no significant difference between them.

Iv-C Summary

To summarize, in addition to aligning the source and target distributions via DANN, we propose to impose an additional constraint to make the alignment unilateral towards the pre-trained source features. The main objective of the unilateral alignment is to preserve the inter-class relationships learned from the source data, where knowledge on all classes are available.

V Experiments on MNIST MNIST-M Task

Fig. 3: Training samples from the MNIST MNIST-M task. Figure taken from [ganin2016domain].

Following the experiment setup used by [ganin2016domain], we first evaluate our model using the MNIST [lecun1998gradient] MNIST-M [ganin2016domain] task 111 Our code for this experiment is available at https://github.com/diagnosisda/dxda. The MNIST-M dataset blends digits from MNIST over patches randomly extracted from color photos from BSDS500 [arbelaez2010contour]. Samples from both datasets are shown in Figure 3. It is a classification task with 10 digits as 10 classes. The publicly-available implementation222https://github.com/pumpikano/tf-dann of DANN is used for our MNIST experiments.

We evaluate the proposed unilateral alignment model against a source-only baseline and the DANN alignment without unilateral constraint. We conduct experiments on different level of missing classes as shown in Table I. For example, in the setup, the unlabeled target training data contains samples from label . In the setup, the unlabeled target training data contains samples from label

. The source-only baseline is trained for 10 Epochs. Both DANN and DANN+Unilateral are trained for 50 Epochs. We adopt the training parameters used by 

[ganin2016domain]. We use a batch size of 64 and learning rate of 0.01.

Results on this task are shown in Table I. All reported results are based on the average accuracy of five different runs. As shown in the third column in the table, without the unilateral constraint, the benefit of the alignment is significantly decreasing when there are more missing classes. This is because the misalignment effect is larger when more classes are missing. By simply adopting our additional unilateral loss, the inter-class relationship learned from the source domain is strengthened, and the adaptation is significantly improved. When there are 9 out of 10 classes missing in the target training set, the unilateral model performance is less than worse than the full alignment, comparing to the degradation without our additional loss. This demonstrates the effectiveness of our proposed method.

Note that in the last row, when there is no missing classes, the additional unilateral loss is limiting the adaptation ability of DANN, thus slightly hurts the model performance.

# of missing class Baseline DANN [ganin2016domain] DANN+Unilateral (Ours)
9 / 10 57.49% 27.03% 72.99%
8 / 10 57.49% 54.72% 74.06%
7 / 10 57.49% 72.76% 74.72%
6 / 10 57.49% 72.97% 74.40%
5 / 10 57.49% 73.91% 74.99%
2 / 10 57.49% 74.76% 75.03%
0 / 10 57.49% 76.95% 76.74%
TABLE I: Experiment results on the MNIST MNIST-M task.

Vi Experiments on Fault Diagnosis Problems

In the following section, we demonstrate the benefits of the proposed approach on a different fault diagnosis task. Fault diagnosis is a classification task where our method can be directly applied on. Usually the label space consists of healthy and fault conditions. In this section, we consider the following variant of fault diagnosis problem:

Unsupervised domain adaptation when of classes are missing in the target training set.

[b] Fault Class Label 0 1 2 3 4 5 6 7 8 9 Loc NA IF IF IF BF BF BF OF OF OF Size 0 7 14 21 7 14 21 7 14 21

  • Fault location not applicable because class 0 is the healthy state.

TABLE II: Class Definition for the CWRU Dataset [smith2015rolling]. Table taken from [wang2019domain].

Vi-a Dataset

We conduct our experiments on the fault diagnosis dataset: the Case Western Reserve University (CWRU) [smith2015rolling].

Vi-A1 Cwru

The publicly-available bearing dataset from CWRU is used. It is a widely used dataset in the field of domain adaptation for fault diagnosis [li2018cross, lu2016deep, zhang2018adversarial, wang2019domain].

We follow the setup used by [li2018cross] whenever possible. Thus drive-end accelerometer data are used as our input. The list of labels concerned in this paper is shown in Table II, namely three different fault types, along with one healthy state are considered. IF stands for Inner race fault, BF stands for Bearing fault, and OF stands for outer race fault. Each fault type contains three sub-types, with fault diameters of 7, 14, 21 mils. Sampling rate of 12 kHz is adopted. Whenever data are unavailable at 12 kHz, we down-sample them to ensure a consistent sampling rate of 12 kHz in all experiments.

There are four different loads in the CWRU dataset. Domain adaptation is applied across these four different loads. For example, Task means working load 0 is the source domain with labeled training samples, and working load 1 is the target domain we want to improve model performance on.

Vi-B Pre-processing

For the CWRU dataset, we follow the same pre-processing steps as in [li2018cross, wang2019domain]. As shown in Figure 4

, first, we downsample and truncate each raw recording. Second, we divide each it into 200 sequences of 1024 points. Finally, using the Fast Fourier Transform 


, each sequence is converted into a vector of 512 Fourier coefficients.

Vi-C Model Implementation

We visualize the details of the backbone model and our discriminator in Figure 5. We use the same architecture following [li2018cross, wang2019domain] to enable a fair comparison.

Fig. 4: Preprocessing steps taken from [li2018cross, wang2019domain].

The backbone architecture [li2018cross] contains two components: a feature extractor, and a classifier. Each convolutional layer has a filter length of 3, and a hidden size of 10. Dropout layers are added accordingly afterwards with a dropout rate of 0.5. The signal is then flattened and transformed to features of size 256 by a fully-connected layer. The classifier is a single hidden layer network of size 256, using the softmax cross-entropy loss.

We train this backbone network as our baseline model using source load data only. To implement our models, a discriminator is needed additionally which comprises two fully connected hidden layers followed by the softmax cross-entropy loss.

The architectures of the backbone network and of the additional discriminator are illustrated in Fig 5

. The CWRU models are trained for 2000 Epochs with Sigmoid activation function. We report 5-run average accuracy and standard deviation.

Vi-D Experiment Results

Vi-D1 Unsupervised Domain Adaptation Experiment on CWRU Dataset with 80% Missing Classes

In this experiment, the target training dataset is composed of a subset of the classes. We would like to again emphasize that none of the target class labels are used for training. Formally, we consider the following set of unlabeled training data from the target machine

for the missing classes experiments. We consider for our demonstration.

Fig. 5: Backbone network used in all our fault diagnosis experiments. The feature extractor and classifier are taken from [li2018cross, wang2019domain].

Note that in this setup we choose arbitrarily the first 2 classes of the dataset () to demonstrate the effectiveness of our unilateral alignment approach, but one may expect similar behavior when the subset of fault types is chosen differently.

[b] Setup Baseline DANN [ganin2016domain] DANN+Unilateral Task 0-1 93.49 1.75 92.40 1.84 96.36 0.81 Task 0-2 93.65 0.96 92.86 1.53 97.38 1.66 Task 0-3 91.02 1.92 92.84 1.66 95.78 1.90 Task 1-0 97.93 0.24 96.92 0.74 97.49 0.52 Task 1-2 100.00 0.00 99.96 0.06 99.94 0.04 Task 1-3 98.26 1.63 99.50 0.09 99.58 0.15 Task 2-0 91.63 1.82 93.49 0.79 93.77 2.09 Task 2-1 97.09 0.92 97.56 0.19 97.60 0.36 Task 2-3 99.78 0.17 99.90 0.09 99.86 0.08 Task 3-0 87.96 0.18 88.41 0.22 88.42 0.72 Task 3-1 89.42 0.96 90.53 1.14 93.45 1.68 Task 3-2 99.65 0.17 99.14 0.90 99.83 0.06 Average 94.99 95.29 96.62

  • Reported numbers are based on average and standard deviation over five runs.

  • In training phase, labeled source samples from all classes and unlabeled target samples from the first 20% classes are provided. In test phase, target samples from all classes, including those that are unseen in target during training are evaluated.

TABLE III: Unsupervised Domain Adaptation Experiment Result on CWRU Dataset with 80 Missing Classes on CWRU Dataset

The results under this new setup are shown in Table III, for the source only baseline, DANN and DANN with our additional unilateral constraint. Compared to the baseline, DANN alone does not provide a significant improvement. This is likely due to the negative effect of trying to align the source features of all classes with target features containing only 20% of the classes. This could result in a distortion of the inter-class relationships.

The proposed unilateral alignment method, tackles this negative effect and strengthens the alignment. By adding the additional consistency loss, it provides an additional 1.33% absolute accuracy improvement over the naive implementation of DANN.

Vi-D2 Discussion

The performed fault diagnosis experiments on the case study demonstrate that the unilateral alignment is able to improve model performance of domain adaptation problems, when there are missing classes in the target training set.

There are few cases where we observe that unilateral alignment may slightly hurt the performance. In such cases, results show that the alignment was actually not required: the baseline is already providing a very high accuracy. Our results demonstrate that in such cases, the drop in accuracy is very small or even insignificant while in many other cases, the gain in accuracy is significant. Overall, the results are significantly improved by the proposed methodology.

Vii Conclusions

In this paper, we demonstrated that when there are missing classes in the target training dataset, directly applying adversarial domain adaptation techniques results in performance decrease. To overcome this problem, we proposed the unilateral alignment, a simple yet effective training strategy that leverages the inter-class relationships of the source domain. We showed in the MNIST experiment that by adding the additional consistency loss that enforces the unilateral alignment the model is able to be robust against missing classes. The additional experiments on fault diagnosis tasks show a promising potential of the proposed domain adaptation method for industrial applications where the problem of missing classes imposes a significant limitation on the applied approaches. Exploring the performance of the proposed models in case of corrupted samples is one of the future directions.


The authors thank Prof. Wen Li for fruitful discussions.