Object category learning and retrieval with weak supervision

01/26/2018 ∙ by Steven Hickson, et al. ∙ 0

We consider the problem of retrieving objects from image data and learning to classify them into meaningful semantic categories with minimal supervision. To that end, we propose a fully differentiable unsupervised deep clustering approach to learn semantic classes in an end-to-end fashion without individual class labeling using only unlabeled object proposals. The key contributions of our work are 1) a kmeans clustering objective where the clusters are learned as parameters of the network and are represented as memory units, and 2) simultaneously building a feature representation, or embedding, while learning to cluster it. This approach shows promising results on two popular computer vision datasets: on CIFAR10 for clustering objects, and on the more complex and challenging Cityscapes dataset for semantically discovering classes which visually correspond to cars, people, and bicycles. Currently, the only supervision provided is segmentation objectness masks, but this method can be extended to use an unsupervised objectness-based object generation mechanism which will make the approach completely unsupervised.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 5

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Unsupervised discovery of common patterns is a long standing task for artificial intelligence as shown in

Barlow (1989); Bengio, Courville, and Vincent (2012)

. Recent deep learning approaches have offered major breakthroughs in classification into multiple categories with millions of labeled examples (e.g. 

Krizhevsky (2009); Szegedy et al. (2015); He et al. (2016) and many others). These methods rely on a lot of annotated data for training in order to perform well. Unfortunately, labeling is an inefficient and expensive progress, so learning from unlabeled data is desirable for many complex tasks. At the same time, much of human knowledge and learning is obtained by unsupervised observations Grossberg (1994).

The goal of this work is to show that semantically meaningful classes can be learned with minimal supervision. Given a set of objectness proposals, we use the activations of foreground objects in order to learn deep features to cluster the available data while simultaneously learning the embedding in an end-to-end manner. More specifically, we propose a differentiable clustering approach that learns better separability of classes and embedding. The main idea is to store the potential cluster means as

weights

in a neural network at the higher levels of feature representation. This allows them to be

learned jointly with the potential feature representation. This differentiable clustering approach is integrated with Deep Neural Networks (e.g. Szegedy et al. (2015)) to learn semantic classes in an end-to-end fashion without manual class labeling.

The idea of doing this ‘end-to-end’ is that gradient descent can not only learn good weights for clustering, it can also change the embedding to allow for better clustering without the use of labels. We see that this leads to better feature representation. Our results show also that different object categories emerge and can later be retrieved from test images never before seen by the network, resulting in clusters of meaningful categories, such as cars, persons, bicycles.

In this work we use given segmentation objectness masks, which are candidate objects without labels. This can be extended by using an independent objectness-based object generation mechanism Pathak et al. (2017); Faktor and Irani (2014) or by using unsupervised motion segmentation in videos or structure from motion Vijayanarasimhan et al. (2017).

2 Related Work

Unsupervised learning (Barlow (1989)) and unsupervised deep learning (Bengio, Courville, and Vincent (2012)Bengio (2012)Bengio and others (2009)

) are central topics to Machine Learning. Unsupervised deep learning has been shown to improve results on classification tasks per

Erhan et al. (2010)

, especially given small datasets and complicated high dimensional data such as video. This has been explored by many representations including sequence to sequence learning and textual representations (

Radford, Jozefowicz, and Sutskever (2017), Ramachandran, Liu, and Le (2016)).

Our work focuses on unsupervised deep learning for discovering visual object categories. This has also been shown to improve results such as in Doersch and Zisserman (2017). Unsupervised discovery of visual objects has been a large topic of interest in computer vision (Sivic et al. (2005); Russell et al. (2006); Singh, Gupta, and Efros (2012); Bach and Jordan (2005); Kwak et al. (2015); Pathak et al. (2017)).

Building specialized, deep embeddings to help computer vision tasks is also a popular approach  such as in Agrawal, Carreira, and Malik (2015)

. Transfer learning from supervised tasks has proven to be very successful. Further,  

Agrawal, Carreira, and Malik (2015) propose learning the lower dimensional embedding through unsupervised learning and show improved performance when transfered to other supervised tasks.

Despite the popularity of building different embeddings, there is little work investigating the use of clustering to modify the embedding in an end-to-end deep learning framework. Bottou and Bengio (1995) investigate a differentiable version of the kmeans algorithm and examine its convergence properties. Our work focuses on learnable feature representations (instead of fixed ones as in Bottou and Bengio (1995)) and introduces memory units for the task.

3 Unsupervised deep clustering

Our unsupervised deep clustering is inspired by Bottou and Bengio (1995), who consider differentiable clustering algorithms.

We differ from this approach because the features we cluster also change with backpropogation. In our work, we add a kmeans-like loss that is integrated end-to-end. Our idea is to store the potential cluster means as weights in the network and thus have them be learned.

The proposed clustering is done simultaneously while building an embedding. Given information of a potential object vs background (binary labels), clustering in a differentiable way provides a better embedding for the input data. We show that this method can be used for meaningful semantic retrieval of related objects.

3.1 Embedding with clustering

We train a convolutional neural network (CNN) to predict foreground and background using oracle labels of patches of objects and background images. Concurrently, we learn the clustering of objects by imposing constraints that will force the embedding to be partitioned into multiple semantically coherent clusters of objects without explicit labels for different objects.

For our experiments, we use random initialization on the fully-connected layers (the last two layers) and we add the differentiable clustering module after the second to last layer. Note that we only cluster the foreground labels as background activations are not of interest for clustering; the classifier can predict foreground vs background with high accuracy (above 90%).

The objective function is shown in Equation 1.

(1)

In this equation, is the number of samples, is the number of defined clusters, is the “weight” (theoretically and typically the mean of the cluster) for each , and is the activations from the fully connected layer before the classification fully connected layer. This is differentiable and the gradient descent algorithm is shown in Equation 2.

(2)

where and is the learning rate. We also add regularization over the weights to the loss . Furthermore, we use a custom clustering metric that informs us if the clusters are evenly distributed as defined in Equation 3 and Equation 4. This is important since we have no labels to evaluate our data during training.

(3)
(4)

The final loss to be optimized is shown in (Equation 5)

(5)

where and

are hyperparameters which are tuned during the training.

is the standard cross-entropy loss for our foreground vs background classification. For our method, we use and

. We apply this loss to every point that is labeled as potentially an object and ignore the background ones when clustering. This way we learn foreground vs background and then learn clustering of the foreground activations. Optimization was performed with a ‘RMSProp’ optimizer, with a learning rate of 0.045, momentum 0.9, decay factor 0.9, and

of 1.0.

4 Experimental evaluation

We experiment with a toy example using CIFAR10 and a more challenging example using Cityscapes.

4.1 CIFAR10 dataset

We first test the proposed unsupervised clustering approach on the CIFAR10 Krizhevsky (2009) dataset. The goal of this experiment is to test if clustering can uncover separate categories in a simple toy problem with a two class setting.

Clusters Automobile Dog
Cluster 0 68.5% 17.9%
Cluster 1 31.5% 82.1%
Table 1: Unsupervised clustering results on CIFAR10 for discovery of two classes. Per cluster accuracy for each of the two given classes on the test set (class labels are unknown during training).

We selected as an example the dog and automobile classes to label as foreground. We then train a network from scratch based on the Network in Network architecture (NiN) of Lin, Chen, and Yan (2013) from scratch for our experiments. All other classes of CIFAR are considered background for this experiment. By attaching our modified clustering objective function to the next to last layer, we attempt to cluster dog and automobile without labels.

We can see in our simple experimental results that classes are naturally clustered with the majority of examples correctly assigned. Table 1 shows quantitative results on the test set. As seen 68.5% of the automobile classes and 82.1% of the dog examples are correctly assigned to separate clusters. Note that in these cases, the concepts and classes of dog and automobile are unknown to the training algorithm and we are just looking at them after clustering for evaluation.

4.2 Cityscapes dataset

The Cityscapes dataset (Cordts et al. (2016)) is a large-scale dataset that is used for evaluating various classification, detection, and segmentation algorithms related to autonomous driving. It contains 2975 training, 500 validation, and 1525 test images, where the test set is provided for the purposes of the Cityscape competition only. In this work, we used the training set for training and the validation set for testing and visualizing results (as the test set has no annotation results). Annotation is provided for classes which represent moving agents in the scene, such as pedestrian, car, motorcycle, bicycle, bus, truck, rider, train. In this work we only use foreground/background labels and intend to discover semantic groups from among the moving objects.

4.3 Weakly supervised discovery of classes

In this experiment we considered the larger, real-life dataset, Cityscapes (Cordts et al. (2016)), described above to see if important class categories, e.g. the moving objects in the scene can be clustered into semantically meaningful classes. We extract the locations and extents of the moving objects and use that as weak supervision. Note the classes are uneven and car and person dominate.

We show results clustering 8 categories into 2 and 3 clusters despite the rarity of some of them (such as bicycle). All results below are presented on the validation set. We report the results in terms of the number of object patches extracted from the available test images. For this dataset, the CNN architecture is based on the Inception architecture proposed by Szegedy et al. (2015). Since there are a small number of examples, we pre-train only the convolutional layers of the network.

Results on clustering the 8 classes of moving objects into 2 and 3 clusters are presented in Table 2 and Table 3 respectively for the learned embedding by the proposed approach and the baseline embedding. The baseline embedding is calculated by fine-tuning the same architecture in the same manner, but without our loss (Equation 5); it uses the same amount of information as input as our embedding. For this experiment, we apply standard kmeans on both activations after training is completed. We see here that our method provides better clustering for the two dominant classes in the dataset (car and person). On the other hand, the baseline embedding clusters on one class only, similar to the two class case. We have consistently observed this behavior for different runs and hypothesize this is due to the sparse nature of the baseline embedding and it’s activations.

Figure 1 visualizes the three retrieved clusters (color-coded) when clustering into 3 clusters with our approach. We can see that people (in blue) and cars (in green) are often correctly retrieved. Bikes are more rare and may be more often mistaken, for example in cases where a portion of the patch contains part of a car, or since the bicycle very often has a person riding it. Still this is exciting result, given that it is learned by not providing a single class label during training.

Classes Cluster 0 Cluster 1
Person 4320 138
Rider 676 138
Car 1491 4399
Truck 60 69
Bus 49 89
Train 17 16
Motorcycle 88 205
Bicycle 795 787
Table 2: Unsupervised clustering of objects from Cityscapes using our method. The table shows number of examples assigned to each learned cluster (for K=2).
Our method Baseline
Classes Cluster 0 Cluster 1 Cluster 2 Cluster 0 Cluster 1 Cluster 2
Person 151 4315 17 4482 1 0
Rider 258 551 7 816 0 0
Car 5195 950 180 6312 13 0
Truck 89 39 5 131 2 0
Bus 127 20 5 152 0 0
Train 25 9 1 35 0 0
Motorcycle 127 76 4 207 0 0
Bicycle 1128 541 450 2119 0 0
Table 3: Unsupervised clustering on the Cityscapes dataset with 3 clusters. The table shows the number of examples assigned to each learned cluster. Our method (left) and baseline (right). Our method results in 69.98% accuracy.
Figure 1: Visualization of clusters learned by our method (for K=3). From the figure, the green class is responsible for retrieving cars, the blue one persons, and the red one bicycles. We can see that both cars and persons are discovered well but bicycles, a rarer class, can be confused with a person or with a partially visible car in the background.

5 Conclusions

We propose a differentiable clustering objective which learns to separate classes during learning and build a better embedding. The key idea is to be able to learn the clusters which are stored as weights, and simultaneously learn the feature representation and the clustering of the data. Our results show that the proposed approach is useful for extracting semantically related objects.

References

  • Agrawal, Carreira, and Malik (2015) Agrawal, P.; Carreira, J.; and Malik, J. 2015. Learning to see by moving. CVPR.
  • Bach and Jordan (2005) Bach, F. R., and Jordan, M. I. 2005.

    Learning spectral clustering.

    NIPS.
  • Barlow (1989) Barlow, H. 1989. Unsupervised learning. Neural computation.
  • Bengio and others (2009) Bengio, Y., et al. 2009. Learning deep architectures for ai. Foundations and trends® in Machine Learning 2(1):1–127.
  • Bengio, Courville, and Vincent (2012) Bengio, Y.; Courville, A. C.; and Vincent, P. 2012. Unsupervised feature learning and deep learning: A review and new perspectives. CoRR, abs/1206.5538.
  • Bengio (2012) Bengio, Y. 2012. Deep learning of representations for unsupervised and transfer learning. In Proceedings of ICML Workshop on Unsupervised and Transfer Learning, 17–36.
  • Bottou and Bengio (1995) Bottou, L., and Bengio, Y. 1995.

    Convergence properties of the k-means algorithms.

    In Advances in neural information processing systems, 585–592.
  • Cordts et al. (2016) Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; and Schiele, B. 2016.

    The cityscapes dataset for semantic urban scene understanding.

    In

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition

    , 3213–3223.
  • Doersch and Zisserman (2017) Doersch, C., and Zisserman, A. 2017. Multi-task self-supervised visual learning. arXiv preprint arXiv:1708.07860.
  • Erhan et al. (2010) Erhan, D.; Bengio, Y.; Courville, A.; Manzagol, P.-A.; Vincent, P.; and Bengio, S. 2010. Why does unsupervised pre-training help deep learning? Journal of Machine Learning Research 11(Feb):625–660.
  • Faktor and Irani (2014) Faktor, A., and Irani, M. 2014. Video segmentation by non-local consensus voting. BMVC.
  • Grossberg (1994) Grossberg, S. 1994. 3-d vision and figure-ground separation by visual cortex. Perception and Psychophysics.
  • He et al. (2016) He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. CVPR.
  • Krizhevsky (2009) Krizhevsky, A. 2009. Learning multiple layers of features from tiny images.
  • Kwak et al. (2015) Kwak, S.; Cho, M.; Laptev, I.; Ponce2, J.; and Schmid, C. 2015. Unsupervised object discovery and tracking in video collections. ICCV.
  • Lin, Chen, and Yan (2013) Lin, M.; Chen, Q.; and Yan, S. 2013. Network in network. arXiv preprint arXiv:1312.4400.
  • Pathak et al. (2017) Pathak, D.; Girshick, R.; Dollar, P.; Darrell, T.; and Hariharan, B. 2017. Learning features by watching objects move. CVPR.
  • Radford, Jozefowicz, and Sutskever (2017) Radford, A.; Jozefowicz, R.; and Sutskever, I. 2017. Learning to generate reviews and discovering sentiment. arXiv preprint arXiv:1704.01444.
  • Ramachandran, Liu, and Le (2016) Ramachandran, P.; Liu, P. J.; and Le, Q. V. 2016. Unsupervised pretraining for sequence to sequence learning. arXiv preprint arXiv:1611.02683.
  • Russell et al. (2006) Russell, B. C.; Efros, A. A.; Sivic, J.; Freeman, W. T.; and Zisserman, A. 2006. Using multiple segmentations to discover objects and their extent in image collections. CVPR.
  • Singh, Gupta, and Efros (2012) Singh, S.; Gupta, A.; and Efros, A. A. 2012. Unsupervised discovery of mid-level discriminative patches. ECCV.
  • Sivic et al. (2005) Sivic, J.; Russell, B. C.; Efros, A. A.; Zisserman, A.; and Freeman, W. T. 2005. Discovering objects and their location in images. ICCV.
  • Szegedy et al. (2015) Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; and Rabinovich, A. 2015. Going deeper with convolutions. CVPR.
  • Vijayanarasimhan et al. (2017) Vijayanarasimhan, S.; Ricco, S.; Schmid, C.; Sukthankar, R.; and Fragkiadaki, K. 2017. Sfm-net: Learning of structure and motion from video.