Learning Deep Visual Object Models From Noisy Web Data: How to Make it Work

02/28/2017 ∙ by Nizar Massouh, et al. ∙ 0

Deep networks thrive when trained on large scale data collections. This has given ImageNet a central role in the development of deep architectures for visual object classification. However, ImageNet was created during a specific period in time, and as such it is prone to aging, as well as dataset bias issues. Moving beyond fixed training datasets will lead to more robust visual systems, especially when deployed on robots in new environments which must train on the objects they encounter there. To make this possible, it is important to break free from the need for manual annotators. Recent work has begun to investigate how to use the massive amount of images available on the Web in place of manual image annotations. We contribute to this research thread with two findings: (1) a study correlating a given level of noisily labels to the expected drop in accuracy, for two deep architectures, on two different types of noise, that clearly identifies GoogLeNet as a suitable architecture for learning from Web data; (2) a recipe for the creation of Web datasets with minimal noise and maximum visual variability, based on a visual and natural language processing concept expansion strategy. By combining these two results, we obtain a method for learning powerful deep object models automatically from the Web. We confirm the effectiveness of our approach through object categorization experiments using our Web-derived version of ImageNet on a popular robot vision benchmark database, and on a lifelong object discovery task on a mobile robot.



There are no comments yet.


page 2

page 4

page 6

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Joystick - ILSVRC12 (IN) Joystick - WINC Table Lamp - ILSVRC12 (IN) Table Lamp - WINC
Fig. 1: Exemplar images for the classes “Joystick” and “Table Lamp” for ImageNet ILSVRC12 (IN) and its replica created automatically from the Web (WINC). We see the evident differences in style for both classes, brought by time.

Deep networks and ImageNet are the key ingredients of the most recent and successful object categorization approaches in robot vision today, across 2D [1, 2], 2,5D [3, 4] and 3D [5, 6] data. The progress achieved by learning methods in object classification over the last five years has been impressive, and it goes hand-in-hand with the development of deep architectures. In turn, the success obtained so far relies heavily on the availability of very large scale annotated data collections. ImageNet [7], and to a lesser extent Places [8]

, are, at the time of writing, the only two publicly available resources suitable for training deep learning-based object categorization algorithms. The overwhelming majority of such systems use these two collections for pre-training, usually followed by some form of network adaptation to a specific task (for a review of the relevant literature see Section


As with any other annotated database, the perceptual object knowledge in ImageNet is static, i.e. its main body was collected during a limited and specific period (circa 2010111http://image-net.org/about-stats.). This can result in some classes becoming dated over time, especially those representing man-made objects (see Figure 1). Robots deployed in everyday environments, such as offices, homes or hospitals, will encounter a wide variety of man-made objects. Being trained on out of date models will reduce their performance. Moreover, large-scale image datasets are inevitably collected by multiple annotators, which might involuntarily inject some of their own views and bias on object categories, resulting in dataset bias [9]. Fine-tuning, i.e. adapting a pre-trained architecture to a new classification task, requires annotated data, computational resources and computing time. For several robotic settings, these might not be feasible options.

The awareness that we need more diverse data collections to feed data-hungry deep architectures is growing in the community. A promising research thread is the attempt to replace real images with synthetically generated data [10]. An alternative approach attempts to create large-scale databases by downloading images directly from Web search engines [11], without any subsequent cleaning by human annotators. Here the main challenge is to cope with noisy (automatic) annotations [12], while at the same time making sure that the collected images are sufficiently representative of all the visual facets of a given object class.

This paper contributes to this latter approach. We first argue that to take full advantage of noisily labeled data, one needs to study the effect of noise on labels in controlled settings. To this end, we inject different percentages of noise, of various kinds, into the labels of ImageNet, and study its impact on accuracy prediction in two different deep architectures. Our study shows that while the accuracy of the popular AlexNet architecture [13] degrades smoothly as the level of noise increases, GoogLeNet [14] maintains a level of accuracy very close to its top performance for levels of noise up to , then it experiences very severe convergence issues, de facto preventing its use for classification. This finding provides an indirect way to assess the level of noise in a Web-generated database, without any need for manual annotation: if GoogLeNet converges, then the noise level in the automatically generated data collection will be between and , and it will have a minimal impact on the accuracy achievable by the architecture on those data. This result is the first contribution of this paper.

Our second contribution is an algorithm for the automatic creation of large data collections from the Web that minimizes noise and maximizes image variability through visual and natural language processing, performing a concept expansion of the query. We use our approach to re-create automatically ImageNet from the Web and we compare what we get against a dataset collected without query expansion, performing a qualitative evaluation in terms of visible label noise and a quantitative analysis in terms of CNN classification accuracy when using our two deep nets of choice.

We further demonstrate the effectiveness of our approach in two robotics scenarios. We show that CNN networks trained on Web-collected data can be used as feature extractors for images of a popular robot object database [15] and allows to get on it a classification accuracy on par to that obtained when using ImageNet-trained CNN models. Moreover, for a robot deployed in an indoor environment and able to use deep models to overcome gaps into its own visual knowledge base without human intervention, our Web-generated models provide a significant increase in performance compared to using the original ImageNet.

The rest of the paper is organized as follows: after a review of the relevant literature (section II), section III describes the protocol developed to study the effect of noisy labels on deep networks, and reports our findings. Section IV describes our query conceptual expansion strategy, illustrating the impact of each component with an experimental study. Section V showcases the power of our approach on object categorization and robotic scenarios. We conclude identifying open issues and future research directions we plan to pursue.

Ii Related Work

Since the seminal work of Krizhevsky [13]

, deep learning has significantly changed the research landscape in visual object recognition. Over the last few years, the use of convolutional neural networks trained over large-scale databases has become the cornerstone of most robot visual systems. An important question is how to tailor the performance of such networks to the specific task at hand. Deep networks need large scale annotated databases during training. To this end, ImageNet and particularly the related Large Scale Visual Recognition Challenge subset (ILSVRC12,

[7]) are widely used in combination with various deep architectures [1, 10, 6, 4]

. To adapt such 1000-categories classifiers to a new problem, a common strategy is fine-tuning

[16], i.e. the procedure of re-training a part of the network using annotated data from the new classification problem of interest, while keeping the parameters in other parts of the network fixed. The array of fine-tuning strategies is very large [16, 1]

, and although effective, the approach depends strongly on heuristics. Moreover, the assumption of having an annotated dataset of

images from the new task might be unrealistic in several robotics domains, where a system should be able to perceive and act without lengthy training on the site, that would be in any case static as well.

The idea of looking at the Web as a source of knowledge from which to learn is not new in robotics. Previous work looked at obtaining semantic information from the Web in various forms, e.g. from structured and semi-structured sources like WikiPedia, DBPedia and WordNet [17, 18, 19], or combined with deep networks pre-trained on ImageNet [20]. Mining the Web to learn about the appearance of objects or scenes is much less investigated. Ventura et al proposed using the Web to call upon human annotators if a robot detected gaps in its perceptual knowledge [21]. Wohlkinger et al downloaded 3D CAD models from public Web resources, creating a 3D object database acting as a proxy for the Web [22]. Recent work by Young et al [20] proposed the combination of semantic and perceptual Web knowledge, using curated resources like DBPedia and ImageNet as Web proxies. We are not aware of previous work in robot vision for automatically mining the Web for RGB images, in order to learn perceptual object models on a robot without use of manual annotators.

The automatic creation of image databases by mining the Web has been researched in the computer vision community. Focusing explicitly on the work done within the context of deep network training, research efforts have followed two different directions. The first acknowledges the fact that databases created from the Web without manual intervention will inevitably contain some level of noise, and so attempt to design loss layers able to cope with this issue

[23, 12]. The second tackles this issue by proposing query expansion strategies that should reduce the level of noisily annotated images while preserving the visual variability necessary to successfully train deep nets [11, 24, 25]. Our work sits in the middle of these two directions: as in the first, we assume that images downloaded automatically from the Web will have some degree of noise associated with their label. Rather than designing regularization layers to cope with this, we instead study the effect of noisy labels on two different deep architectures and use the insights of this study to our advantage. This, combined with a new recipe for automatic query expansion able to reduce noisily annotated data while spanning the visual concept space significantly, lead to a powerful strategy for robot visual classification in the wild.

Iii Studying the effect of noisy labels on CNNs

When automatically harvesting the Web to build databases for object classification, the labels of the images can be expected to contain noise. However, the level of noise for any given strategy is seldom investigated, beyond the manual generation of ground truth. When building databases of the scale necessary for deep learning, assessing this noise directly is a daunting task.

We therefore assess the level of noise associated with labels created by a given automatic database collection procedure through a reference task. We consider ImageNet, specifically its ILSVRC12 set of categories, as our classification problem. We evaluate how the accuracy of different deep architectures changes when the original dataset is injected with different levels and types of noise. Assessing such changes in classification accuracy gives us information about how well networks trained from scratch on general purpose databases will perform on new, Web-collected data without human annotation.

Fig. 2: Schematic illustration of the procedure used to inject different kind of noise into the original ILSVRC12 data collection. Left: in the external noise case, ImageNet images of categories not involved in the ILSVRC12 set are labelled with the name of the ILSVRC12 categories and substituted to various percentage of the original images. Right: in the internal noise case, some of the images have their label switched to that of another class in the set.

We focus on two main types of label noise:

  • external noise describes an image that is associated with a label, but does not actually belong to any label in the dataset;

  • internal noise describes an image that is associated with a label, but actually belongs to another label in the dataset.

We created replicas of the ILSVRC12 collection with increasing percentages of noise. We created separate replicas for internal and external label noise, ranging from to in increments of . We distributed erroneously labeled images uniformly over the classes, substituting the original correct images with samples pooled randomly either from other ImageNet classes not belonging to ILSVRC12 (external noise) or from different classes of the same set (internal noise). This is illustrated in Figure 2.

We tested two different CNN architectures to evaluate their robustness to this type of noise: AlexNet [26] and GoogLeNet [14]

. The first is the naïve standard CNN in object categorization, while the second is a deeper architecture which includes Inception modules designed to improve the network effectiveness without an uncontrolled blow-up of the computational complexity. In all our experiments we trained the networks on the noisy sets, evaluating the classification performance on the original ILSVRC12 validation set. We used the Caffe

[27] CNN implementation available on NVIDIA Deep Learning GPU Training System (DIGITS) for all our experiments. The learning rate was set to

initially and divided by 10 after every third of the total epochs. We trained both networks with an SGD solver for 30 epochs and we repeated each experiments 5 times reshuffling the image list and using batch size of 128 for AlexNet and 32 for GoogLeNet.

Fig. 3: Top-1 classification accuracy produced by networks trained on ILSVRC12 data containing different percentage of noise and evaluated on the original ILSVRC12 validation set.

The average accuracy results are shown in Figure 3. We see that a low percentage of internal or external noise () induces only a moderate reduction in classification performance. For larger amounts of noise (), the networks appear more robust to external noise than to internal noise. Moreover, GoogLeNet outperforms AlexNet for a low noise percentage, while the trend inverts for an amount of noise equal or over 35%. Although the former result is expected [14], the latter is quite surprising, indicating that GoogLeNet, besides being deeper, is also less robust to noise than AlexNet.

A further investigation conducted on the external noise case indicated that for 35% noise, GoogLeNet converged during training in three out of the five repetitions of the experiment, while for 45% noise it converged in only one of the runs. Since the repetitions only differed in the data order, this tells us that the content of each data batch fed to GoogLeNet is important. When every batch is noisy it becomes difficult to optimize the large number of network parameters. We also tried doubling the batch size with no significant difference in the overall result trend.

To the best of our knowledge, this GoogLeNet behavior has not been observed or discussed before. Besides drawing attention to a limitation of very deep CNN architectures, this result suggests that GoogLeNet might be used as a rough indicator of the level of label noise present in an image database generated automatically from the Web: if for a given collection GoogLeNet converges, this means that (a) the level of noise on the labels is at most , and (b) the recognition accuracy using this training data is expected to be roughly lower than would be achieved by using a noise-free version of the same database. In the rest of the paper we will show how this result, coupled with an effective approach for automatic Web mining, leads to state of the art results in lifelong object discovery [20].

Iv Learning from the Web

This section illustrates our strategy for collecting images from the Web in order to build a training set for learning deep classification models. We present a qualitative analysis on the gathered data and a quantitative evaluation on the accuracy obtained when using them to train deep nets. For a reference task we focus on reproducing the ILSVRC12 database of 1000 categories [28].

Iv-a Dataset Collection

Collecting images of a given category from the Web presents some practical challenges. Many popular search engines have restrictions that limit downloads to a maximum of 1000 images per query. Considering the large number of duplicates for some categories, the number of images obtained per query can be very limited. This limitation means that the statistics of the original ILSVRC12 dataset, composed by several categories each represented by images, are hard to recreate with direct query methods.

We can tackle this issue with two possible solutions:

  • category name & single search engine - use the category name to query a search engine that has weak download restrictions and provides a large number of low quality images;

  • query expansion & multiple search engines - expand the query by adding extra textual information to the category name and use multiple search engines, each with strict download limitations.

The first strategy can use search engines like Picsearch (www.picsearch.com) from which we can get more than 7000 images per class, but the level of control on the output is quite limited (e.g. it is not possible to choose the image type to exclude drawings), and the relevance of the images with respect to the query decreases quickly. The second strategy needs external expert knowledge for each category to provide related text and expand the original query.

One widely adopted method to obtain meaningful expansions is to leverage WordNet [29], either considering the synonyms in the synset of each category, the parent nodes, or even the whole category textual description. However, some of the WordNet synsets contain a single word (e.g. “jay”, “boxer”, “crane”) so no synonyms are available. When searching for parents, deciding at which depth level of the ontology to stop is not trivial (e.g. “jay”, “corvin”, “oscine”, “passerine”, “bird”, “vertebrate”, etc.): the hierarchy of classes is not uniform, with some parts more or less dense than others due to different numbers of siblings. Hence, it is hard to find a meaningful rule for all categories. Using the textual description of a category requires a natural language pre-processing step able to isolate relevant words and avoid the confusing ones (e.g. for ”crane”: “large long-necked wading bird of marshes and plains in many parts of the world” the query crane+bird can be useful, but crane+world provides misleading results). Finally, to get a sufficient amount of images, one or two words are not enough, as preliminary tests show that more than ten are needed to reach our goal of reproducing the ILSVRC12 class statistics.

In the following we will use IN to indicate the original ImageNet ILSVRC12 dataset and WIN to indicate our Web-collected ImageNet ILSVRC12 version, obtained adopting following the category name & single search engine strategy. We also create WINC using query expansion & multiple search engines strategy. Below we describe our query expansion and preprocessing procedures.

Query Expansion. Since there isn’t a single universal strategy for query expansion, we target simplicity. We assume that our database creation process can only access the Web to gain knowledge and cannot ask for a human for disambiguation or expert help. Under such conditions, Search Engine Optimization (SEO) practice can provide guidance. SEO evaluates the relation between how users query a search engine and the preference they express in the returned results. A highly preferred result, usually top ranked by the search engine, is indexed with keywords considered as very relevant to the original query. By exploiting SEO tools, in particular the ones that target image search, we can easily get a large number of relevant keywords associated to visual concepts. Here we use the MyKeyworder service (www.mykeyworder.com), giving a category name as input and collecting the top 20 output keywords. These can be considered intrinsically visual, because they are associated with image search queries, semantic concepts and photographic tags. Each of the obtained keywords are then used as query expansions for an object class name when querying three image search engines: Google, Yahoo! and Bing. Although the maximum number of available images from each of them is 1000, we limited ourself to the top 200 results as in [30]. This choice reduces the risk of introducing irrelevant and noisy images (which usually appears after the first result page), and guarantees a large variability in the final image set since multiple keywords cover different visual aspects of the same concept. Considering the whole set of 1000 ILSVRC12 classes, the download phase leaves us with images and a large space for further refinement.

Duplicates Removal. An essential preprocessing step consists of removing duplicate images. When using multiple search engine as well as multiple concept expansions, we deal with two kinds of image repetitions. The per-query duplicates are exact copies or very similar images obtained by using the same query across three different search engines. These duplicates are eliminated immediately after the end of the download process. The cross-expansion duplicates are exact copies or very similar images obtained when expanding a concept with different keywords. These duplicates are removed in a second processing step; the number of images shared between each expansion and the original concept, not expanded, is saved as a measure of diversity. Both the duplicate removal processes are executed by extracting the image fingerprints through perceptual hashing and then comparing them to find close image pairs.

Variability Maximization. To induce better generalization in the classifier that will be learned on this data we aim to maximize the quality and variability of each class. Therefore we select only high quality expansions to include in the dataset. Given the -th expansion of the -th class , we can reasonably suppose that it is of high quality if,

  • the images of the expansion lie close to the class images , i.e. they have a small L2 distance in the feature space;

  • the images of the expansions

    are not sparse, i.e. there is small standard deviation

    in the image set;

  • the number of duplicates shared by the pair is small, which ensures variability.

We define an inverted score for each expansion (minimal score corresponds to maximal quality, see Figure 4 for the formula and a graphical illustration), sort them, pick the top 5 expansions out of the 20 per class and merge the corresponding images to define the pool from which we randomly select samples with the same statistics of the ILSVRC12 collection.

Fig. 4: Schematic illustration of the relation between a concept and its expansions in the feature space. By using the distribution of the samples we define a score for each expansions which helps to select them by maximizing the variability coverage for each class. In our experiments we used off-the-shelf CNN fc7 activations of an AlexNet pre-trained on ImageNet ILSVRC12 as features [31].

Table I shows the effect of the query expansion and of the variability maximization procedure, indicating the extra keywords obtained with the former and the selection operated by the latter for four categories.

class name keywords for concept expansion selected concepts
siberian husky dog, siberian, animal, canine, cute, portrait, pet, breed, mammal, purebred, pet,isolated,beautiful,purebred,dog
domestic, young, white, fur, beautiful, nature, wolf, isolated, adorable, looking
grey whale mammal, ocean, sea, wildlife, marine, mexico, gray, animal, life, water, mexico,ocean,swimming,lagoon,water
lagoon, cetacean, fin, baja, blue, swimming, tail, watching, america, powerful
desk office, top, table, work, business, background, white, view, blank, empty, office, space, design, above, phone
above, space, computer, paper, workplace, wooden, coffee, phone, notebook, design
vending machine

business, buy, dispenser, vector, automatic, drink, food, service, coin, isolated,

beverage, food, automatic, can, drink
illustration, consumer, can, beverage, snack, choice, symbol, sale, merchandise, button
TABLE I: For the four categories listed in the first column, this table shows the top 20 keywords produced as output by the SEO tool (second column) and the 5 keywords selected by the variability maximization procedure (third column).

Iv-B Qualitative Analysis

Icecream - ILSVRC12 (IN) Icecream - WIN Icecream - WINC ILSVR red, WINC blue
Pillow - ILSVRC12 (IN) Pillow - WIN Pillow - WINC ILSVR red, WINC blue
Fig. 5: Examples of images extracted from ImageNet ILSVRC12 (IN), WIN and WINC dataset for two categories. The last column on the right shows dot plots of image samples for IN and WINC in a bi-dimensional subspace obtained applying PCA on the fc7 activations of AlexNet trained on ImageNet ILSVRC12.

To show the difference between the reference IN dataset and WIN/WINC we show samples of two object categories in Figure 5. For both “Icecream” and “Pillow” we see that (1) WIN images often do not focus on the single object and consider instead a whole scene; moreover, some of the images are irrelevant. (2) WINC images appear much more consistent with the corresponding assigned label; (3) by projecting the feature maps (fc7 activations of AlexNet trained on ImageNet) of the images in the same class of IN and WINC into a bi-dimensional space through PCA we can look at their overall distribution. The plots show that WINC, besides spanning the same part of the visual space of IN, presents also a larger variability which is not only due to noise, but also to several relevant images (Icecream cones and Pillows of different shapes on white background). Similar conclusions can be drawn when looking at other object categories.

Iv-C Quantitative Analysis

val:IN val:WINC
trainn. val. classes 10 50 100 200 1000 10 50 100 200
AlexNet IN 57.8 59.5 61.7 58.5 49.5 58.8 61.1 57.9 56.4
AlexNet WIN 38.8 31.9 33.2 32.5 31.0 47.2 38.8 38.7 39.1
AlexNet WINC 42.2 39.7 42.3 37.9 33.6 62.1 61.0 59.7 59.2
GoogLeNet IN 64.8 65.8 66.0 65.3 62.2 63.4 65.0 62.1 61.4
GoogLeNet WINC 52.2 51.2 50.5 49.6 42.7 69.2 70.4 65.1 65.0
TABLE II: Top-1 accuracy performance (): classification results of two CNN architectures trained on the curated IN dataset and on its web-collected replicas WIN and WINC.

We complete the evaluation of the Web-collected datasets by training CNNs on them and comparing their performance with that obtained on the curated ImageNet ILSVRC12 (IN). Specifically, we train AlexNet and GoogLeNet, testing both of them on the original ILSVRC12 validation set (val:IN) as well as on a cleaned Web-based validation set (val:WINC). This was collected by selecting 50 images out of the samples of WINC that were not involved in the respective training set and manually verifying their labels. Although this manual image annotation is time expensive, having this extra testbed allows us to assess the effect of a possible bias between ImageNet and the Web-collected datasets as discussed in the introduction. To keep the time dedicated to the annotation limited, we considered a subset of 200 out of 1000 classes. For completeness we also ran tests on groups of 10, 50 and 100 classes, showing that the overall performance of the different models remain consistent regardless of the cardinality of the labeled set.

The top-1 classification accuracy results obtained with AlexNet are shown in the first rows of Table II. They indicate that the models trained on IN outperform those trained on WIN/WINC when both are tested on val:IN. Between WINC and WIN the first shows higher accuracy than the second, as we could expect from the qualitative analysis. The trend between IN and WINC changes when testing on val:WINC. Here training on WINC provides equal or better performance than training on IN

. These results prove empirically the presence of a domain shift between the ImageNet and Web underlying probability distributions

[32]. They also allow us to claim the reliability of the WINC model. The behavior of the WIN model remains instead unchanged regardless of the validation set, indicating that this version of the Web-collected dataset is significantly worse than the original IN. Because of this, we will not consider it in the following analysis, concentrating only on our clean WINC.

When using GoogLeNet (bottom part of Table II) all results grow in absolute terms, but the relative performance of the IN and WINC models remains the same as with AlexNet, confirming what already stated before. A further important message provided by the GoogleNet results is that it is possible to train this network on WINC without encoutering network convergence issues. From our results in Section III we can conclude that the amount of noise present in our Web-collected version of ImageNet is below the critical threshold of .

V Experiments

Having shown that it is possible to learn reasonable CNN classification models from Web-collected images, we now assess their performance when facing real scenarios from robotic applications. We consider two tasks: object classification using the models for feature extraction, and lifelong object discovery using the models to produce category proposals.

Fig. 6: First row: examples of images from the Washington dataset [15]. Second row: examples of images of Dataset A from [20].

Feature Extraction. We conducted experiments on the Washington RGB-D object dataset [15]. This is a widely used testbed in the robotic community, consisting of 41,877 RGB-D images organized into 300 instances divided in 51 classes of common indoor objects (e.g. scissors, cereal box, keyboard etc. see top row of Figure 6). Each object instance was positioned on a turntable and captured from three different viewpoints while rotating. Since two consecutive views are extremely similar, only 1 frame out of 5 is used for evaluation purposes. We did not consider the depth information and focused only on the RGB images: following [10], we represented the images with the last fully convolutional layer of the AlexNet network trained on the IN dataset and learned a linear SVM model on them. Note that the results obtained with this protocol can be considered the off-the-shelf state of the art in the robot vision community. Finally we ran a test on the splits provided with the dataset and compared the results with those obtained by using AlexNet trained on WINC. We also repeated the evaluation when using a pre-trained GoogLeNet for both IN and WINC, considering their last inception layer as features222Multiple validation experiments have shown that, when using GoogLeNet, the inception_5b

layer produces better results than the following pooling and softmax layers.

. Results shown in Table III indicate that, regardless of the considered network, the average classification accuracy obtained with WINC models is just slightly worse than that obtained with IN models and they can be considered equal within their standard deviations. It is worthwhile to remember that the Washington database was collected in 2011, so with respect to age it can be considered closer to the IN dataset than to WINC. Nevertheless the WINC model is robust enough to provide good results when used for feature extraction.

model classification accuracy (%)
IN AlexNet
WINC AlexNet
IN GoogLeNet
WINC GoogLeNet
TABLE III: SVM classification accuracy on the Washington database [15]. The first column indicates the training data and network architecture used to extract the image features.
Fig. 7: WUP values between ground truth and predicted object labels. The title of each figure indicates the CNN architecture adopted to learn the visual classification models which generate the category proposals, while the legend specifies their training data.

Category Proposals. For a mobile service robot in human environment it is important not only to recognize known object categories, but also to be able to generate hypotheses about previously unseen categories. CNN models pre-trained on thousand of categories can be used as a potential source of information to extend a robot’s situated knowledge. For this task we considered the same setting described in [20]: a long-term autonomous mobile service robot equipped with an RGB-D camera recording everyday scenes in a human environment [33]. It observes cabinet tops, counters, desks, shelves and other surfaces taking multiple views from various angles. From these surfaces it detects and segments objects which it isn’t able to recognise [34]. During the test phase, a cropped image of the unknown object is given as input to a pre-trained CNN model. Although the correct object label might not be in the predicted set, the success is measured by evaluating the WUP similarity [35] between the ground truth and the proposed category with values ranging in .

As we deal with autonomously gathered images in real environments they appear smaller, with lower-resolution and from less favourable angles than both the images in ImageNet and in our Web-collected dataset (see bottom row of Figure 6). Nevertheless, note that using a deep network trained on a database created automatically from the Web is a strong proxi for autonomous lifelong object discovery from Web images.

We run the experiments on Dataset A from [20], starting from AlexNet models trained both on IN and WINC. Figure 7 shows the results per each of 20 object categories of the dataset and indicate that in most of the cases the WUP score between the ground truth and predicted labels is higher when using WINC than when using IN. On average we get a WUP score of 0.52 with IN and 0.58 with WINC. The observed performance is also confirmed by the results obtained when using the GoogLeNet architecture with average WUP score of respectively 0.62 and 0.64. We see these results as a first, important step towards robots able to learn the visual appearance of previously unseen objects autonomously from the Web.

Vi Conclusions

The aim of this work is to pave the way for robots able to learn perceptual object models from images found on the Web, without the need of human annotators. We took the ImageNet ILSVRC12 1000 categories database as a reference task, and we studied how injecting noise into it affected the performance of two popular deep architectures, AlexNet and GoogLeNet. We built on our findings, presenting a query expansion approach for automatically re-creating ILSVRC12 from Web downloads, and we demonstrated experimentally that this Web-derived version of ImageNet leads to deep models that are better at capturing the visual facets of object categories when used to extend a robot on-board perceptual object knowledge.

We see this work as the first step on a long and intriguing road. The possibility to automatically create data collections from the Web, large enough to train a deep net, should be further explored, aiming for much larger data collections in terms of classes and images per class. This would allow us to study questions such as when would we reach the upper bound of accuracy for a given architecture? Another avenue we intend to explore is the possibility of creating task specific, rather than all-purpose, databases, which intuitively should allow a robot to exploit any prior knowledge it might have on its situated settings and goals. Finally, we plan to continue our investigations on the effect of noise on the behavior of deep architectures, both in terms of further experiments and in-depth analysis trying to pinpoint the specific layers that are more affected by it.


  • [1]

    M. Schwarz, H. Schulz, and S. Behnke, “Rgb-d object recognition and pose estimation based on pre-trained convolutional neural network features,” in

    Proc ICRA, 2015.
  • [2] C. Li, A. Reiter, and G. D. Hager, “Beyond spatial pooling: Fine-grained representation learning in multiple domains,” in Proc CVPR, 2015.
  • [3] S. Gupta, R. Girshick, P. Arbeláez, and J. Malik, “Learning rich features from rgb-d images for object detection and segmentation,” in Proc ECCV, 2014.
  • [4] A. Eitel, J. T. Springenberg, L. Spinello, M. Riedmiller, and W. Burgard, “Multimodal deep learning for robust rgb-d object recognition,” in Proc. IROS.
  • [5] R. Socher, B. Huval, B. P. Bath, C. D. Manning, and A. Y. Ng, “Convolutional-recursive deep learning for 3d object classification.” in Proc NIPS, 2012.
  • [6] H. F. Zaki, F. Shafait, and A. Mian, “Convolutional hypercube pyramid for accurate rgb-d object category and instance recognition,” in Proc ICRA, 2016.
  • [7] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “Imagenet: A large-scale hierarchical image database,” in Proc CVPR, 2009.
  • [8]

    B. Zhou, A. Lapedriza, J. Xiao, A. Torralba, and A. Oliva, “Learning deep features for scene recognition using places database.” in

    Proc NIPS, 2014.
  • [9] A. Torralba and A. A. Efros, “Unbiased look at dataset bias,” in Proc CVPR, 2011.
  • [10] F. M. Carlucci, P. Russo, and B. Caputo, “A deep representation for depth images from synthetic data,” in Proc ICRA, 2017.
  • [11] D. S. Cheng, F. Setti, N. Zeni, R. Ferrario, and M. Cristani, “Semantically-driven automatic creation of training sets for object recognition,” Computer Vision and Image Understanding, vol. 131, pp. 56–71, 2015.
  • [12] S. Sukhbaatar and R. Fergus, “Learning from noisy labels with deep neural networks,” CoRR, vol. abs/1406.2080, 2014.
  • [13] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Proc NIPS, 2012.
  • [14] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going deeper with convolutions,” in Proc. CVPR, 2015.
  • [15] K. Lai, L. Bo, X. Ren, and D. Fox, “A large-scale hierarchical multi-view rgb-d object dataset,” in Proc ICRA, 2011.
  • [16] K. Chatfield, K. Simonyan, A. Vedaldi, and A. Zisserman, “Return of the devil in the details: Delving deep into convolutional nets,” in Proc BMVC, 2014.
  • [17] M. Beetz, D. Beßler, J. O. Winkler, J. Worch, F. Balint-Benczedi, G. Bartels, A. Billard, A. K. Bozcuoglu, Z. Fang, N. Figueroa, A. Haidu, H. Langer, A. Maldonado, A. L. P. Ureche, M. Tenorth, and T. Wiedemeyer, “Open robotics research using web-based knowledge services,” in Proc ICRA, 2016.
  • [18] M. Samadi, T. Kollar, and M. Veloso, “Using the Web to Interactively Learn to Find Objects,” in Proc AAAI, 2012.
  • [19] J. Young, V. Basile, L. Kunze, E. Cabrio, and N. Hawes, “Towards lifelong object learning by integrating situated robot perception and semantic web mining,” in

    23rd European Conf. on Artificial Intelligence (ECAI’16)

    , The Hague, Netherlands, 2016.
  • [20] J. Young, V. Basile, E. Cabrio, N. Hawes, and B. Caputo, “Semantic web-mining and deep vison for lifelong object discovery,” in Proc ICRA, 2017.
  • [21] R. Ventura, B. Coltin, and M. Veloso, “Web-Based Remote Assistance to Overcome Robot Perceptual Limitations,” in Proc AAAI, 2013.
  • [22] W. Wohlkinger, A. Aldoma, R. B. Rusu, and M. Vincze, “3dnet: Large-scale object class recognition from CAD models,” in Proc ICRA, 2012.
  • [23]

    X. Chen and A. Gupta, “Webly supervised learning of convolutional networks,” in

    Proc ICCV, 2015.
  • [24] S. K. Divvala, A. Farhadi, and C. Guestrin, “Learning everything about anything: Webly-supervised visual concept learning,” in Proc CVPR, 2014.
  • [25] E. Golge and P. Duygulu, “Conceptmap: Mining noisy web data for concept learning,” in Proc ECCV, 2014.
  • [26] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Proc. NIPS, 2012.
  • [27] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell, “Caffe: Convolutional architecture for fast feature embedding,” in Proc ACM Multimedia, 2014.
  • [28] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, “ImageNet Large Scale Visual Recognition Challenge,” International Journal of Computer Vision (IJCV), vol. 115, no. 3, pp. 211–252, 2015.
  • [29] C. Fellbaum, Ed., WordNet: an electronic lexical database.   MIT Press, 1998.
  • [30] K. Chatfield and A. Zisserman, “Visor: Towards on-the-fly large-scale object category retrieval,” in Proc ACCV, 2012.
  • [31] A. S. Razavian, H. Azizpour, J. Sullivan, and S. Carlsson, “Cnn features off-the-shelf: An astounding baseline for recognition,” in CVPR Workshop, ser. CVPRW ’14, 2014.
  • [32] K. Saenko, B. Kulis, M. Fritz, and T. Darrell, “Adapting visual category models to new domains,” in European Conference on Computer Vision (ECCV), 2010.
  • [33] N. Hawes, C. Burbridge, F. Jovan, L. Kunze, B. Lacerda, L. Mudrová, J. Young, J. L. Wyatt, D. Hebesberger, T. Körtner, R. Ambrus, N. Bore, J. Folkesson, P. Jensfelt, L. Beyer, A. Hermans, B. Leibe, A. Aldoma, T. Faulhammer, M. Zillich, M. Vincze, M. Al-Omari, E. Chinellato, P. Duckworth, Y. Gatsoulis, D. C. Hogg, A. G. Cohn, C. Dondrup, J. P. Fentanes, T. Krajník, J. M. Santos, T. Duckett, and M. Hanheide, “The STRANDS project: Long-term autonomy in everyday environments,” IEEE Robotics & Automation Magazine (RAM), In Press.
  • [34] T. Faeulhammer, R. Ambrus, C. Burbridge, M. Zillich, J. Folkesson, N. Hawes, P. Jensfelt, and M. Vincze, “Autonomous learning of object models on a mobile robot,” IEEE Robotics and Automation Letters (RA-L), vol. 2, no. 1, pp. 26 – 33, 2016.
  • [35] Z. Wu and M. Palmer, “Verbs semantics and lexical selection,” in Annual Meeting on Association for Computational Linguistics, 1994.