GET-AID: Visual Recognition of Human Rights Abuses via Global Emotional Traits

02/11/2019 ∙ by Grigorios Kalliatakis, et al. ∙ University of Essex 6

In the era of social media and big data, the use of visual evidence to document conflict and human rights abuse has become an important element for human rights organizations and advocates. In this paper, we address the task of detecting two types of human rights abuses in challenging, everyday photos: (1) child labour, and (2) displaced populations. We propose a novel model that is driven by a human-centric approach. Our hypothesis is that the emotional state of a person -- how positive or pleasant an emotion is, and the control level of the situation by the person -- are powerful cues for perceiving potential human rights violations. To exploit these cues, our model learns to predict global emotional traits over a given image based on the joint analysis of every detected person and the whole scene. By integrating these predictions with a data-driven convolutional neural network (CNN) classifier, our system efficiently infers potential human rights abuses in a clean, end-to-end system we call GET-AID (from Global Emotional Traits for Abuse IDentification). Extensive experiments are performed to verify our method on the recently introduced subset of Human Rights Archive (HRA) dataset (2 violation categories with the same number of positive and negative samples), where we show quantitatively compelling results. Compared with previous works and the sole use of a CNN classifier, this paper improves the coverage up to 23.73 child labour and 57.21 trained models are available online at



There are no comments yet.


page 1

page 2

page 3

page 4

page 6

page 7

page 9

page 10

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

In the era of mobile phones with photo and video capabilities, social media and the Internet, citizen media and other publicly available footage can provide documentation of human rights violations and war crimes. However, the omnipresence of visual evidence may deluge those accountable for analysing it. Currently, information extraction from human-rights-related video and imagery requires manual labour by human rights analysts and advocates. Such analysis is utterly expensive and time consuming, while it is emotionally traumatic to focus solely on images of horrific events without context. This paper attempts to address the problem of predicting human rights abuses from a single image, and provides technical insight and solutions to address the above issues. Note that naive schemes based on object detection or scene recognition are doomed to fail in this binary classification problem as illustrated in Fig.


. If we can exploit existing smartphone cameras, which are ubiquitous, it may be possible to turn human rights image analysis into a powerful and cost-effective computer vision application.

Recent work by Kalliatakis et al. [15] has shown that fine-tuned deep CNNs can address the human rights abuses classification problem to a certain extent. In this paper, we investigate the potency of representation learning methods in solving two independent, binary classification problems in the context of human rights image analysis:

  1. child labour / no child labour

  2. displaced populations / no displaced populations

As will be shown in the experimental section, we found that the integration of novel global emotional traits with a fine-tuned CNN reports an improvement in coverage up to 23.73% for child labour and 57.21% for displaced populations, over the sole use of a CNN classifier that is trained end-to-end using the given training data. Based on these new global emotional traits, our approach makes the following technical contributions:

(a) Children playing. (b) Child labour.
(c) Camping. (d) Displaced populations.
Fig. 1: In many instances, emotional traits can be the only notifying difference between an image that depicts human rights abuse and a non-violent situation.

First, we describe the design and implementation of an end-to-end system, which is used to estimate the emotional states of people, adopting the continuous dimensions of the

VAD Emotional State Model [23]. This model describes emotions using three numerical dimensions: Valence (V) to measure how positive or pleasant an emotion is; Arousal (A) to measure the agitation level of the person; and Dominance (D) that measures the control level of the situation by the person.

Our second technical contribution comprises a new mechanism capable of characterising an image based on the emotional states of all people in the scene, termed global emotional traits (GET). This mechanism exploits two of the continuous dimensions of the VAD emotional state model which are relevant to human rights image analysis. As will be explained in the following, global emotional traits are learned by jointly analysing each person and the entire scene similar to [17].

We present a human-centric, end-to-end model for recognising two types of human rights violations. Our central observation is that the proposed global emotional traits of an image, which expose the entire mood of a situation are highly informative for inferring potential human rights abuses, when used along with a standard image classification system. Finally, we describe HRA—Binary, a subset of the HRA dataset [15] as benchmark for this classification task. This is used to evaluate our GET-AID model.

The remainder of the paper is organized as follows. In Section II we review the related work on object detection, emotion recognition and human rights abuses recognition. In Section III, we introduce our method, GET-AID. In Section IV, we describe the subset of the HRA dataset. Section V discusses our results on human rights abuse classification. We conclude this paper in Section VI.

Ii Related work

This section gives an overview of the related work on human rights abuses recognition, which can be regarded as a category in image understanding. The background of object detection and emotion recognition are also investigated since they are important components in our human-centric approach.

Ii-a Human rights abuses understanding

Human rights abuses understanding plays a crucial role in human rights advocacy and accountability efforts. Automatic perception of potential human rights violations enables to discover content, that may otherwise be concealed by sheer volume of visual data. These automated systems are not producing evidence, but are instead narrowing down the volume of material that must be examined by human analysts who are making legitimate claims, that they then present in justice, accountability, or advocacy settings [2].

One recent example of the potential use of AI to advance human rights analysis can be found in the work of the Center for Human Rights Science (CHRS) at Carnegie Mellon University [25], where they have created computer vision methods to rapidly process and analyse large amount of video by detecting audio (explosions and gunshots), detecting and counting people and synchronising multiple videos from different sources. Another example is the Event Labeling through Analytic Media Processing (E-LAMP) computer vision-based video analysis system [3] which is capable of detecting objects, sounds, speech, text, and event types in a video collection.

Though the above methods have shown good performance in their respective applications, a significant concern associated with human rights-related technology, is that the ability to extract large video collections tends to be limited to institutions with large staff base or access to expensive, technologically advanced tools and techniques [16]. A different group of methods based on a single given image alongside the first ever image dataset for the purpose of human rights violations recognition was introduced in [14]. Recently, Kalliatakis et al. [15] proposed to recognise human rights violations from a single image by fine-tuning object-centric and scene-centric CNNs on a larger, verified-by-experts image dataset. That work also introduced a web-demo for recognising human rights violations, accessible through computer or mobile device browsers.

Fig. 2: Model Architecture. Our model consists of (a) an object detection branch, (b) an emotional traits branch, and (c) an abuse-centric branch. The image features and their layers are shared between the emotional traits and abuse branches (blue boxes).

Ii-B Object detection

Object detection, the process of determining the instance of the class to which an object belongs and estimating the location of the object by outputting the bounding box around the object, is one of the computer vision areas that has improved steadily in the past few years. It is possible to group object detectors using different aspects, but one of the most accepted categorisation is to split object detectors into two categories: one-stage detectors and two-stage detectors.

One-stage detectors: The OverFeat model [29] which applies a sliding window approach based on multi-scaling for jointly performing classification, detection and localization was one of the first modern one-stage object detectors based on deep networks. More recently YOLO[22, 10] and SSD [26, 27] have revived interest in one-stage methods, mainly because of their real time capabilities, although their accuracy trails that of two-stage methods. One of the main reasons being due to the class imbalance problem [20].

Two-stage detectors: the leading model in modern object detection is based on a two-stage approach which was established in [31]. The first stage generates a sparse set of candidate proposals that should contain all objects, and the second stage classifies the proposals into foreground classes or background. R-CNN, a notably successful family of methods [11, 12] enhanced the second-stage classifier to a convolutional network, resulting in large accuracy improvements. The speed of R-CNN has improved over the years by integrating region proposal networks (RPN) with the second-stage classifier into a single convolution network, known as the Faster R-CNN framework [28].

In this work, we adopt the one-stage approach of RetinaNet framework [20] which handles class imbalance by reshaping the standard cross entropy loss to focus training on a sparse set of hard examples and down-weights the loss assigned to well-classified examples.

(a) (b)
Fig. 3: Examples of people marked with the red bounding box that have been labelled with different scores of Valence, Arousal and Dominance.

Ii-C Emotion recognition

Recognising people’s emotional states from images is an active research area among the computer vision community. One established categorisation of emotion recognition methods in the literature is to split them into two categories based on how emotions are represented: discrete categories and continuous dimensions.

Discrete categories: most of the research in computer vision to recognise people’s emotional states is explored by facial expression analysis [8, 7], where a large variety of methods have been developed to recognise the 6 basic emotions defined in [6]. Many of these methods are based on a set of specific localised movements of the face, called Action Units, in order to encode the facial expressions [9, 5]. More recently emotion recognition systems based on facial expressions use CNNs to recognise the Action Units [8].

Continuous dimensions: instead of recognising discrete emotion categories, this family of methods use the continuous dimensions of the VAD Emotional State Model [23, 24] to represent emotions. The VAD model uses a 3-dimensional approach to describe and measure the emotional experience of humans: Valence (V) describes affective states from highly negative (unpleasant) to highly positive (pleasant); Arousal (A) measures the intensity of affective states ranging from calm to excited or alert; and Dominance (D) represents the feeling of being controlled or influenced by external stimuli. In recent times the VAD model has been utilised for facial expression recognition [30].

In this paper, we adopted the tridimensional model of affective experience alongside a joint analysis of the person and the entire scene in order to recognise rich information about emotional states, similar to [17].

Iii Method

We now describe our method for detecting two types of human rights abuses based on the global emotional traits of people within the image. Our goal is to label challenging everyday photos as either human-rights-abuse positive (‘child labour’ or ‘displaced populations’) or human-rights-abuse negative (‘no child labour’ or ‘no displaced populations’ respectively).

To detect the global emotional traits of an image, we need to accurately localise the box containing a and the associated object of interaction (denoted by and , respectively), as well as identify the emotional states of each human using the VAD model. Our proposed solution adopts the RetinaNet [20] object detection framework alongside an additional emotional traits branch that estimates the continuous dimensions of each detected person and then determines the global emotional traits of the given image.

Specifically, given a set of candidate boxes, RetinaNet outputs a set of object boxes and a class label for each box. While the object detector can predict multiple class labels, our model is concerned only with the ‘person’ class. The region of the image comprising the person whose feelings are to be estimated at is used alongside the entire image for simultaneously extracting their most relevant features. These features, are fused and used to perform continuous emotion recognition in VAD space. Our model extends typical image classification by assigning a triplet score to pairs of candidate human boxes and an abuse category a. To do so, we decompose the triplet score into three terms:


We discuss each component next, followed by details for training and inference. Fig. 2 illustrates each component in our full framework.

Iii-a Model components

Iii-A1 Object detection

The object detection branch of our network, shown in Fig. 2(a), is identical to that of RetinaNet [20] single stage classifier, although the SSD [22] detector has also been tested. First, an image is forwarded through ResNet-50 [13]

, then in the subsequent pyramid layers, the more semantically important features are extracted and concatenated with the original features for improved bounding box regression. Although the combination of the proposal and classification components into a single network results in a more complicated training process and lower accuracy compared to their two-stage counterparts, RetinaNet manages to counterbalance that with higher detection speed by utilising a smarter loss function named focal loss.

(a) (b)
(c) (d)
Fig. 4: Estimating continuous emotions in VAD space v global emotional traits from the combined body and image features. These global emotional traits will be integrated with the standard image classification scores to detect human rights abuses. The left column shows the predicted emotional states and their scores from the person region of interest (RoI), while the right column show the same images analysed for global emotional traits.

Iii-A2 Emotional traits recognition

The first role of the emotional traits branch is to assign an emotion classification score to each human box and emotion . Similar to [17], we use an end-to-end model with three main modules: two feature extractors and a fusion module. The first module takes the region of the image comprising the person whose emotional traits are to be estimated, , while the second module takes as input the entire image and extracts global features. This way the required contextual support is accommodated in our emotion recognition process. Finally, the third module takes as input the extracted image and body features and estimates the continuous dimensions in VAD space. Fig. 3 shows qualitative results of the method.

The second role of the emotional traits branch is to assign a two-dimensional emotion classification score which characterises the entire input image, based on two emotion classification scores and for each human. To the best of our knowledge, this particular score, called global emotional traits (GET), is the first attempt to establish a method for summarising the overall mood of a situation depicted in a single image. We decompose the two-dimensional GET score into two terms:


In the above, is the encoding of the global valence score relative to human box and entire image , that is:


Similarly, is the encoding of the global dominance score relative to human box and entire image , that is:


In Fig. 4 (a),(c) the three different emotional states over target objects location are estimated, while (b), (d) illustrate the global emotional traits proposed here. For the sake of completeness all three predicted numerical dimensions are depicted. However, only valence and dominance are considered to be relevant to human rights violations recognition tasks since the agitation level of a person, denoted by arousal, can be ambiguous for several situations. For example, both Fig. 4 (a) and (c) depict people with arousal values close to 5.5, but the activities captured are utterly different from a human rights image analysis perspective.

Iii-A3 Abuse recognition

The first role of the abuse branch, shown in Fig. 2

(c), is to assign an abuse classification score to the input image. Just like in the two-phase transfer learning scheme deployed previously for HRA-CNNs

[15], we train an end-to-end model for binary classification of everyday photos as either human-rights-abuse positive (‘child labour’ or ‘displaced populations’) or human-rights-abuse negative (‘no child labour’ or ‘no displaced populations’).

In order to improve the discriminative power of our model, the second role of the abuse branch is to integrate in the recognition pipeline. Specifically, the raw image classification score is readjusted based on the recognised global emotional traits. Each GET unit, that is deltas from the neutral state, is expressed as a numeric weight varying between 1 and 10, while the neutral states of GET are assigned between 4.5 and 5.5 based on the number of examples per each of the scores in the continuous dimensions reported in [17]

. The GET feature of each input image can be written in the form of a 2-element vector:


where and

refer to the weights of valence and dominance, respectively. The adjustment that will be assigned to the raw probability,

is the weight of valence or dominance multiplied by a factor of 0.11 which has been experimentally set. When the input image depicts positive valence or positive dominance, the adjustment factor is subtracted from the positive human-rights-abuse probability and added to the negative human-rights-abuse probability. Similarly, when the input image depicts negative valence or negative dominance the adjustment factor is added to the negative human-rights-abuse probability and subtracted from the positive human-rights-abuse probability. This is formally written in Algorithm 1.

Finally, when no were detected from the object detection branch, (1) is reduced into plain image classification as follows:


Iii-B Training

We approach human rights abuses classification as a cascaded, multi-task learning problem. Due to different datasets, convergence times and loss imbalance, all three branches have been trained separately. For object detection we adopted an existing implementation of the RetinaNet object detector, pre-trained on the COCO dataset [21], with a ResNet-50 backbone.

For emotion recognition in continuous dimensions, we formulate this task as a regression problem using the Euclidean loss. The two feature extraction modules described in Section


, are designed as truncated versions of various well-known convolutional neural networks and initialised using pretrained models on two large-scale image classification datasets, ImageNet

[18] and Places [32]. The truncated version of those CNNs removes the fully connected layer and outputs features from the last convolutional layer in order to maintain the localisation of different parts of the images which is significant for the task at hand. Features extracted from these two modules (red and blue boxes in Fig. 2

(b)) are then combined by a fusion module. This module first uses a global average pooling layer to reduce the number of features from each network and then a fully connected layer, with an output of a 256-D vector, functions as a dimensionality reduction layer for the concatenated pooled features. Finally, we include a second fully connected layer with 3 neurons representing valence, arousal and dominance. The parameters of the three modules are learned jointly using stochastic gradient descent with momentum of 0.9. The batch size is set to 54 and we use dropout with a ratio of 0.5.

For human rights abuse classification, we formulate this task as a binary classification problem. We train an end-to-end model for classifying everyday images as human-rights-abuse positive or human-rights-abuse negative, based on the context of the images, for two independent use cases, namely child labour and displaced populations. Following the two-phase transfer learning scheme proposed in [15], we fine-tune various CNN models for the two-class classification task. First, we conduct feature extraction utilising only the convolutional base of the original networks in order to end up with more generic representations as well as retaining spatial information similar to emotion recognition pipeline. The second phase consists of unfreezing some of the top layers of the convolutional base and jointly training a newly added fully connected layer and these top layers.

All the CNNs presented here were trained using the Keras Python deep learning framework


over TensorFlow

[1] on Nvidia GPU P100.

Iii-C Inference

Object Detection Branch: We first detect all objects (including the person class) in the input image. We apply a threshold on boxes with scores higher than 0.5, which is set conservatively to retain most objects. This yields a set of boxes with scores and . These boxes are used as input to the emotional trait branch.

Emotional Traits Branch: Next, we apply the emotional traits branch to all detected objects that were classified as human. We feed each human box alongside the entire input image to the VAD emotion recognition model. For each , we predict valence and arousal scores, and then compute the global emotional traits that describes the entire input image .

Abuse Branch: If no human box has been detected, for example when a plain beach without people or kitchen appliances was given as input image, the branch predicts the two abuse scores and directly from the binary classifier. On the other hand, when one or more people have been detected, the branch weights the raw predictions from the binary classifier based on the computed global emotional traits of the input image according to Algorithm 1.

   {: child labour/disp. populations}
   {: no child labour/no disp. populations}
  if  4.5 and 5.5 then
  else if  then
  else if  then
  end if
  if  4.5 and 5.5 then
  else if  then
  else if  then
  end if
Algorithm 1 Calculate

Iv Dataset and metrics

There are a limited number of image datasets for human rights violations recognition [15, 14]. The most relevant for this work is HRA (Human Rights Archive) [15]. In order to find the main test platform on which we could demonstrate the effectiveness of GET-AID and analyse its various components, we construct a new image dataset by maintaining the verified samples intact for the two categories with more images, child labour and displaced populations. The HRA—Binary dataset contains 1554 images of human rights abuses and the same number of no violation counterparts for training, as well as 200 images collected from the web for testing and validation. Note that each abuse category is treated as an independent use case for our experiments. The dataset is made publicly available for future research.

Following [15], we evaluate GET-AID with two metrics accuracy and coverage and compare its performance against the sole use of a CNN classifier trained on HRA—Binary. We also evaluate the continuous dimensions using error rates - the difference (in average) between the true value and the regressed value.

V Experiments

V-a Implementation details

Our emotion recognition implementation is based on the emotion recognition in context (EMOTIC) model [17], with the difference that our model estimates only continuous dimensions in VAD space. We train the three main modules (Section III-B) on the EMOTIC database, which contains a total number of 18,316 images with 23,788 annotated people, using pre-trained CNN feature extraction modules. We treat this multiclass-multilabel problem as a regression problem by using a weighted Euclidean loss to compensate for the class imbalance of EMOTIC.

Table I shows the results for the continuous dimensions using error rates. The best result is obtained by utilising model ensembling, which consists of pooling together the predictions of a set of different models in order to produce better predictions. We pool the predictions of classifiers (ensemble the classifiers) by conducting weighted average of their prediction at inference time. The weights are learned on the validation data - usually the better single classifiers are given a higher weight, while the worse single classifiers are given a lower weight. However, ensembling the classifiers results in prolonged inference times, which causes us to turn our focus onto single classifiers for the remainder of the experiments.

Body Feature Backbone Mean error rate
VGG16 1.59
VGG19 1.57
ResNet50 1.69
VGG16 + ResNet50 1.40
VGG16 + VGG19 1.36
VGG19 + ResNet50 1.48
VGG19 + ResNet50 + VGG16 1.36
TABLE I: Mean error rate obtained (average of all three VAD dimensions) for different body feature backbone CNNs. The image feature backbone CNN was kept constant for all cases, namely VGG16-Places365 [32].

Following the two-phase transfer learning scheme proposed in [15], we fine-tune our human-rights-abuse classification models, Fig. 2(c), for 50 iterations on the HRA—Binary set with a learning rate of 0.0001 using the stochastic gradient descent (SGD) [19] optimizer for cross-entropy minimization. These vanilla models will be examined against GET-AID. Here, vanilla means pure image classification using solely fine-tuning without any alteration.

(a) (b) (c) (d)
(e) (f) (g) (h)
Fig. 5: Human rights abuses detected by our method. Each image shows two predictions alongside their probabilities. Top prediction is given by GET-AID, while the bottom prediction is given by the respective vanilla CNN. Green colour implies that no human right abuse is detected, while red colour signifies that a human right abuse is detected. In some instances such as (a), (b), (c), (e) and (h), GET-AID overturns the initial-false prediction of the vanilla CNN, where in other instances such as (d), (f) and (g), GET-AID strengthens the initial-true prediction, resulting in higher coverage.

V-B Baseline

To enable a fair comparison between vanilla CNNs and GET-AID, we use the same backbone combinations for all modules described in Fig. 2. We report comparisons in both accuracy and coverage metrics for fine-tuning up to three convolutional layers. The per-network results for child labour and displaced populations are shown in Table II and Table III respectively. The implementation of vanilla CNNs is solid: it has up to 65% accuracy on the child labour classification and up to 80% accuracy on the displaced populations classification, 29.82 and 44.82 points higher 111Only cases with a single or two layers fine-tuned were considered for calculating these numbers in order to be consistent with the implementation of [15] than the 35.18 reported in the single-label, multiclass classification problem of [15]. Regarding coverage, vanilla CNNs achieve up to 64% for child labour classification and up to 67% child displaced populations classification, which is in par with the 64% maximum coverage reported in [15]. We believe that this accuracy gap is mainly due to the fact that [15] deals with a multiclass classification problem, whereas in this work we classify inputs into two mutually exclusive classes. This comparison showed that it is possible to trade coverage with accuracy in the context of human rights image analysis. One can always obtain high accuracy by refusing to process a number of examples, but this reduces the coverage of the system. Nevertheless, vanilla CNNs provide a strong baseline to which we will compare our method.

Our method, GET-AID, has a mean coverage of 64.75% for child labour and 46.11% for displaced populations on the HRA—Binary set. This is an absolute gain of 12.42 and 16.78 points over the strong baselines of 52.33% and 29.33% respectively. This is a relative improvement of 23.73% and 57.21% respectively. In relation to accuracy, GET-AID has a mean accuracy of 54% for child labour and 61.25% for displaced populations, which is an absolute drop of 4 and 5.5 points over the strong baselines of 58% and 66.75% accordingly. This indicates a relative loss of only 6.89% and 8.23% respectively. We believe that this negligible drop in accuracy is mainly due to the fact that the HRA—Binary set is not solely made up of images with people in their context, it also contains images of generic objects and scenes, where only the sole classifier’s prediction is taken into account.

vanilla CNN GET-AID
Top-1 acc. Coverage Top-1 acc. Coverage
VGG16 1 62% 73% 56% 78%
VGG19 65% 30% 57% 55%
ResNet50 51% 0% 50% 24%
Places365 59% 71% 54% 81%
VGG16 2 61% 77% 59% 78%
VGG19 61% 64% 59% 76%
ResNet50 52% 0% 49% 33%
Places365 54% 44% 52% 65%
VGG16 3 56% 83% 56% 84%
VGG19 55% 87% 55% 82%
ResNet50 50% 99% 48% 91%
Places365 67% 0% 53% 30%
mean - 58% 52.33% 54% 64.75%
TABLE II: Detailed results on HRA—Binary for the child labour scenario. We show the main baseline and GET-AID for various network backbones. We bold the leading entries on coverage.
vanilla CNN GET-AID
Top-1 acc. Coverage Top-1 acc. Coverage
VGG16 1 58% 0% 56% 24.4%
VGG19 69% 3% 59% 33%
ResNet50 60% 0% 53% 29%
Places365 64% 3% 54% 32%
VGG16 2 63% 43% 60% 58%
VGG19 77% 54% 70% 59%
ResNet50 42% 1% 44% 33%
Places365 80% 49% 73% 58%
VGG16 3 72% 69% 67% 71%
VGG19 82% 64% 77% 68%
ResNet50 53% 0% 51% 22%
Places365 81% 66% 71% 66%
mean - 66.75% 29.37% 61.25% 46.11%
TABLE III: Detailed results on HRA—Binary for the displaced populations scenario. We show the main baseline and GET-AID for various network backbones. We bold the leading entries on coverage.

V-C Qualitative results

(a) (b)
Fig. 6: False positive detections of our method.

We show our human rights abuse detection results in Fig. 5. Each subplot illustrates two predictions alongside their probability scores. The top of the two predictions is given by GET-AID, while the bottom one is given by the respective vanilla CNN sole classifier. Our method can successfully classify human right abuses by overturning the initial-false prediction of the vanilla CNN in Fig. 5 (a), (b), (c), (e) and (h). Moreover, GET-AID can strengthen the initial-true prediction of the sole classifier as shown in Fig. 5 (d), (f) and (g) respectively.

V-D Failure cases

Fig. 6 shows some failure detections. Our method can be incorrect, because of false global emotional traits inferences. Some of them are caused by a failure of continuous dimensions emotion recognition, which is an interesting open problem for future research.

Vi Conclusion

We have presented a human-centric approach for classifying two types of human rights abuses. This two-class human rights abuse labelling problem is not trivial, given the high-level image interpretation required. Understanding what a person is experiencing from his frame of reference is closely related with situations where human rights are being violated. Thus, the key to our computational framework are people’s emotional traits, which resonate well with our own common sense in judging potential human right abuses. We introduce global emotional traits of the image which are responsible for weighting the classifiers prediction during inference. We benchmark performance of our GET-AID model against sole CNN classifiers. Our experimental results showed that this is an effective strategy, which we believe has good potential beyond human rights abuses classification. We hope this paper will spark interest and subsequent work along this line of research. All our code and data are publicly available.


  • [1] Martín Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, et al.

    Tensorflow: a system for large-scale machine learning.

    In OSDI, volume 16, pages 265–283, 2016.
  • [2] Jay D Aronson. Computer vision and machine learning for human rights video analysis: Case studies, possibilities, concerns, and limitations. Law & Social Inquiry, 2018.
  • [3] Jay D Aronson, Shicheng Xu, and Alex Hauptmann. Video analytics for conflict monitoring and human rights documentation. Center for Human Fights Science Technical Report, 2015.
  • [4] François Chollet et al. Keras., 2015.
  • [5] Wen-Sheng Chu, Fernando De la Torre, and Jeffrey F Cohn. Selective transfer machine for personalized facial expression analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39(3):529–545, 2017.
  • [6] Paul Ekman and Wallace V Friesen. Constants across cultures in the face and emotion. Journal of personality and social psychology, 17(2):124, 1971.
  • [7] Stefanos Eleftheriadis, Ognjen Rudovic, and Maja Pantic. Joint facial action unit detection and feature fusion: A multi-conditional learning approach. IEEE Transactions on Image Processing, 25(12):5727–5742, 2016.
  • [8] C Fabian Benitez-Quiroz, Ramprakash Srinivasan, and Aleix M Martinez. Emotionet: An accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild. In

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition

    , pages 5562–5570, 2016.
  • [9] E Friesen and P Ekman. Facial action coding system: a technique for the measurement of facial movement. Palo Alto, 1978.
  • [10] Cheng-Yang Fu, Wei Liu, Ananth Ranga, Ambrish Tyagi, and Alexander C. Berg. Dssd : Deconvolutional single shot detector. CoRR, abs/1701.06659, 2017.
  • [11] Ross Girshick. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, pages 1440–1448, 2015.
  • [12] Ross Girshick, Jeff Donahue, Trevor Darrell, and Jitendra Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 580–587, 2014.
  • [13] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 770–778, 2016.
  • [14] Grigorios Kalliatakis, Shoaib Ehsan, Maria Fasli, Ales Leonardis, Juergen Gall, and Klaus D. McDonald-Maier. Detection of human rights violations in images: Can convolutional neural networks help? In Proceedings of the 12th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications - Volume 5: VISAPP, (VISIGRAPP 2017), pages 289–296. SciTePress, 2017.
  • [15] Grigorios Kalliatakis, Shoaib Ehsan, Ales Leonardis, Maria Fasli, and Klaus D. McDonald-Maier. Exploring object-centric and scene-centric cnn features and their complementarity for human rights violations recognition in images. IEEE Access, 7:10045–10056, 2019.
  • [16] Grigorios Kalliatakis, Shoaib Ehsan, and Klaus D. McDonald-Maier. A Paradigm Shift: Detecting Human Rights Violations Through Web Images. In: Proceedings of the Human Rights Practice in the Digital Age Workshop; 2017.
  • [17] Ronak Kosti, Jose M Alvarez, Adria Recasens, and Agata Lapedriza. Emotion recognition in context. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), volume 1, 2017.
  • [18] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, pages 1097–1105, 2012.
  • [19] Yann LeCun, Bernhard Boser, John S Denker, Donnie Henderson, Richard E Howard, Wayne Hubbard, and Lawrence D Jackel. Backpropagation applied to handwritten zip code recognition. Neural Computation, 1(4):541–551, 1989.
  • [20] Tsung-Yi Lin, Priyal Goyal, Ross Girshick, Kaiming He, and Piotr Dollár. Focal loss for dense object detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018.
  • [21] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European Conference on Computer Vision, pages 740–755. Springer, 2014.
  • [22] Wei Liu, Dragomir Anguelov, Dumitru Erhan, Christian Szegedy, Scott Reed, Cheng-Yang Fu, and Alexander C Berg. Ssd: Single shot multibox detector. In European Conference on Computer Vision, pages 21–37. Springer, 2016.
  • [23] Albert Mehrabian. Framework for a comprehensive description and measurement of emotional states. Genetic, social, and general psychology monographs, 1995.
  • [24] Albert Mehrabian and James A Russell. An approach to environmental psychology. the MIT Press, 1974.
  • [25] Enrique Piracés. The Future of Human Rights Technology, page 289-308. Cambridge University Press, 2018.
  • [26] Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 779–788, 2016.
  • [27] Joseph Redmon and Ali Farhadi. Yolo9000: Better, faster, stronger. 2017 IEEE Conference on Computer Vision and Pattern Recognition, pages 6517–6525, 2017.
  • [28] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in neural information processing systems, pages 91–99, 2015.
  • [29] Pierre Sermanet, David Eigen, Xiang Zhang, Michaël Mathieu, Rob Fergus, and Yann LeCun. Overfeat: Integrated recognition, localization and detection using convolutional networks. International Conference on Learning Representations (ICLR), 2014.
  • [30] Mohammad Soleymani, Sadjad Asghari-Esfeden, Yun Fu, and Maja Pantic. Analysis of eeg signals and facial expressions for continuous emotion detection. IEEE Transactions on Affective Computing, (1):17–28, 2016.
  • [31] Jasper RR Uijlings, Koen EA Van De Sande, Theo Gevers, and Arnold WM Smeulders. Selective search for object recognition. International Journal of Computer Vision, 104(2):154–171, 2013.
  • [32] Bolei Zhou, Agata Lapedriza, Aditya Khosla, Aude Oliva, and Antonio Torralba. Places: A 10 million image database for scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(6):1452–1464, 2018.