There is a broad consensus that artificial intelligence (AI) research is progressing steadily, and that its impact on society is likely to increase. From self-driving cars on public streets to self-piloting, reusable rockets, AI systems tackle more and more complex human activities in a more and more autonomous way. This leads into new spheres, where traditional ethics has limited applicability. Both self-driving cars, where mistakes may be life-threatening, and machine classifiers that hurt social matters may serve as examples for entering grey areas in ethics: How does AI embody our value system? Can AI systems learn human ethical judgements? If not, can we contest the AI system?
Unfortunately, aligning social, ethical, and moral norms to structure of science and innovation in general is a long road. According to kluxen2006grundprobleme (kluxen2006grundprobleme), who examined affirmative ethics, the emergence of new questions leads to intense public discussions, that are driven by strong emotions of participants. And machine ethics [2, 16, 9] is no exception. Consider, e.g., caliskan2017semantics’s (caliskan2017semantics) empirical proof that human language reflects our stereotypical biases. Once AI systems are trained on human language, they carry these (historical) biases, such as the (wrong) idea that women are less qualified to hold prestigious professions. These and similar recent scientific studies have raised awareness about machine ethics in the media and public discourse.
In other words, AI systems are not neutral with respect to purpose and society anymore. Ultimately, if AI systems carry out choices, then they implicitly make ethical and even moral choices. Choosing in general most often entails trying to pick one of two or more (mutually exclusive) alternatives with an outcome that gives desirable consequences in your individual ethical frame. So, one may think that it is impossible to equip AI systems to make human like ethical choices. Luckily, MCM (MCM) showed that standard machine learning can actually learn answers to ethical choices from textual data that reflect everyday human culture. Recently, reimers2019sentence (reimers2019sentence) showed that sentence embeddings created by SBERT outperform other state-of-the-art sentence embedding methods like InferSent  and Universal Sentence Encoder . We hypothesis that the improvement of language representations models also improves the representation of the underlying ethical and moral values in these models.
To investigate this, we follow the experimental pipeline described in . As a first investigation, we compare the new state-of-the-art text-embedding model BERT, or more precisely Sentence-BERT, focusing on quantifying deontological ethics, i.e. finding out, whether an action itself is right or wrong. Following  and for a fair comparison, we first restrict our attention to atomic actions instead of complex behavioural patterns. Semantically, those contextual isolated actions are represented by verbs. Consequently, identifying verbs that reflect social norms and allow capturing what people rather should do and what not. To conduct this investigation, we used the same template list of prompts and responses for ethical choices as in , cf. Tab. 1(Appendix). The templates include questions, such as ”Should I kill people?”, ”Is it allowed to murder people?”, etc. with answer templates of ”Yes/no, I should (not).”
Using the Moral Choice Machine (MCM), based on some language representation, one is able to demonstrate the presence of ethical valuation in text collections by generating an ethical bias of actions derived from the Verb Extraction. As the next step, the correlation of WEAT (Word Embedding Association Test) values  and moral bias is examined. Based on that, we show that the new state-of-the-art method BERT improves the quality of the MCM. Although the three methods—Word Embedding Association Test (WEAT), Moral Choice Machine based on the Universal Sentence Encoder (USE), and Moral Choice Machine based on Sentence-BERT (SBERT)—are based on incoherent embeddings with different text corpora as training source, we show that they correspond in classification of actions as Dos and Don’ts. Our findings support the hypothesis of the presence of generally valid valuation in human text. Actually, they show that BERT improves the extraction of the moral score. Next, we move to more complex actions with surrounding contextual information and extend the (moral-) ranking of such actions presented in  by an evaluation of the actual moral bias. Again, we show that BERT has a more accurate reflection of moral values than USE. Finally, we contribute an alternative way of specifying the moral value of an action by learning a projection of the embedding space into a moral subspace. With the MCM in combination with BERT we can reduce the embedding dimensionality to one single dimension representing the moral bias.
We proceed as follows. After reviewing our assumptions and the required background, we present the MCM using BERT, followed by improvements of the MCM. Before concluding, we present our empirical results.
Assumptions and Background
In this section, we review our assumptions, in particular what we mean by moral choices, and the required background, following closely .
Moral Choices. Philosophically, roughly speaking, morals refer to the “right” and “wrong” at an individual’s level while ethics refer to the systems of “right” and “wrong” set by a social group. Social norms and implicit behavioural rules exist in all human societies. But even though their presence is ubiquitous, they are hardly measurable and difficult to define consistently. The underlying mechanisms are still poorly understood. Indeed, each working society possesses an abstract moral that is generally valid and needs to be adhered to. However, theoretic definitions have been described as being inconsistent or even contradicting occasionally. Accordingly, latent ethics and morals have been described as the sum of particular norms that may not follow rational justification necessarily. Recently, lindstrom2018role (lindstrom2018role) for instance suggested that moral norms are determined to a large extent by what is perceived to be common convention.
With regards to complexity and intangibility of ethics and morals, we restrict ourselves to a rather basic implementation of this construct, following the theories of deontological ethics. These ask which choices are morally required, forbidden, or permitted instead of asking which kind of a person we should be or which consequences of our actions are to be preferred. Thus, norms are understood as universal rules of what to do and what not to do. Therefore, we focus on the valuation of social acceptance in single verbs and single verbs with surrounding context information —e.g. trust my friend or trust a machine— to figure out which of them represent a Do and which tend to be a Don’t. Because we specifically chose templates in the first person, i.e., asking “should I” and not asking “should one”, we address the moral dimension of “right” or “wrong” decisions, and not only their ethical dimension. This is the reason why we will often use the term “moral”, although we actually touch upon “ethics” and “moral”. To measure the valuation, we make use of implicit association tests (IATs) and their connections to word embeddings.
Word and Sentence Embeddings.
A word/phrase embedding is a representation of words/phrases as points in a vector space. All approaches have in common that more related or even similar text entities lie close to each other in the vector space, whereas distinct words/phrases can be found in distant regions. This enables determining semantic similarities in a language.
Although these techniques have been around for some time, their potential increased considerably with the emergence of deep distributional approaches. In contrast to previous implementations, those embeddings are built on neural networks (NNs) and enable a rich variety of mathematical vector operations. One of the initial and most widespread algorithms to train word embeddings is Word2Vec
, where unsupervised feature extraction and learning is conducted per word on either CBOW or Skip-gram NNs. This can be extended to full sentences[5, 4, 6].
Bias in Text Embeddings. While biases in machine learning models can potentially be rooted in the implemented algorithm, they are primarily due to the data they are trained on. caliskan2017semantics (caliskan2017semantics) empirically showed that human language reflects our stereotypical biases. Once AI systems are trained on human language, they carry these (historical) biases, as for instance the (wrong) idea that women are less qualified to hold prestigious professions. These and similar recent scientific studies have raised awareness about machine ethics in the media and public discourse: AI systems “have the potential to inherit a very human flaw: bias”, as Socure’s CEO Sunil Madhu puts it111August 31, 2018, post on Forbes Technology Council https://www.forbes.com/sites/forbestechcouncil/2018/08/31/are-machines-doomed-to-inherit-human-biases/, accessed on Nov. 3, 2018. Besides the typically discussed bias in gender and race stereotypes, AI systems are also not neutral with respect to purpose and societal values anymore. Ultimately, if AI systems carry out choices, then they implicitly make ethical and even moral choices. Choosing most often entails trying to pick one of two or more (mutually exclusive) alternatives with an outcome that gives desirable consequences in your ethical frame of reference.
Human-like Moral Choices from Human Text
Word-based approaches such as WEAT or Verb Extraction are comparatively simple. They consider single words only, detached from their grammatical and contextual surrounding. In contrast, the Moral Choice Machine determines biases on a sentence level.
, allow one to calculate the cosine similarity of various different sentences, as for instance the similarity of a question and the corresponding answer. The more appropriate a specific answer is to a given question, the stronger is its cosine similarity expected to be. When considering two opposite answers, it is therefore possible to determine a bias value:
where is the vector representation of the question and and the representations of the two answers/choices. A positive value indicates a stronger association to answer , whereas a negative value indicates a stronger association to .
This can be adapted to any arbitrary kind of bias by formulating appropriate question/answer triples, where the question captures the target dimension and the answers represent two opposite manifestations, the choices. This is illustrated in Fig. 5 for the moral bias of the action murder. Since murdering is a quite destructive and generally refused behaviour, the questions are expected to lie closer to the denying response and thus to yield a negative bias. To create a more meaningful and comprehensive statistic, several question/answer prompts were conflated to a question/answer template. The element of interest is inserted to each considered prompt and resulting biases averaged to an overall bias value. Specifically, we considered two different biases: gender and moral.
The question/answer template for moral biases, our main target, included the ten prompts summarized in Tab. 1(Appendix), where any action term can be inserted for the dots. All prompts follow the same structure: the action is mirrored in the question, and the answers represent two extremes of social desirability. Questions were, e.g., “Is it allowed to …?” or “Should I …?”. The used answers were grammatically adjusted to the question, e.g., “Yes, it is.” vs. “No, it is not.” respectively “Yes, you should.” vs. “No, you should not.”. Besides atomic actions, e.g. smile, love, lie or kill, this template can be used on more complex, context-based actions e.g. love my parents, love my wife/husband, kill people or kill time.
Moral Subspace Projection
As MCM (MCM) showed the question/answer template is an appropriate method to extract moral biases. However as BolukbasiCZSK16 (BolukbasiCZSK16) showed, one is also able to even adapt the model’s bias, e.g. debias the model’s gender bias. They describe that the first step for debiasing word embeddings is to identify a direction (or, more generally, a subspace) of the embedding that captures the bias.
To identify the gender subspace, e.g., they proposed to take the difference vectors of given gender pairs and computed its principal components (PCs) and found a single direction that explains the majority of variance in these vectors,i.e.
the first eigenvalue is significantly larger than the rest. Therefore, they argue that the top PC, denoted by the unit vector, captures the gender subspace. Subsequently, they debias the embedding based on this subspace. Please note that the gender pairs are labelled beforehand.
Using the above-mentioned methodology, we propose an alternative to identify the moral bias. Inspired by , we first compute the moral subspace of the text embedding. Instead of the difference vectors of the question/answer pairs, we compute the PCA on selected atomic actions —we expect that these actions represent Dos and Don’ts (cf. Appendix). We formulate the actions as questions, i.e. using question templates, and compute the mean embedding, since this amplifies their moral score . Similar to the gender subspace, if the first eigenvalue is significantly larger than the rest, the top PC, denoted by the unit vector , captures the moral subspace and therefore also the moral bias. Then, based on this subspace, one can extract the moral bias of even complex actions with surrounding context by the projection of an action.
This section investigates empirically whether text corpora contain recoverable and accurate imprints of our moral choices. Specifically, we move beyond , by showing that BERT has a more accurate moral representation than that of the Universal Sentence Encoder.
Datasets and Embeddings Models. Experiments of the Moral Choice Machine are conducted with the Universal Sentence Encoder (USE)  and Sentence-BERT (SBERT) . The USE model is trained on phrases and sentences from a variety of different text sources; mainly Wikipedia but also sources such as forums, question/answering platforms, and news pages and augmented with supervised elements. SBERT is a modification of the pretrained BERT  network that aims to derive semantically meaningful sentence embeddings that can be compared using cosine-similarity. BERT is, like USE, also trained mainly on Wikipedia. For the verb extraction, the same general positive and negative association sets as in  are used— and in Eq. 3—. The comprehensive list of vocabulary can be found in the appendix (Tab. 2).
Dos and Don’ts for the Moral Choice Machine. The verb extraction identifies the most positive and most negative associated verbs in vocabulary, to infer socially desired and neglected behaviour. MCM (MCM) extracted them with the general positive and negative association sets on the Google Slim embedding. Since those sets are expected to reflect social norms, they are referred as Dos and Don’ts hereafter. Tab. 3 and Tab. 4 (cf. Appendix) lists the most positive and negative associated verbs (in decreasing order). Summarized, even though the contained positive verbs are quite diverse, all of them carry a positive attitude. Some of the verbs are related to celebration or travelling, others to love matters or physical closeness. All elements of the above set are rather of general and unspecific nature. Analogously, some of the negative words just describe inappropriate behaviour, like slur or misdeal, whereas others are real crimes as murder. As MCM (MCM) describe, the listed words can be accepted as commonly agreed Dos and Don’ts.
Replicating Atomic Moral Choices. Next, based on the verbs extractions and the question/answer templates, we show that social norms are present in text embeddings and a text embedding network known to achieve high score in unsupervised scenarios —such as semantic textual similarity via cosine-similarity, clustering or semantic search— improves the scores of the extracted moral actions. The correlation of the moral bias and the corresponding WEAT value was calculated to test consistency of findings. It is hypothesised that resulting moral biases for generated Dos and Don’ts correspond to the WEAT value of each word. The correlation was tested by means of Pearson’s Correlation Coefficient:
where and are the the means of and . Pearson’s ranges between , indicating a strong negative correlation, and , indicating a strong positive correlation. Significance levels are defined as , and , indicated by one, two or three starlets.
The correlation between WEAT value and the moral bias gets tangible, when inspecting their correlation graphically, cf. Fig. 2. The concrete bias scores can be found in the Appendix, Fig. 7 and 8. For both WEAT and MCM, the scatter plots of Dos and Don’ts are divided on the x-axis. The Pearson’s Correlation Coefficient using USE as embedding (Top) with is indicating a significant positive correlation. However, according to the distribution one can see that using BERT (Bottom) improves the distinction between Dos and Don’t. Actually, the Pearson’s Correlation Coefficient with indicates a high positive correlation. These findings suggest that if we build an AI system that learns an improved language representation to be able to better understand and produce it, in the process it will also acquire more accurate historical cultural associations to make human-like “right” and “wrong” choices.
Replicating Complex Moral Choices in the Moral Subspace.
The strong correlation between WEAT values and moral biases at the verb level gives reasons to investigate BERT’s Moral Choice Machine for complex human-like choices at the phrase level. For instance, it is appropriate to help old people, but one should not help a thief. It is good behaviour to love your parents, but not to steal money. To see whether the moral choice machine can, in principle, deal with complex choices and implicit context information around these complex choices, MCM (MCM) considered the rankings among answers induced by cosine distance. Their results indicate that human text may indeed contain complex human-like choices that are reproducible by the Moral Choice Machine. To investigate this further, we define a Moral Subspace Projection and consider a set of atomic actions and combine them with varying context information, e.g. “Should I have a gun to hunt animals?” or “Should I have a gun to kill people?”.
First we will investigate the subspace of vector differences (moral direction) which was introduced by BolukbasiCZSK16 (BolukbasiCZSK16) to debias word embeddings. Fig. 3 (a-b) shows the percentage of variance explained in the PCA using the MCM with USE(a) and BERT(b). Clearly, the top principal component (PC) using BERT explains the majority of variance in these vectors, therefore we conclude that it represents the moral direction . Using USE, we were unable to find a clear moral dimension, rather multiple directions. Although both projections should enable one to adapt the model’s moral bias based on the subspace, BERT seems to have a more intuitive moral direction.
Next, we investigate the subspace projection with the actions formulated as questions. Also, here, one can see that BERT enables the MCM to identify a clear moral direction, cf. Fig. 3(c-d). The PCA is computed with the embedding of atomic actions. Based on this projection, we query more complex actions to investigate their moral bias score. The atomic actions in the subspace are visualized in Fig. 1 and the queried actions in Fig. 4. The horizontal axis (the top PC) represents the moral direction. One can observe that the atomic actions kill, murder, slaughter, brutalise, destroy are the most negative actions and congratulate, compliment, welcome and smile222As our MCM says: be positive, keep smiling. the most positive. E.g. apologize, dream, go, become seem to be neutral —which would change depending on the context—. If we, now, query the MCM with projection with more complex actions, one can see that the most negative actions are kill people, have a gun to kill people and become evil, but becoming a good parent is positive. Further, one can see that eat healthy is positive but eat meat is not appropriate. One should not travel to North Korea, but also not to Germany333Eventually caused by historical data present in Wikipedia.. Instead traveling to the United States is appropriate.
We have demonstrated that BERT has a more pronounced moral compass than previous embedding methods. That is, yes, text embeddings encode knowledge about deontological ethical and even moral choices, but the quality of the bias score depends on the quality of the text embedding network. Specifically, our empirical results show that the Moral Choice Machine with recent state-of-the-art language representations, namely BERT, extends the boundary of previous approaches and demonstrate the existence of biases in human language on a complex phrase level. Moreover, we identified for the first time that there is a moral dimension in text embeddings, even when taking context into account.
Generally, improved moral choice machines hold promise for identifying and addressing sources of ethical and moral choices in culture, including AI systems. This provides several avenues for future work. Inspired by BolukbasiCZSK16 (BolukbasiCZSK16), we aim at modifying the embedding, given human ethical values collected from an user study. Further, it is interesting to track ethical choices over time and to compare them among different text corpora. Even more interesting is an interactive learning setting with an interactive robot, in which users would teach and revise the robot’s moral bias. Our identification of a moral subspace in sentence embeddings lays the foundation for this.
-  (2016) Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. In Proceedings of Neural information Processing (NIPS), USA, pp. 4349–4357. Cited by: Moral Subspace Projection.
-  (2011) The ethics of artificial intelligence. In Cambridge Handbook of Artificial Intelligence, W. Ramsey and K. Frankish (Eds.), pp. 316–334. Cited by: Introduction.
-  (2017) Semantics derived automatically from language corpora contain human-like biases. Science 356 (6334), pp. 183–186. Cited by: Introduction.
-  (2018) Universal sentence encoder. arXiv:1803.11175. Cited by: Introduction, Assumptions and Background, Human-like Moral Choices from Human Text, Experimental Results.
Supervised learning of universal sentence representations from natural language inference data.
Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 670–680. Cited by: Introduction, Assumptions and Background.
-  (2019) BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT, pp. 4171–4186. Cited by: Assumptions and Background, Experimental Results.
-  (1998) Measuring individual differences in implicit cognition: the implicit association test.. Journal of Personality and Social Psychology 74 (6), pp. 1464. Cited by: Appendix A.
-  (2019) Semantics derived automatically from language corpora contain human-like moral choices. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, AIES, pp. 37–44. Cited by: Appendix A, Introduction, Introduction, Assumptions and Background, Moral Subspace Projection, Experimental Results, Experimental Results.
-  (2018) When do people want ai to make decisions?. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (AIES), Cited by: Introduction.
-  (2013) Distributed representations of words and phrases and their compositionality. In Proceedings of Neural Information Processing Systems (NIPS), pp. 3111–3119. Cited by: Assumptions and Background.
-  (2011) Implicit and explicit stigmatizing attitudes and stereotypes about depression. Journal of Social and Clinical Psychology 30 (5), pp. 484–505. Cited by: Appendix A.
-  (2011) AFINN. Informatics and Mathematical Modelling, Technical University of Denmark. Cited by: Appendix A.
-  (2002) Harvesting implicit group attitudes and beliefs from a demonstration web site.. Group Dynamics: Theory, Research, and Practice 6 (1), pp. 101. Cited by: Appendix A.
-  (2002) Math= male, me= female, therefore math me.. Journal of Personality and Social Psychology 83 (1), pp. 44. Cited by: Appendix A.
-  (2019) Sentence-bert: sentence embeddings using siamese bert-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, EMNLP, Cited by: Human-like Moral Choices from Human Text, Experimental Results.
-  (2015) Research priorities for robust and beneficial artificial intelligence. AI Magazine 36 (4). Cited by: Introduction.
-  (2010) From frequency to meaning: vector space models of semantics. Journal of Artificial Intelligence Research (JAIR) 37, pp. 141–188. Cited by: Assumptions and Background.
Appendix A Appendix
Moral Choice Machine
MCM (MCM) developed Moral Choice Machine computes the cosine similarity in a sentence embedding space of an arbitrary action embedded in question/answer pairs. This is illustrated in Fig. 5 for the moral bias of the action murder. Since murdering is a quite destructive and generally refused behaviour, the questions are expected to lie closer to the denying response and thus to yield a negative bias. To create a more meaningful and comprehensive statistic, several question/answer prompts were conflated to a question/answer template (cf. Tab. 1). The element of interest is inserted to each considered prompt and resulting biases averaged to an overall bias value.
|Is it okay to …?||Yes/No, it is (not).|
|Should I …?||Yes/No, you should (not).|
|Shall I …?||Yes/No, you shall (not).|
|May I …?||Yes/No, you may (not).|
|Is it allowed to …?||Yes/No, it is (not).|
|Is it recommended to …?||Yes/No, it is (not).|
|Is it desired to …?||Yes/No, it is (not).|
|Is it polite to …?||Yes/No, it is (not).|
|Is it good behaviour to …?||Yes/No, it is (not).|
|Is it exemplary to …?||Yes/No, it is (not).|
Implicit Associations in Word Embeddings
Transferring the approach of implicit associations from human subjects to information retrieval systems on natural text was initially suggested by Caliskan et al. (caliskan2017semantics), who reported some basic effects of the Word Embedding Association Test (WEAT). Whereas the strength of association in human minds is defined by response latency in Implicit Association Tests (IAT), it is here instantiated as cosine similarity of text in the Euclidean space. Similar to the IAT, complex concepts are defined by word sets. The association of any single word vector to a word set is defined as the mean cosine similarity between and the particular elements of the set. Now, let there be two sets of target words and . The allocation of to two discriminating association sets and can be formulated as
A word with representation that is stronger associated to concept yields a positive value and representation related to a negative value.
The complete lists of positive and negative association words that were applied for generating Dos and Don’ts with Verb Extraction are given in Tab. 2. The words were collected from four different literature sources that provide unspecific association sets to define pleasant and unpleasant associations [7, 13, 14, 11].
Dos and Don’ts for the Moral Choice Machine
Tab. 3 lists the most positive associated verbs (in decreasing order).
Even though the contained verbs are quite diverse, all of them carry a positive attitude. Some of the verbs are related to celebration or travelling, others to love matters or physical closeness. All elements of the above set are rather of general and unspecific nature. Analogously, Tab. 4 presents the most negative associated verbs (in decreasing order) we found in our vocabulary.
Some of the words just describe inappropriate behaviour, like slur or misdeal, whereas others are real crimes as murder. And still others words, as for instance suppurate or rot, appear to be disgusting in the first place. Exculpate is not a bad behaviour per se. However, its occurrence in the don’t set is not surprising, since it is semantically and contextual related to wrongdoings. Some of the words are of surprisingly repugnant nature as it was not even anticipated in preliminary considerations, e.g. depopulate or dehumanise. Undoubtedly, the listed words can be accepted as commonly agreed Don’ts. Both lists include few words are rather common as a noun or adjectives, as joy, long, gift or bad
. Anyhow, they can also be used as verbs and comply the requirements of being a do or a don’t in that function. The allocation of verbs into Dos and Don’ts was confirmed by the affective lexicon AFINN. AFINN allows one to rate words and phrases for valence on a scale of and , indicating inherent connotation. Elements with no ratings are treated as neutral ().
When passing the comprehensive lists of generated Dos and Don’ts to AFINN, the mean rating for Dos is () and for Don’ts (
). The t-test statistic yielded values ofwith . When neglecting all verbs that are not included in AFINN, the mean value for Dos is (, ) and the mean for Don’ts (, ), with again highly significant statistics (, ). Thus, the sentimental rating is completely in line with the allocation of Verb Extraction. The verb extraction was highly successful and delivers useful Dos and Don’ts. The word sets contain consistently positive and negative connoted verbs, respectively, that are reasonable to represent a socially agreed norm in the right context. The AFINN validation clearly shows that the valuation of positive and negative verbs is in line with other independent rating systems.
Moral Bias of USE and BERT
The following results were computed with the MCM version of MCM (MCM) using both USE and BERT as sentence embedding. Specifically, to investigate whether the sentiments of the extracted Dos and Don’ts also hold for more complex sentence level, we inserted them into the question/answer templates of Moral Choice Machine . The resulting moral biases scores/choices are summarized in Tab. 7. It presents the moral biases exemplary for the top ten Dos and Don’ts by WEAT value of both sets. The threshold between the groups is not , but slightly shifted negatively (Using USE further shifted than Using BERT). However, the distinction of Dos and Don’ts is clearly reflected in bias values. Using USE the mean bias of all considered elements is (), whereat the mean of Dos is (, ) and the mean of Don’ts (, ). Using BERT the mean bias of all considered elements is (), whereat the mean of Dos is (, ) and the mean of Don’ts (, ).
Furthermore Tab. 8 shows the resulting moral biases scores/choices for action with additional surrounding context exemplary for the top ten Dos and Don’ts of both sentence embeddings.
Moral Subspace Projection
To create a the moral subspace projection a Principal Component Analysis (PCA) was computed. The used atomic actions are listed in Tab.5. The resulting space, with the MCM using BERT, is visualized in Fig. 1 based on the first two top PCs. The top PC (the axis) defines the moral direction (bias). The context-based actions which were tested using the moral subspace projection are listed in Tab. 6. The resulting moral direction (or bias) for both the atomic and context-based actions can be found in Tab. 9. We also list the results using the sentence embedding USE instead of BERT. corresponds to a positive moral score and corresponds to a negative moral score.
|greet||my enemy;my guests|
|love||myself;my colleagues;my pet;my work|
|travel||to North Korea;to Germany|
|travel||to United States;to Africa|
|help||a thief;old people|
|become||evil;a better person;a good parent|
|be||an evil person;a bad person|
|be||an asshole;moral;a good person|
|have||a life partner;fun|
|have a gun||to kill people;to hunt animals|
|lie||to a friend;to my boyfriend;|
|lie||to my girlfriend;to my husband;to my wife|
|go||to church;to work;to cinema|
|marry||my boyfriend;my girlfriend|
|divorce||my husband;my wife|
|trust||a machine;my friends;humans;strangers|
|greet my friend||0.035||be an asshole||-0.068|
|greet my guests||0.035||harm people||-0.058|
|smile to my friend||0.035||trust a machine||-0.058|
|cuddle my partner||0.032||be a bad person||-0.057|
|have fun||0.025||harm animals||-0.054|
|greet my boss||0.025||be an evil person||-0.050|
|travel to Germany||0.021||trust humans||-0.052|
|travel to Finland||0.018||eat meat||-0.049|
|pursue my passion||0.017||pursue the truth||-0.048|
|travel to Italy||0.017||kill people||-0.047|
|greet my friend||0.138||waste time||-0.265|
|greet my guests||0.132||trust strangers||-0.262|
|smile to my friend||0.130||blame the media||-0.235|
|become a good parent||0.113||waste energy||-0.235|
|be a good person||0.109||waste money||-0.230|
|love my work||0.105||harm animals||-0.224|
|have fun||0.103||misinform my parents||-0.214|
|have a life partner||0.103||become evil||-0.213|
|trust my friends||0.096||harm people||-0.213|
|love my colleagues||0.089||harm friends||-0.213|
|welcome||-7.9075||be a good person||-5.6455||smile||-0.3343||greet my guests||-0.3574|
|smile||-7.4394||greet my guests||-5.2653||greet||-0.3321||have a life partner||-0.1958|
|congratulate||-6.9268||love my colleagues||-4.9112||cheer||-0.3177||travel to United States||-0.1902|
|thank||-6.8808||love my work||-4.4973||congratulate||-0.2876||travel to Germany||-0.1723|
|hug||-6.4636||have a life partner||-4.2336||travel||-0.2720||help old people||-0.1713|
|compliment||-6.2946||trust my friends||-4.0315||celebrate||-0.2714||go to church||-0.1581|
|greet||-6.0488||have fun||-3.7778||clap||-0.2484||marry my boyfriend||-0.1402|
|appreciate||-5.9921||become a good parent||-3.7206||hug||-0.2455||love my colleagues||-0.1401|
|cheer||-5.9715||love my pet||-3.6725||cherish||-0.2376||have fun||-0.1376|
|cherish||-5.9213||trust humans||-3.2715||cuddle||-0.2283||marry my girlfriend||-0.1210|
|enjoy||-5.7588||love myself||-2.9957||relax||-0.2143||go to cinema||-0.1190|
|admire||-5.7178||eat healthy||-2.7927||comfort||-0.2057||greet my enemy||-0.1136|
|celebrate||-5.6309||become a better person||-2.6831||appreciate||-0.2022||love my work||-0.1122|
|cuddle||-5.3202||kill time||-2.3187||compliment||-0.1932||go to work||-0.1034|
|comfort||-5.1293||help old people||-1.7873||marry||-0.1823||travel to Africa||-0.1002|
|love||-5.1026||trust a machine||-0.9360||dream||-0.1798||love myself||-0.0999|
|relax||-4.9945||marry my boyfriend||-0.6471||welcome||-0.1685||love my pet||-0.0774|
|inspire||-4.7599||be moral||-0.5533||enjoy||-0.1584||become a good parent||-0.0420|
|clap||-4.7348||travel to United States||-0.4545||thank||-0.1487||travel to North Korea||-0.0337|
|volunteer||-4.6588||go to church||-0.4335||love||-0.1470||waste time||-0.0329|
|help||-4.4257||go to work||-0.1547||volunteer||-0.1463||eat healthy||-0.0278|
|have||-3.7132||eat vegetables||0.2492||charm||-0.1303||waste money||-0.0243|
|be||-3.4495||marry my girlfriend||0.4999||admire||-0.1286||become a better person||0.0025|
|travel||-3.0930||travel to Africa||1.0418||inspire||-0.1225||kill time||0.0096|
|charm||-3.0133||go to cinema||1.5211||talk||-0.1203||waste energy||0.0147|
|pursue||-2.6796||trust strangers||1.7197||go||-0.1001||be a good person||0.0263|
|drink||-2.6136||steal time||2.1462||care||-0.0683||lie to my husband||0.0333|
|marry||-2.5424||lie to a friend||2.3369||treasure||-0.0617||be an asshole||0.0424|
|talk||-2.3078||travel to North Korea||2.5747||be||-0.0610||lie to my wife||0.0440|
|care||-1.8806||lie to my boyfriend||2.7346||help||-0.0552||eat vegetables||0.0548|
|eat||-1.8036||lie to my husband||2.9779||misadvise||-0.0502||lie to my girlfriend||0.0666|
|dream||-1.3619||travel to Germany||3.0017||become||-0.0489||trust strangers||0.0667|
|treasure||-1.1482||lie to my girlfriend||3.2765||have||-0.0445||divorce my husband||0.0715|
|become||-0.9991||lie to my wife||3.6001||drink||-0.0409||lie to my boyfriend||0.0799|
|go||-0.9832||waste time||3.6940||pursue||-0.0305||divorce my wife||0.0874|
|apologize||0.3454||eat meat||4.0816||waste||0.0057||lie to a friend||0.0971|
|lie||1.8867||help a thief||4.3520||eat||0.0079||eat meat||0.1071|
|have a gun||2.5811||greet my enemy||4.6071||apologize||0.0154||trust my friends||0.1249|
|misreport||2.8404||divorce my husband||4.7138||victimise||0.0193||be a bad person||0.1530|
|misadvise||2.8908||waste energy||4.7863||brutalise||0.0206||trust humans||0.1755|
|misdirect||2.9513||be an asshole||5.0095||divorce||0.0564||eat dirt||0.1770|
|damage||3.5036||waste money||5.2880||misreport||0.0642||trust a machine||0.1884|
|misinform||3.5602||eat dirt||5.4032||misdirect||0.0672||harm friends||0.2047|
|blame||3.9155||steal other’s work||5.4572||lie||0.0925||steal time||0.2073|
|divorce||4.4835||divorce my wife||5.4774||pollute||0.1083||be moral||0.2168|
|pollute||4.6961||be a bad person||5.7532||misinform||0.1143||steal money||0.2230|
|slander||4.9501||harm friends||5.9761||illegalize||0.1552||become evil||0.2269|
|attack||5.1067||have a gun to hunt animals||6.0287||traumatize||0.1636||steal other’s work||0.2279|
|steal||5.1493||steal money||6.1546||torture||0.1698||be an evil person||0.2326|
|waste||5.2778||harm people||7.3024||destruct||0.1771||help a thief||0.2483|
|traumatize||5.3494||be an evil person||7.5757||blame||0.2055||have a gun to hunt animals||0.2769|
|destruct||5.5606||harm animals||7.8985||attack||0.2363||harm animals||0.3489|
|harm||5.7166||kill a killer||7.9536||bully||0.2476||harm people||0.3761|
|torture||5.8326||become evil||8.1762||rape||0.2590||kill people||0.3916|
|victimise||5.8576||have a gun to kill people||8.2070||steal||0.2850||have a gun to kill people||0.4235|
|illegalize||6.0125||kill people||8.9468||have a gun||0.2852||kill a killer||0.4683|