## 1. Introduction

It has been observed a predictive learning task can admit a hypothesis of high prediction accuracy but low prediction fairness (i.e. the model predictions are illegally biased against disadvantaged instances). This has raised an argument for the fairness-accuracy trade-off in machine learning, and many studies propose methods to balance prediction fairness and accuracy.

However, existing studies focus on a single learning task,
while in many problems one can (formalize and) jointly learn
multiple related tasks in hope of boosting their overall
prediction accuracy – a topic commonly referred
as transfer learning (Pan and Yang, 2010).
What is the interplay between fairness and transfer
learning? This motivates the presented study.
^{†}^{†}This paper is presented as a
poster at the 2017 Workshop on Fairness, Accountability and
Transparency in Machine Learning (FAT/ML 2017).

In this paper, we show standard transfer learning can improve prediction accuracies of participating tasks (as usual), but at the price of lowering their originally high prediction fairness. We name this phenomenon discriminatory transfer. Our study reveals the information transfer process as a new cause of unfair model predictions, even when each participating task by itself admits a model with fair predictions. How does discriminatory transfer happen, and how to mitigate it while retaining the traditional gain of information transfer as much as possible remain open questions.

Our empirical examination is based on two popular and generic transfer learning algorithms, i.e. the hypothesis transfer algorithm formalized by Kuzborskij and Orabona(Kuzborskij and Orabona, 2013) and the multi-task algorithm formalized by Ciliberto et al (Ciliberto et al., 2015)

. The two algorithms correspond to two settings: the former assumes one task is pre-learned and its solution is used to assist learning another task; the latter assumes all tasks are learned jointly with constraints on task relatedness. Our main fairness measure is based on equalized odds proposed by Hardt et al

(Hardt et al., 2016), which is a conditional refinement of the legal notion disparate impact. We experiment on the Communities and Crime data set (Redmond and Baveja, 2002).## 2. Preliminaries

### 2.1. Notations and Problem Setting

Without loss of generality, we consider two supervised learning tasks

and . Task consists of a population , a distribution on , and a target predictive function ; the goal is to learn based on a random sample drawn from and labeled by , plus any information transferred from the other task. Task is defined in a similar manner. We assume both are embedded in for some , but they are not necessarily identical.### 2.2. Hypothesis Transfer Algorithm

Hypothesis transfer is a popular transfer learning paradigm, which aims to improve prediction accuracy of a task by additionally using a pre-learned predictive model of another related task. For more information on this topic, see (Kuzborskij and Orabona, 2013) and the reference therein.

We will examine a generic algorithm called Regularized Least Square (RLS) formalized and theoretically justified by Kuzborskij and Orabona (Kuzborskij and Orabona, 2013). The algorithm focuses on linear predictive functions, i.e. any function is uniquely parameterized by a

-dimensional vector

. Suppose task is pre-learned and its learned predictive function is used to assist task . Let be an training example of task and be its label. The RLS algorithm solves the following problem(1) |

where is a regularization coefficient. The authors showed the solution to (1) has an analytic form

(2) |

where is a data matrix with row representing example , and is a label vector with element representing the label of example .

It is clear controls the degree of information transferred from task to task , in a sense that larger will bias towards more strongly.

When two tasks do have similar target hypotheses, one may expect reasonably larger leads to a more accurate . (Of course, improperly large may mis-bias learning and lower the accurate, a problem known as negative transfer.) While this used to be a happy ending, our experimental study will show larger can meanwhile hurt prediction fairness.

### 2.3. Multi-Task Algorithm

Multi-task learning is another popular transfer learning paradigm, which aims to improve prediction accuracies of both tasks by learning them jointly with proper constraints on task relations.

We will examine a recent multi-task learning algorithm
developed by Ciliberto et al
at ^{†}^{†}https://github.com/cciliber/matMTL.
Recall is a random sample of task with
example and its label . Similarly, let
be a random sample of task with example
and its label .
The algorithm finds that minimize the following problem

(3) |

where

is loss function and

is Hilbert space of vector-valued functions with scalar components . The term encodes relation between and through a matrix such that, based on the Representer Theorem and other mild conditions, one has(4) |

where = and =. Micchelli and Pontil (Micchelli and Pontil, 2005) show proper choices of allow to capture certain established task relations.

Again, one may expect larger to give more accurate and , provided they are similar. Our experimental study will show this is indeed the case, but increasing also lowers prediction fairness of both and .

### 2.4. Fairness Measure

We will measure fairness based on the notion of equalized odds recently proposed by Hardt et al (Hardt et al., 2016). In each task, suppose the population is divided into two groups and , and one aims to examine prediction fairness between these groups. Let be a model for the task, we define the Equalized Odds Ratio (E.O.R.) of as

(5) |

Our proposed measure is first inspired by the well-known
-rule, which states gives fair prediction if the
following ratio^{†}^{†}We will refer this ratio
as Disparate Impact Ratio (D.I.R.). is no less than 80%

(6) |

However, it is argued the probabilities in (

6) do not guarantee equality of opportunity (Dwork et al., 2012), and authors in (Hardt et al., 2016) refine each probability by furthering conditioning it on (i.e. not disparate impact within truly qualified population and truly unqualified population). Our proposed measure is obtained by simply replacing the probabilities in (6) with their refinements in (Hardt et al., 2016).Similar to prior studies, we say the prediction fairness of is improved as approaches value 1, and is lowered as deviates from 1.

## 3. Experiment

### 3.1. Experimental Setup

We experimented on the Communities and Crime data set (Redmond and Baveja, 2002) obtained from the UCI data repository. The original data set collects for 1994 communities their crime rates and 122 normalized predictive attributes. The general goal of learning is to predict community crime rates based on these attributes.

We considered crime rate as label and binarized it such that crime rates above 50% were encoded as 1 and others encoded as -1; the attribute “percentage of population that is African American” was considered sensitive and binarized such that percentages above 50% were encoded as 1 and others encoded as 0 – based on this, all communities were partitioned into either African American (AA) communities (percentages ¿50%) or non-African American communities. We examined prediction discrimination against AA communities.

We simulated four transfer learning settings based on the data set, each consisting of two tasks.

Setting 1 partitioned communities based on the
population attribute, such that task is to
predict crime rates for communities with populations greater
than 0.05, and task for other communities^{†}^{†}For
all settings, our choice of the threshold (here, 0.05) was close
to attribute mean, but also for demonstrating smoother results..

Setting 2 partitioned communities based on the median house income attribute, such that task is to predict for communities with incomes greater than 0.25, and task for other communities.

Setting 3 partitioned communities based on the number of people under poverty level attribute, such that task is to predict for communities with poverty count greater than 0.02, and task for other communities.

Setting 4 partitioned communities based on the number of homeless people in shelters attribute, such that task is to predict for communities with homeless count greater than 0.03, and task for the rest.

In all settings, task was assumed pre-learned and
used to assist task . When switching tasks, we observed
similar trends but the results were numerically unstable and
thus not reported here.
Finally, all performance were evaluated over 10-fold cross
validation, with 1 fold used for training and 9 for testing,
and the averaged results were reported^{†}^{†}We chose 1 fold
for training to better demonstrate the well-known advantage of
transfer learning on smaller samples.
Our source code is at https://github.com/petelan/FATML2017.

### 3.2. Example Results

In this section, we show example results of discriminatory transfer in the hypothesis transfer (H.T.) algorithm under setting 2. For the pre-learned task, we used ridge regression as the base learner, with regularization coefficient fixed to 1. To see the interaction between fairness and transfer learning, we varied coefficient

in (1) and obtained results in Figure 1.From Figure 1 (a), we observe prediction accuracy is improved as increases from to , then mildly degrades and eventually converges. This is a somewhat common observation: it shows the traditional benefit of hypothesis transfer within range , and implies the true task relation is captured at around ; as one reinforces task relation beyond that range, however, prediction accuracy can decrease.

Our new observations come from Figure 1 (b). We see as increases from 1e-8 to 1e-2, prediction accuracy is increased while prediction fairness is decreased – this is the discriminatory transfer phenomenon.

One may wonder whether the prediction fairness of the pre-learned hypothesis would have any impact on the result. Our next result suggests it may have an impact, but discriminatory transfer may still exist.

Recall we have fixed in the pre-learned task. We noticed decreasing could result in hypotheses with higher prediction fairness though lower prediction fairness. See Table 1. So one may say our results in Figure 1 is based on an ‘unfair’ pre-learned hypothesis. We now repeat the same experiment but with smaller ’s; results are reported in Figure 2.

Let us focus on the case when =1e-4, which means the pre-learned hypothesis is fair. As increases from 1e-7 to 1e-6, we observe what may still be considered as discriminatory transfer – prediction accuracy increases while prediction fairness decreases. As continues to increase from 1e-6 to 1e-5, however, prediction fairness starts to grow back while prediction accuracy continues to increase. This is something new, and suggests there may be ways to mitigate discriminatory transfer. When grows larger than 1e-5, prediction accuracy starts to drop (hence the traditional benefit of transfer learning is gone); we are not particularly interested in the results.

How does discriminatory transfer occur? How to mitigate it while maximally maintaining the traditional accuracy improvement? What are other interactions between fairness and transfer learning? These are open questions. We conjecture standard transfer process may rule out some fair hypotheses when biasing target task learning.

Accuracy | Fairness (E.O.R / D.I.R.) | |
---|---|---|

1 | 0.7962 | 0.2908 / 0.1048 |

1e-1 | 0.7518 | 0.4585 / 0.2532 |

1e-2 | 0.6827 | 0.7125 / 0.4955 |

1e-3 | 0.6277 | 0.9370 / 0.7152 |

1e-4 | 0.6068 | 0.9787 / 0.7991 |

1e-5 | 0.6039 | 0.9729 / 0.8105 |

### 3.3. Other Results

In this section, we show further evidence of discriminatory transfer on both hypothesis transfer and multi-task algorithms over four transfer learning settings. For convenience, in one experiment both prediction accuracy and fairness will be shown in the same figure; in these cases, the left vertical axis of the figure will represent accuracy and the right vertical axis will represent fairness (as measured by the equalized odds ratio).

We first examined the performance of hypothesis transfer algorithm on four transfer settings. The results are shown in Figure 3. One can easily observe discriminatory transfer in all settings.

Next, we examined the performance of multi-task algorithm (3). The source code provides multiple options of the output kernel learning modalities, and we only presented results based on the Frobenius modality (as similar trends were observed using other modalities). The performance of each task in setting 1 is shown in Figure 4, where we gradually increased to enforce stronger task relation.

Discriminatory transfer can be seen in each task, i.e. as increases, both tasks have improved prediction accuracies but lowered fairness. Similar phenomena are observed in other settings, as shown in Figure 5.

## 4. Discussion

The presented study shows standard transfer learning can improve prediction accuracy of target tasks at the cost of lowering their prediction fairness. How to explain this phenomenon in principle, and how to minimize fairness damage while retaining the traditional accuracy improvement gained by transfer learning remain open questions.

Zemel et al (Zemel et al., 2013) had a very interesting result suggesting fairness may be transferrable: they showed a fair feature representation learned for one task can be used in another task to improve the latter’s prediction fairness. We studied a fundamentally different problem, and showed discrimination can be induced from the transfer process, even when the pre-learned hypothesis is fair. Besides, their transfer setting is different from ours: they focused on feature learning and treated different label sets of the same population as different tasks, while ours focused on predictive learning and treated the same label set of different populations as different tasks – a setting of domain adaptation (Pan and Yang, 2010).

The present study is motivated by negative transfer proposed by Rosenstein et al in (Rosenstein et al., 2005); they showed learning tasks jointly may not improve their prediction accuracies if the tasks are not as similar as presumed. We had not come up with a similarly concrete hypothesis for discriminatory transfer, however, and only conjectured transfer process may rule out fair hypotheses when biasing target task learning.

###### Acknowledgements.

The authors would like to thank anonymous reviewers for their valuable comments and suggestions.## References

- (1)
- Ciliberto et al. (2015) Carlo Ciliberto, Youssef Mroueh, Tomaso Poggio, and Lorenzo Rosasco. 2015. Convex learning of multiple tasks and their structure. In International Conference on Machine Learning (ICML).
- Dwork et al. (2012) Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. In Innovations in Theoretical Computer Science Conference.
- Hardt et al. (2016) Moritz Hardt, Eric Price, and Nathan Srebro. 2016. Equality of opportunity in supervised learning. In Advances in Neural Information Processing Systems (NIPS).
- Kuzborskij and Orabona (2013) Ilja Kuzborskij and Francesco Orabona. 2013. Stability and hypothesis transfer learning. In International Conference on Machine Learning (ICML).
- Micchelli and Pontil (2005) Charles A Micchelli and Massimiliano Pontil. 2005. Kernels for multi-task learning. In Advances in Neural Information Processing Systems.
- Pan and Yang (2010) Sinno Jialin Pan and Qiang Yang. 2010. A survey on transfer learning. IEEE Transactions on Knowledge and Data Engineering 22, 10 (2010), 1345–1359.
- Redmond and Baveja (2002) Michael Redmond and Alok Baveja. 2002. A data-driven software tool for enabling cooperative information sharing among police departments. European Journal of Operational Research 141, 3 (2002), 660–678.
- Rosenstein et al. (2005) Michael T Rosenstein, Zvika Marx, Leslie Pack Kaelbling, and Thomas G Dietterich. 2005. To transfer or not to transfer. In NIPS 2005 Workshop on Transfer Learning, Vol. 898.
- Zemel et al. (2013) Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. 2013. Learning fair representations. In International Conference on Machine Learning (ICML).

Comments

There are no comments yet.