Classifying Documents within Multiple Hierarchical Datasets using Multi-Task Learning

06/06/2017
by   Azad Naik, et al.
0

Multi-task learning (MTL) is a supervised learning paradigm in which the prediction models for several related tasks are learned jointly to achieve better generalization performance. When there are only a few training examples per task, MTL considerably outperforms the traditional Single task learning (STL) in terms of prediction accuracy. In this work we develop an MTL based approach for classifying documents that are archived within dual concept hierarchies, namely, DMOZ and Wikipedia. We solve the multi-class classification problem by defining one-versus-rest binary classification tasks for each of the different classes across the two hierarchical datasets. Instead of learning a linear discriminant for each of the different tasks independently, we use a MTL approach with relationships between the different tasks across the datasets established using the non-parametric, lazy, nearest neighbor approach. We also develop and evaluate a transfer learning (TL) approach and compare the MTL (and TL) methods against the standard single task learning and semi-supervised learning approaches. Our empirical results demonstrate the strength of our developed methods that show an improvement especially when there are fewer number of training examples per classification task.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/27/2018

Multi-task Learning of Pairwise Sequence Classification Tasks Over Disparate Label Spaces

We combine multi-task learning and semi-supervised learning by inducing ...
research
11/19/2017

Compression-Based Regularization with an Application to Multi-Task Learning

This paper investigates, from information theoretic grounds, a learning ...
research
11/29/2021

Learning Multiple Dense Prediction Tasks from Partially Annotated Data

Despite the recent advances in multi-task learning of dense prediction p...
research
11/14/2018

Efficient and Scalable Multi-task Regression on Massive Number of Tasks

Many real-world large-scale regression problems can be formulated as Mul...
research
11/18/2015

Efficient Output Kernel Learning for Multiple Tasks

The paradigm of multi-task learning is that one can achieve better gener...
research
05/22/2018

Infinite-Task Learning with Vector-Valued RKHSs

Machine learning has witnessed the tremendous success of solving tasks d...
research
07/25/2017

A Survey on Multi-Task Learning

Multi-Task Learning (MTL) is a learning paradigm in machine learning and...

Please sign up or login with your details

Forgot password? Click here to reset