Multi-Objective Optimization for Self-Adjusting Weighted Gradient in Machine Learning Tasks

06/03/2015
by   Conrado Silva Miranda, et al.
0

Much of the focus in machine learning research is placed in creating new architectures and optimization methods, but the overall loss function is seldom questioned. This paper interprets machine learning from a multi-objective optimization perspective, showing the limitations of the default linear combination of loss functions over a data set and introducing the hypervolume indicator as an alternative. It is shown that the gradient of the hypervolume is defined by a self-adjusting weighted mean of the individual loss gradients, making it similar to the gradient of a weighted mean loss but without requiring the weights to be defined a priori. This enables an inner boosting-like behavior, where the current model is used to automatically place higher weights on samples with higher losses but without requiring the use of multiple models. Results on a denoising autoencoder show that the new formulation is able to achieve better mean loss than the direct optimization of the mean loss, providing evidence to the conjecture that self-adjusting the weights creates a smoother loss surface.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/03/2020

Multi-Loss Weighting with Coefficient of Variations

Many interesting tasks in machine learning and computer vision are learn...
research
12/30/2021

A Unified DRO View of Multi-class Loss Functions with top-N Consistency

Multi-class classification is one of the most common tasks in machine le...
research
02/08/2021

Multi-Objective Learning to Predict Pareto Fronts Using Hypervolume Maximization

Real-world problems are often multi-objective with decision-makers unabl...
research
09/16/2021

SLAW: Scaled Loss Approximate Weighting for Efficient Multi-Task Learning

Multi-task learning (MTL) is a subfield of machine learning with importa...
research
04/25/2023

Loss and Reward Weighing for increased learning in Distributed Reinforcement Learning

This paper introduces two learning schemes for distributed agents in Rei...
research
01/01/2022

Some connections between higher moments portfolio optimization methods

In this paper, different approaches to portfolio optimization having hig...
research
02/03/2016

Single-Solution Hypervolume Maximization and its use for Improving Generalization of Neural Networks

This paper introduces the hypervolume maximization with a single solutio...

Please sign up or login with your details

Forgot password? Click here to reset