Optimization with access to auxiliary information

06/01/2022
by   El Mahdi Chayti, et al.
0

We investigate the fundamental optimization question of minimizing a target function f(x) whose gradients are expensive to compute or have limited availability, given access to some auxiliary side function h(x) whose gradients are cheap or more available. This formulation captures many settings of practical relevance such as i) re-using batches in SGD, ii) transfer learning, iii) federated learning, iv) training with compressed models/dropout, etc. We propose two generic new algorithms which are applicable in all these settings and prove using only an assumption on the Hessian similarity between the target and side information that we can benefit from this framework.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/15/2021

A Method to Reveal Speaker Identity in Distributed ASR Training, and How to Counter It

End-to-end Automatic Speech Recognition (ASR) models are commonly traine...
research
11/16/2021

Wyner-Ziv Gradient Compression for Federated Learning

Due to limited communication resources at the client and a massive numbe...
research
06/25/2021

Implicit Gradient Alignment in Distributed and Federated Learning

A major obstacle to achieving global convergence in distributed and fede...
research
05/26/2022

Aggregating Gradients in Encoded Domain for Federated Learning

Malicious attackers and an honest-but-curious server can steal private c...
research
12/01/2020

MYSTIKO : : Cloud-Mediated, Private, Federated Gradient Descent

Federated learning enables multiple, distributed participants (potential...
research
06/05/2021

FedNL: Making Newton-Type Methods Applicable to Federated Learning

Inspired by recent work of Islamov et al (2021), we propose a family of ...
research
07/08/2023

Learning to Group Auxiliary Datasets for Molecule

The limited availability of annotations in small molecule datasets prese...

Please sign up or login with your details

Forgot password? Click here to reset