Federated Distillation of Natural Language Understanding with Confident Sinkhorns

10/06/2021
by   Rishabh Bhardwaj, et al.
2

Enhancing the user experience is an essential task for application service providers. For instance, two users living wide apart may have different tastes of food. A food recommender mobile application installed on an edge device might want to learn from user feedback (reviews) to satisfy the client's needs pertaining to distinct domains. Retrieving user data comes at the cost of privacy while asking for model parameters trained on a user device becomes space inefficient at a large scale. In this work, we propose an approach to learn a central (global) model from the federation of (local) models which are trained on user-devices, without disclosing the local data or model parameters to the server. We propose a federation mechanism for the problems with natural similarity metric between the labels which commonly appear in natural language understanding (NLU) tasks. To learn the global model, the objective is to minimize the optimal transport cost of the global model's predictions from the confident sum of soft-targets assigned by local models. The confidence (a model weighting scheme) score of a model is defined as the L2 distance of a model's prediction from its probability bias. The method improves the global model's performance over the baseline designed on three NLU tasks with intrinsic label space semantics, i.e., fine-grained sentiment analysis, emotion recognition in conversation, and natural language inference. We make our codes public at https://github.com/declare-lab/sinkhorn-loss.

READ FULL TEXT

page 15

page 17

research
09/07/2016

Sentiment Classification of Food Reviews

Sentiment analysis of reviews is a popular task in natural language proc...
research
04/05/2021

Exploring Transformers in Emotion Recognition: a comparison of BERT, DistillBERT, RoBERTa, XLNet and ELECTRA

This paper investigates how Natural Language Understanding (NLU) could b...
research
08/21/2023

SeqGPT: An Out-of-the-box Large Language Model for Open Domain Sequence Understanding

Large language models (LLMs) have shown impressive ability for open-doma...
research
07/01/2023

FedCP: Separating Feature Information for Personalized Federated Learning via Conditional Policy

Recently, personalized federated learning (pFL) has attracted increasing...
research
03/25/2022

Canary Extraction in Natural Language Understanding Models

Natural Language Understanding (NLU) models can be trained on sensitive ...
research
01/24/2022

On-Device Learning with Cloud-Coordinated Data Augmentation for Extreme Model Personalization in Recommender Systems

Data heterogeneity is an intrinsic property of recommender systems, maki...
research
04/28/2023

Earning Extra Performance from Restrictive Feedbacks

Many machine learning applications encounter a situation where model pro...

Please sign up or login with your details

Forgot password? Click here to reset