Neural Sequence Model Training via α-divergence Minimization

06/30/2017
by   Sotetsu Koyamada, et al.
0

We propose a new neural sequence model training method in which the objective function is defined by α-divergence. We demonstrate that the objective function generalizes the maximum-likelihood (ML)-based and reinforcement learning (RL)-based objective functions as special cases (i.e., ML corresponds to α→ 0 and RL to α→1). We also show that the gradient of the objective function can be considered a mixture of ML- and RL-based objective gradients. The experimental results of a machine translation task show that minimizing the objective function with α > 0 outperforms α→ 0, which corresponds to ML-based methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/22/2019

Dual Skew Divergence Loss for Neural Machine Translation

For neural sequence model training, maximum likelihood (ML) has been com...
research
03/25/2022

Machine-Learning Based Objective Function Selection for Community Detection

NECTAR, a Node-centric ovErlapping Community deTection AlgoRithm, presen...
research
12/16/2013

Comparative Analysis of Viterbi Training and Maximum Likelihood Estimation for HMMs

We present an asymptotic analysis of Viterbi Training (VT) and contrast ...
research
03/13/2021

CACTUS: Detecting and Resolving Conflicts in Objective Functions

Machine learning (ML) models are constructed by expert ML practitioners ...
research
06/01/2019

BreGMN: scaled-Bregman Generative Modeling Networks

The family of f-divergences is ubiquitously applied to generative modeli...
research
05/10/2023

Finding Meaningful Distributions of ML Black-boxes under Forensic Investigation

Given a poorly documented neural network model, we take the perspective ...
research
09/16/2019

Sinkhorn Algorithm as a Special Case of Stochastic Mirror Descent

We present a new perspective on the celebrated Sinkhorn algorithm by sho...

Please sign up or login with your details

Forgot password? Click here to reset