Online Convex Optimization Perspective for Learning from Dynamically Revealed Preferences

08/24/2020
by   Violet Xinying Chen, et al.
0

We study the problem of online learning (OL) from revealed preferences: a learner wishes to learn an agent's private utility function through observing the agent's utility-maximizing actions in a changing environment. We adopt an online inverse optimization setup, where the learner observes a stream of agent's actions in an online fashion and the learning performance is measured by regret associated with a loss function. Due to the inverse optimization component, attaining or proving convexity is difficult for all of the usual loss functions in the literature. We address this challenge by designing a new loss function that is convex under relatively mild assumptions. Moreover, we establish that the regret with respect to our new loss function also bounds the regret with respect to all other usual loss functions. This then allows us to design a flexible OL framework that enables a unified treatment of loss functions and supports a variety of online convex optimization algorithms. We demonstrate with theoretical and empirical evidence that our framework based on the new loss function (in particular online Mirror Descent) has significant advantages in terms of eliminating technical assumptions as well as regret performance and solution time over other OL algorithms from the literature.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/18/2022

Online Convex Optimization with Unbounded Memory

Online convex optimization (OCO) is a widely used framework in online le...
research
11/10/2019

Grinding the Space: Learning to Classify Against Strategic Agents

We study the problem of online learning in strategic classification sett...
research
10/03/2022

Online Self-Concordant and Relatively Smooth Minimization, With Applications to Online Portfolio Selection and Learning Quantum States

Consider an online convex optimization problem where the loss functions ...
research
07/06/2017

Convergence Analysis of Optimization Algorithms

The regret bound of an optimization algorithms is one of the basic crite...
research
10/16/2020

Online non-convex optimization with imperfect feedback

We consider the problem of online learning with non-convex losses. In te...
research
12/30/2020

Risk Guarantees for End-to-End Prediction and Optimization Processes

Prediction models are often employed in estimating parameters of optimiz...
research
07/22/2019

Online optimization of piecewise Lipschitz functions in changing environments

In an online optimization problem we are required to choose a sequence o...

Please sign up or login with your details

Forgot password? Click here to reset