Theoretical Machine Learning Seminar

Relaxing the I.I.D. Assumption: Adaptive Minimax Optimal Sequential Prediction with Expert Advice

We consider sequential prediction with expert advice when the data are generated stochastically, but the distributions generating the data may vary arbitrarily among some constraint set. We quantify relaxations of the classical I.I.D. assumption in terms of possible constraint sets, with I.I.D. at one extreme, and an adversarial mechanism at the other. The Hedge algorithm, long known to be minimax optimal for in the adversarial regime, has recently been shown to also be minimax optimal in the I.I.D. setting. We show that Hedge is sub-optimal between these extremes, and present a new algorithm that is adaptively minimax optimal with respect to our relaxations of the I.I.D. assumption, without knowledge of which setting prevails.

Date & Time

July 14, 2020 | 12:30pm – 1:45pm

Location

Remote Access Only - see link below

Speakers

Jeffrey Negrea

Affiliation

University of Toronto

Notes

We welcome broad participation in our seminar series. To receive login details, interested participants will need to fill out a registration form accessible from the link below.  Upcoming seminars in this series can be found here.

Register Here