ESSEC METALAB

RESEARCH

META-STRATEGY FOR LEARNING TUNING PARAMETERS WITH GUARANTEES

[ARTICLE] This paper proposes a meta-strategy for online learning methods like the online gradient algorithm (OGA) and exponentially weighted aggregation (EWA) to learn tuning parameters from past tasks.

by Pierre Alquier (ESSEC Business School), Dimitri Meunier

Online learning methods, similar to the online gradient algorithm (OGA) and exponentially weighted aggregation (EWA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from past tasks. Our strategy is based on the minimization of a regret bound. It allows us to learn the initialization and the step size in OGA with guarantees. It also allows us to learn the prior or the learning rate in EWA. We provide a regret analysis of the strategy. It allows to identify settings where meta-learning indeed improves on learning each task in isolation.

[Please read the research paper here]

Research list
arrow-right
Résumé de la politique de confidentialité

Ce site utilise des cookies afin que nous puissions vous fournir la meilleure expérience utilisateur possible. Les informations sur les cookies sont stockées dans votre navigateur et remplissent des fonctions telles que vous reconnaître lorsque vous revenez sur notre site Web et aider notre équipe à comprendre les sections du site que vous trouvez les plus intéressantes et utiles.