in ,

A Modern Introduction to Online Learning, Hacker News


  

              

                                   Submitted on Dec)

    

Abstract:In this monograph, I introduce the basic concepts of Online Learning through a modern view of Online Convex Optimization. Here, online learning refers to the framework of regret minimization under worst-case assumptions. I present first-order and second-order algorithms for online learning with convex losses, in Euclidean and non-Euclidean settings. All the algorithms are clearly presented as instantiation of Online Mirror Descent or Follow-The-Regularized-Leader and their variants. Particular attention is given to the issue of tuning the parameters of the algorithms and learning in unbounded domains, through adaptive and parameter-free online learning algorithms. Non-convex losses are dealt through convex surrogate losses and through randomization. The bandit setting is also briefly discussed, touching on the problem of adversarial and stochastic multi-armed bandits. These notes do not require prior knowledge of convex analysis and all the required mathematical tools are rigorously explained. Moreover, all the proofs have been carefully chosen to be as simple and as short as possible.

            

      

Submission history

From: Francesco Orabona [view email]        [v1]Tue, Dec****************** 90: (**********************, ************************************ (UTC) 1, (KB)**********************

What do you think?

Leave a Reply

Your email address will not be published. Required fields are marked *

GIPHY App Key not set. Please check settings

Datamining Bandersnatch, Hacker News

Illinois regulates artificial intelligence like HireVue’s used to analyze online job interviews, Recode

Illinois regulates artificial intelligence like HireVue’s used to analyze online job interviews, Recode