Learning with BOT - Bregman and Optimal Transport divergences - Université de Pau et des Pays de l'Adour Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2021

Learning with BOT - Bregman and Optimal Transport divergences

Résumé

The introduction of the Kullback-Leibler divergence in PAC-Bayesian theory can be traced back to the work of [1]. It allows to design learning procedure with generalization errors based on an optimal trade-off between accuracy on the training set, and complexity. This complexity is penalized thanks to the Kullback-Leibler divergence from a prior distribution, modeling a domain knowledge over the set of candidates or weak learners. In the context of high dimensional statistics, it gives rise to sparsity oracle inequalities or more recently sparsity regret bounds, where the complexity is measured thanks to 0 or 1 −norms. In this paper, we propose to extend the PAC-Bayesian theory to get more generic regret bounds for sequential weighted averages, where (1) the measure of complexity is based on any ad-hoc criterion and (2) the prior distribution could be very simple. These results arise by introducing a new measure of divergences from the prior in terms of Bregman divergence or Optimal Transport.
Fichier principal
Vignette du fichier
Learning_with_BOT.pdf (298.06 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03262687 , version 1 (16-06-2021)
hal-03262687 , version 2 (18-06-2021)

Identifiants

  • HAL Id : hal-03262687 , version 1

Citer

Andrew Chee, Sébastien Loustau. Learning with BOT - Bregman and Optimal Transport divergences. 2021. ⟨hal-03262687v1⟩
261 Consultations
259 Téléchargements

Partager

Gmail Facebook X LinkedIn More