Generalization of Hamiltonian algorithms

Andreas Maurer

Advances in Neural Information Processing Systems 37 (NeurIPS 2024) Main Conference Track

A method to prove generalization results for a class of stochastic learning algorithms is presented. It applies whenever the algorithm generates a distribution, which is absolutely continuous distribution relative to some a-priori measure, and the logarithm of its density is exponentially concentrated about its mean. Applications include bounds for the Gibbs algorithm and randomizations of stable deterministic algorithms, combinations thereof and PAC-Bayesian bounds with data-dependent priors.

10.52202/079017-0834