A Novel Data-Dependent Learning Paradigm for Large Hypothesis Classes
Alireza F. Pour and Shai Ben-David
我们用一组候选模型来解决学习的一般任务,这些模型太大,无法将经验估计与真实损失的统一收敛。 虽然应对这些挑战的常见方法是基于SRM(或正则化)的学习算法,但我们提出了一种新的学习范式,依赖于更有力地整合经验数据,并且需要更少的算法决策才能基于先前的假设。 我们分析了我们方法的概括能力,并在几个常见的学习假设中展示了它的优点,包括近距离点的相似性,将域聚入高度标签同质的区域,标签规则的Lipschitzness假设以及对比学习假设。 我们的方法允许使用这些假设,而无需先验地了解它们的真实参数。
We address the general task of learning with a set of candidate models that is too large to have a uniform convergence of empirical estimates to true losses. While the common approach to such challenges is SRM (or regularization) based learning algorithms, we propose a novel learning paradigm that relies on stronger incorporation of empirical data and requires less algorithmic decisions to be based on prior assumptions. We analyze the generalization capabilities of our approach and demonstrate i...