论文标题

通过普通线性回归学习无限地学习许多响应功能时的风险界限

Risk bounds when learning infinitely many response functions by ordinary linear regression

论文作者

Plassier, Vincent, Portier, François, Segers, Johan

论文摘要

考虑基于相同的输入变量同时学习大量响应函数的问题。训练数据由从共同分布和相关响应的单个独立随机样本组成。输入变量被映射到称为特征空间的高维线性空间中,并将响应函数模拟为映射特征的线性函数,并通过普通最小二乘校准系数。我们通过在响应函数中统一地控制多余风险的收敛率,为最坏情况下的过剩预测风险提供收敛保证。特征图的尺寸允许具有样本量的无穷大。响应函数的收集虽然可能是无限的,但应该具有有限的Vapnik-Chervonenkis维度。在合理的计算时间内构建多个替代模型时,可以应用界限。

Consider the problem of learning a large number of response functions simultaneously based on the same input variables. The training data consist of a single independent random sample of the input variables drawn from a common distribution together with the associated responses. The input variables are mapped into a high-dimensional linear space, called the feature space, and the response functions are modelled as linear functionals of the mapped features, with coefficients calibrated via ordinary least squares. We provide convergence guarantees on the worst-case excess prediction risk by controlling the convergence rate of the excess risk uniformly in the response function. The dimension of the feature map is allowed to tend to infinity with the sample size. The collection of response functions, although potentially infinite, is supposed to have a finite Vapnik-Chervonenkis dimension. The bound derived can be applied when building multiple surrogate models in a reasonable computing time.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源