solidot新版网站常见问题,请点击这里查看。
消息
本文已被查看3665次
Regularization via Mass Transportation. (arXiv:1710.10016v1 [math.OC])
来源于:arXiv
The goal of regression and classification methods in supervised learning is
to minimize the empirical risk, that is, the expectation of some loss function
quantifying the prediction error under the empirical distribution. When facing
scarce training data, overfitting is typically mitigated by adding
regularization terms to the objective that penalize hypothesis complexity. In
this paper we introduce new regularization techniques using ideas from
distributionally robust optimization, and we give new probabilistic
interpretations to existing techniques. Specifically, we propose to minimize
the worst-case expected loss, where the worst case is taken over the ball of
all (continuous or discrete) distributions that have a bounded transportation
distance from the (discrete) empirical distribution. By choosing the radius of
this ball judiciously, we can guarantee that the worst-case expected loss
provides an upper confidence bound on the loss on test data, thus offering new
generalization bou 查看全文>>