新的损失函数的训练结构的预测Training Structured Predictors for Novel Loss Functions |
|
课程网址: | http://videolectures.net/nipsworkshops09_mcallester_tspnlf/ |
主讲教师: | David McAllester |
开课单位: | 芝加哥丰田技术学院 |
开课时间: | 2010-01-19 |
课程语种: | 英语 |
中文简介: | 作为动机,我们考虑PASCAL图像分割挑战。给定图像和目标类(例如人),挑战在于将图像分割成由该类中的对象(人前景)占据的区域和未被该类占用的区域(非人背景)。在现有技术水平下,通过预测所有背景来实现最低像素错误率。然而,使用联合的联合得分和全背景预测得分为零的属性来评估挑战。这提出了一个问题,即如何将特定的损失函数纳入结构化预测器的训练中。标准方法是将所需损失结合到结构化铰链损耗中,并观察到,对于任何损失,结构化铰链损失是所需损耗的上限。然而,这个上限非常松散,并且结构化铰链损失是处理PASCAL评估测量的适当或有用的方法还不是很清楚。本讲座回顾了解决这个问题的各种方法,并提出了一种新的训练算法,我们称之为好标签 - 坏标签算法。我们证明,在数据丰富的情况下,良好标签 - 坏标签算法遵循训练损失的梯度,假设我们只能在给定的图形模型中执行推理。该算法在结构上与结构化铰链损耗(不遵循损耗梯度)的随机次梯度下降相似但显着不同。 |
课程简介: | As a motivation we consider the PASCAL image segmentation challenge. Given an image and a target class, such as person, the challenge is to segment the image into regions occupied by objects in that class (person foreground) and regions not occupied by that class (non-person background). At the present state of the art the lowest pixel error rate is achieved by predicting all background. However, the challenge is evaluated with an intersection over union score with the property that the all-background prediction scores zero. This raises the question of how one incorporates a particular loss function into the training of a structured predictor. A standard approach is to incorporate the desired loss into the structured hinge loss and observe that, for any loss, the structured hinge loss is an upper bound on the desired loss. However, this upper bound is quite loose and it is far from clear that the structured hinge loss is an appropriate or useful way to handle the PASCAL evaluation measure. This talk reviews various approaches to this problem and presents a new training algorithm we call the good-label-bad-label algorithm. We prove that in the data-rich regime the good-label-bad-label algorithm follows the gradient of the training loss assuming only that we can perform inference in the given graphical model. The algorithm is structurally similar to, but significantly different from, stochastic subgradient descent on the structured hinge loss (which does not follow the loss gradient). |
关 键 词: | 计算生物学; 大型图形模型; 标签算法 |
课程来源: | 视频讲座网 |
最后编审: | 2020-06-29:wuyq |
阅读次数: | 75 |