0


和积网络的判别学习

Discriminative Learning of Sum-Product Networks
课程网址: http://videolectures.net/nips2012_gens_discriminative_learning/  
主讲教师: Robert Gens
开课单位: 华盛顿大学
开课时间: 2013-01-16
课程语种: 英语
中文简介:
Sum产品网络是一种新的深度架构,可以在高树宽度模型上实现快速,准确的推理。迄今为止,仅提出了用于训练SPN的生成方法。在本文中,我们提出了第一个SPNs的判别训练算法,结合前者的高精度和后者的表征能力和易处理性。我们证明了易处理的判别式SPNs类比易处理的生成型算法更广泛,并提出了一种有效的反向传播式算法来计算条件对数似然的梯度。标准梯度下降受到扩散问题的影响,但是多层网络可以是学习可靠地使用“硬”梯度下降,其中边缘推断被MPE推理替代(即,推断非证据变量的最可能状态)。生成的更新具有简单直观的形式。我们在标准图像分类任务上测试discriminiminativeSPNs。到目前为止,我们在CIFAR 10数据集上获得了最好的结果,使用了比现有方法更少的特征,SPN架构有区别地融合了局部图像结构。我们还报告了STL 10上发布的最高测试准确度,即使我们只使用数据集的标记部分。
课程简介: Sum-product networks are a new deep architecture that can perform fast, exact inference on high-treewidth models. Only generative methods for training SPNs have been proposed to date. In this paper, we present the first discriminative training algorithms for SPNs, combining the high accuracy of the former with the representational power and tractability of the latter. We show that the class of tractable discriminative SPNs is broader than the class of tractable generative ones, and propose an efficient backpropagation-style algorithm for computing the gradient of the conditional log likelihood. Standard gradient descent suffers from the diffusion problem, but networks with many layers can be learned reliably using ‘’hard’’ gradient descent, where marginal inference is replaced by MPE inference (i.e., inferring the most probable state of the non-evidence variables). The resulting updates have a simple and intuitive form. We test discriminative SPNs on standard image classification tasks. We obtain the best results to date on the CIFAR-10 dataset, using fewer features than prior methods with an SPN architecture that learns local image structure discriminatively. We also report the highest published test accuracy on STL-10 even though we only use the labeled portion of the dataset.
关 键 词: 高树宽度模型; 多层网络; 对数似然
课程来源: 视频讲座网
最后编审: 2019-09-06:lxf
阅读次数: 44