0


结构的无监督发现、简洁的表示和稀疏性

Unsupervised Discovery of Structure, Succinct Representations and Sparsity
课程网址: http://videolectures.net/icml09_ng_udssrs/  
主讲教师: Andrew Ng
开课单位: 斯坦福大学
开课时间: 2009-08-26
课程语种: 英语
中文简介:
我们描述了一类无监督学习方法,它们学习训练数据的稀疏表示,从而识别有用的特征。此外,我们展示了这些思想的深度学习(多层)版本,基于稀疏DBN的思想,学习了丰富的特征层次结构,包括对象的部分整体分解。这是“概率最大池”的概念,它允许我们大规模地实现卷积DBN,同时保持概率上合理的语义。在图像的情况下,在最低级别,该方法学习检测边缘;在下一个层面,它将边缘放在一起形成“对象部分”;最后,在最高层将对象部分放在一起形成整个“对象模型”。该方法学习的特征对于广泛的任务非常有用,包括对象识别,文本分类和音频分类。我们还展示了将模型的两层版本(在自然图像上训练)与大脑中的视觉皮层区域V1和V2(皮层中的视觉处理的第一和第二阶段)进行比较的结果。最后,我们将结束对未来研究的一些开放性问题和方向的讨论。
课程简介: We describe a class of unsupervised learning methods that learn sparse representations of the training data, and thereby identify useful features. Further, we show that deep learning (multilayer) versions of these ideas, ones based on sparse DBNs, learn rich feature hierarchies, including part-whole decompositions of objects. Central to this is the idea of "probabilistic max pooling", which allows us to implement convolutional DBNs at a large scale, while maintaining probabilistically sound semantics. In the case of images, at the lowest level this method learns to detect edges; at the next level, it puts together edges to form "object parts"; and finally, at the highest level puts together object parts to form whole "object models". The features this method learns are useful for a wide range of tasks, including object recognition, text classification, and audio classification. We also present the result of comparing a two-layer version of the model (trained on natural images) to visual cortical areas V1 and V2 in the brain (the first and second stages of visual processing in the cortex). Finally, we'll conclude with a discussion on some open problems and directions for future research.
关 键 词: 无监督学习; 训练数据; 深度学习
课程来源: 视频讲座网
最后编审: 2019-04-24:cwx
阅读次数: 38